US20060099563A1 - Computerized teaching, practice, and diagnosis system - Google Patents

Computerized teaching, practice, and diagnosis system Download PDF

Info

Publication number
US20060099563A1
US20060099563A1 US10/982,439 US98243904A US2006099563A1 US 20060099563 A1 US20060099563 A1 US 20060099563A1 US 98243904 A US98243904 A US 98243904A US 2006099563 A1 US2006099563 A1 US 2006099563A1
Authority
US
United States
Prior art keywords
student
practice
sequence
answers
presentation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US10/982,439
Inventor
Zhenyu Lawrence Liu
Xin Chen
Lang Shie Chen
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
MULTIPI Inc
Original Assignee
MULTIPI Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by MULTIPI Inc filed Critical MULTIPI Inc
Priority to US10/982,439 priority Critical patent/US20060099563A1/en
Assigned to MULTIPI, INC. reassignment MULTIPI, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CHEN, LANG SHIE, CHEN, XIN, LIU, ZHENYU LAWRENCE
Priority to TW094112864A priority patent/TWI270023B/en
Priority to CNA2005100783943A priority patent/CN1770222A/en
Publication of US20060099563A1 publication Critical patent/US20060099563A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B7/00Electrically-operated teaching apparatus or devices working with questions and answers
    • G09B7/02Electrically-operated teaching apparatus or devices working with questions and answers of the type wherein the student is expected to construct an answer to the question which is presented or wherein the machine gives an answer to the question presented by a student

Definitions

  • One-on-one instruction is a particularly desirable teaching technique because the teacher gains real-time incite to the concepts that are giving a student difficulties, permitting the teacher to tailor instruction to the needs of that student.
  • extensive one-on-one instruction is economically impractical in most primary schools.
  • Parents can play a significant role in one-on-one teaching of their children, but parents often lack teaching skills and also experience the difficulty of identifying the specific areas where their children may be having difficulties. Further, communication barriers often make it difficult for teachers and parents to co-ordinate teaching efforts and communicate regarding the specific areas where students are having difficulties.
  • Students can also teach themselves through self-study and homework, but homework and self-study generally do not give the student immediate feedback that could guide the student to use their time most efficiently. Additionally, students may not know where to begin when attempting to master new material or concepts they do not understand. Books and other printed materials generally only provide a fixed message that is unable to adapt to the students needs, so that the self-study using printed material may not be targeted at the right concepts.
  • an automated or computer-aided teaching system enhances the traditional education by rapidly providing detailed analysis of a student's performance at the level of basic concepts and skills.
  • the evaluation can be achieved using practice problems that are divided into multiple steps associated with basic concepts and skills, so that the teaching system allows students to give step-by-step answers when solving the practice problems. Analysis of the step-by-step answers identifies specific concepts and skills that the student has yet to master. Conventional education can then be targeted at the specific concepts and skills for the greatest benefit to the students.
  • the teaching system can further use analysis of a student performance for automatic real-time selection of subsequent practice problems that are targeted to the student's needs.
  • the teaching system can periodically provide refresher problems stressing concepts or skills that were historically difficult for the student.
  • a computer-aided teaching system includes a random problem generation system allows intensive drills of the targeted subject matter without repeating or exhausting the available problems.
  • Real time diagnosis can precisely locate the errors and weaknesses and provide real-time teaching insertion through problem explanation and demonstration of problem solutions.
  • the system can therefore adapt to the student's learning behavior and speed.
  • the system can automatically generate examples, practice problems, and fixed material targeted at those concepts and skills.
  • the learning system can revisit the difficult concepts and skills at varied intervals during a learning session. The frequency of revisiting a concept or skill can change (e.g., decrease) as the student's answers to problems mark improvement or show that the student has achieved and maintained mastery of the concept.
  • computers executing a teaching system are networked to permit teachers to efficiently manage and monitor the students learning progress and analyze the results on a real-time basis for accurate adjustment of the next teaching step or strategy.
  • the computer network can be linked within a campus or to off campus computers, for example, through the Internet.
  • the teaching system when networked among teachers, parents, and students can provide efficient communications of detailed information regarding a student's studies at school or home. This interactive, bi-directional communications can improve the efficiency and quality of education.
  • One specific embodiment of the invention is an automated teaching process.
  • the process begins with selecting a problem template from among a set of problem template, and randomly generating values for choices identified in the selected problem template.
  • the random choices can be numbers, words, or identifiers for objects such as images that are part of a presentation of the problem.
  • the process additionally presents a sequence of one or more questions that lead to a solution of the problem. The presentation of each question generally depends on the randomly generated values and may depend on the student's previous answers.
  • the student's answers to the questions can be compared to answers derived from the randomly generated values and formulas from the selected problem template.
  • Diagnosis codes that correspond to the answers can identify performance values that are updated to reflect whether the answers from the student were correct. These performance values can indicate weaknesses and control whether the process provides additional teaching or review problems and the frequency of review or the speed of question presentation.
  • the selected problem template can identify two or more sequence of questions leading to the solution with one sequence being shorter than the other, so that the process can select shorter or longer sequence of questions depending on the evaluation of the student's answers.
  • Another specific embodiment of the invention is an automated teaching process that includes: selecting a problem template from among a set of problem templates; randomly generating values for respective choices identified in the selected problem template; and receiving a selection of either a practice mode or an example mode.
  • the process presents a sequence of one or more questions that depend on the randomly generated values and lead to a solution of the problem.
  • the process presents a sequence of one or more instructions that similarly depend on the randomly generated values and lead to the solution of the problem.
  • the process can enable the student to switch between the practice mode and the example mode, for example, when the student is unsure on how to answer a question.
  • Another specific embodiment of the invention is a teaching system including a problem body table, a practice body table, an example body table, a practice presentation unit, and an example presentation unit.
  • the problem body table contains entries corresponding to problems, and each entry identifies random choices and answers.
  • the practice body table contains entries corresponding to the entries in the problem body table, and each entry in the practice body table includes information representing at least one sequence of questions leading to a solution of the problem associated with the corresponding entry in the problem body table.
  • the example body table contains entries corresponding to the entries in the problem body table, and each entry in the example body table includes information representing a sequence of instructions leading to the solution of the problem associated with the corresponding entry in the problem body table.
  • the practice presentation unit uses information from corresponding entries of the practice body table and the problem body table for presentation of the sequence of questions, and the example presentation unit uses information from corresponding entries of the example body table and the problem body table for presentation of the sequence of instructions.
  • Another specific embodiment of the invention is a teaching system that includes a practice problem presentation system, a user progress table, a diagnosis engine, and a management unit.
  • the practice problem presentation system presents practice problems from a sequence of lessons.
  • the diagnosis engine evaluates answers received from a student during presentation of the practice problems based and then updates the user progress table to indicate which of the lessons are complete.
  • the management unit can access the user progress table and a schedule to determine whether the student is on schedule for progress through the lessons
  • Another specific embodiment of the invention is an automated teaching process that includes: setting a progress clock for completion of a sequence of lessons; presenting lessons to a user at a user controlled rate; comparing the user's progress through the lessons to the progress clock; and determining whether the user is ahead or behind schedule. Depending on whether the user is ahead or behind schedule, the process can then adjust teaching parameters such as the number of practice problems provided per lesson, the difficulty of practice problems, and the number or rate of review problems inserted into lessons.
  • Another specific embodiment of the invention is an automated teaching process including: presenting a current question from a sequence containing multiple questions leading to a solution for the problem; receiving an answer to the current question from a student; and repeating the presenting and receiving steps until the student has reached the solution.
  • the student can control whether consecutive presentations of questions go on to a next question in an order of the sequence.
  • the process can also include highlighting an area of a presentation of a question and presenting an explanation of how content of the area relates to answering the question.
  • the process can include: evaluating each answers to the questions in the sequence; locating diagnosis codes respectively corresponding to the answers; and updating progress categories corresponding to the diagnosis codes to reflect which of the answers where correct.
  • FIG. 1 shows flow diagram of an educational practice process in accordance with an embodiment of the invention.
  • FIG. 2 shows block diagram of a teaching system in accordance with an embodiment of the invention.
  • FIG. 3 is a block diagram of tables suitable for use in the teaching system of FIG. 2 .
  • FIG. 4 is a flow diagram of a teaching process in accordance with an embodiment of the invention that reduces the number of presented steps during practice of a specific type of problem.
  • FIG. 5 is a flow diagram of a teaching process in accordance with an embodiment of the invention where the user selects the logical flow of a multi-step solution.
  • FIG. 6 is a block diagram illustrating the relationship of a teaching insertion unit and a diagnosis engine.
  • FIG. 7 is a block diagram illustrating an embodiment of the invention in which an interactive web site provides a learning system that is accessible to students, teachers, and parents.
  • a teaching system in accordance with an embodiment of the invention provides explanatory materials, examples, and problems on an interactive basis and adapts to the needs of a student. Additionally, the teaching system can provide detailed analysis or diagnosis of the student's performance that can be communicated to teachers and/or parents for use in other teaching environments.
  • FIG. 1 is a flow diagram illustrating a teaching process 100 used by a student.
  • Teaching process 100 is generally a computer-aided process in which either dedicated hardware or software implements portions of process 100 as described further below.
  • Teaching process 100 is generally adaptable to teaching a variety of subjects at many different educational levels.
  • teaching process 100 is directed to primary (K-12) education and the teaching of mathematics, science, or a similar subject in which the solving of problems may be divided into a multi-step logical process.
  • Subjects such as mathematics and science are particularly suited for computer-aided learning because problems in these subjects are often numerically dependent and evaluated.
  • a subject matter and level is selected at the start of process 100 .
  • a person such as a teacher, a parent, or a student in step 110 simply selects a subject, area, and level for process 100 .
  • a teacher may give a class assignment that identifies a particular lesson for students.
  • Such lessons can be categorized using a standardized teaching curriculum that educational authorities commonly create.
  • lessons can be categorized using the academic standards of the California State Board of Education, as set forth for the CAT/6 standardized tests for the K-12 students.
  • Each student selects in step 110 the assigned subject, area, and level.
  • the student may take a placement test in step 120 and an automated assignment in step 125 selects the subject, area, and/or level for the student using teaching process 100 .
  • Step 130 generates a random problem based on the selected or assigned subject, area, and level. Selection of the random problem may further be based on the student's prior performance as described further below.
  • the generated problem is randomized first in selection from among problem templates in a class associated with the subject, area, and level of process 100 .
  • the selected problem template is further randomized through selection of items required for construction of a complete problem.
  • the selections of items for a particular problem template may include, for example, numeric values that can be selected randomly from a target set or range so that numeric solutions or answers are seldom repeated. Further, items in the content, wording, graphics, or other presentation features associated with the template can be randomly selected from associated lists and provide variety in the random problems that will ideally help hold the students interest in solving practice problems.
  • Each problem template will generally be associated with a multi-step problem solving process.
  • the general problem of adding fractions may be associated with the steps of finding a common denominator from the fractions being added, multiplying the numerator and denominators of the fractions by appropriate factors, adding the resulting numerators of the fractions after multiplication, finding common factors in the numerator and denominator of the result, and simplifying the result accordingly.
  • Associated with each step of a template are an instruction, a question, and one or more formulas for solution of that step.
  • Each answer in that the student provides during the multi-step process may have a diagnosis code identifying a skill.
  • the diagnosis codes may correspond to skills such as finding common denominators, multiplying numerator and denominator, simplifying fractions.
  • the instructions associated with a template are generally for use when the teaching system presents the random problem as an example of problem solving technique.
  • Process 100 uses the question associated with a step of the problem when the student is solving practice problems.
  • step 140 chooses whether to see an example solution of a problem, attempt to solve a problem, or exit process 100 . If the student chooses to see an example solution of the problem, step 150 presents a multi-step solution of the random problem. For each step of the problem, process 100 presents the instruction associated with the step as completed using the randomly selected items and numeric values. Teaching process 100 can thus lead the student step by step through the solution of the randomly generated problem.
  • Process 100 executes a practice routine 160 if the student in step 140 decides to practice solving problems.
  • practice routine 160 includes presenting a student with a question associated with a current step of the random problem.
  • the question will typically include the random numbers or items selected in step 130 .
  • the question may seek one or more numeric answers or may ask the student select from a number of choices on how to proceed logically in solving the problem posed.
  • the student answers the question in step 164 before the teaching system in step 166 decides whether practice routine 160 is finished. If practice routine 160 for the current random problem is not complete, step 168 selects the next step associated with the problem.
  • Selection of the next step may be based on the student's answer to the last question and/or the student's past performance on similar problems. For example, if the student answered the last question incorrectly, step 168 may ask the student to try again. If the student answered correctly and prior performance indicates the student has mastered the concept associated with the next question, step 168 may choose a question that combines two or more steps of the problem solution.
  • An evaluation routine 170 evaluates the student's performance and can be begun when practice routine 160 is complete or alternatively after each question.
  • Evaluation routine 170 in step 172 identifies and tabulates the errors that the student made during practice routine 160 . More specifically, the identified errors can be classified by the concept or concepts associated with the incorrect answer, and a table can record the number of right and wrong answers associated with each concept.
  • Step 174 determines whether a student's mastery of a particular concept is weak. A large number of errors associated with a particular concept or skill would indicate a weakness in the student's mastery of that concept or skill.
  • a review routine 180 for the detected weakness can be initiated.
  • Review routine 180 may include a concept review step 182 in which the teaching system provides a description or explanation of the concept to the student. One or more solved examples that use the concept can then be presented in step 184 , and targeted practice problems are given in step 186 .
  • teaching process 100 can return to step 130 for generation of another random problem or return to evaluation routine 170 to complete further evaluation of the student's performance.
  • the review process and routine 180 in general can be executed at anytime after a weakness is detected. In particular, routine 180 may be delayed while the students finishes a block of the problems, so that the student is provided with one of more review processes at the end of a lesson, a section, or a chapter.
  • step 176 calculates a speed for the practice problems of the evaluated type. For example, if a student masters the concepts or skills being presented by a specific type of practice problem, the number of practice problems using the mastered concepts or skills can be reduced. The practice problems can also be presented using fewer steps, and the calculated speed based on the student's mastery of the concepts or skills may control the number of steps used in the multi-step practice problems.
  • Step 178 calculates a frequency for revisiting previously identified weaknesses.
  • the frequency for the revisiting an identified weakness is proportional to the student's error rate for questions corresponding to the identified weakness. For example, if the error rate is 100% for problems corresponding to the weakness, the revisit frequency can be set at a highest level, e.g., a revisit index 9 . If the error rate is 90%, the revisit frequency can be set somewhat lower, e.g., a revisit index 8 . If error rate is below a target level (e.g., 10%), the revisit frequency can be set to zero. After each cycle of practice, e.g., each problem or lesson, the revisit index can be incremented, and a review process can be started when the revisit index reaches a target level, e.g., a revisit index 10 .
  • a target level e.g., a revisit index 10 .
  • Step 190 which in FIG. 1 is executed after evaluation routine 170 , determines whether to revisit any of the concepts that have given the student particular difficulties. If step 190 determines it is time for a revisit, the next practice problem for the student will be a random problem that specifically uses the concept targeted for the revisit. The revisited practice problems can be the same type of problems identified and used in concept review routine 180 . If step 190 , determines it is not yet time to revisit any concepts, teaching process 100 branches back to step 130 to generate another random problem of the type that the student is currently studying.
  • Process 100 as described above can be implemented in hardware or software. Such software may be embodiment as machine-readable instructions in a storage medium such as a floppy disk, a CD ROM, a DVD ROM, or a memory stick to name a few.
  • embodiment process 100 is implemented in software for a standalone computer, one or more computers connected to a local area network, or computers interconnected through a wide area network or the Internet.
  • software that implements process 100 can be accessible and used through a web site.
  • FIG. 2 illustrates the functional blocks of a system 200 for implementing a teaching process in accordance with an embodiment of the invention.
  • System 200 includes a random problem generator 210 , presentation units 212 and 214 , mapping tables 220 , a diagnostic engine 230 , a teaching insertion unit 235 , a review insertion unit 240 , a revisit frequency adjustment unit 245 , a practice speed adjustment unit 250 , a report summary unit 260 , a progress clock 265 , and a management unit 270 .
  • mapping tables 220 uses mapping tables 220 to generate and present random problems for a specific grade, chapter, section, subject, lesson, or level.
  • mapping tables 220 can include a problem body table 310 , an example body table 320 , a practice body table 330 , a diagnosis code table 340 , user progress tables 350 , and a teaching body table 360 .
  • Random problem generator 210 uses the student's level and/or the assigned lesson to generate a pointer that locates related entries in problem body table 310 and then selects one of the entries in problem body table 310 .
  • Each entry generally includes a label identifying information such as a grade, chapter, section, subject, lesson, level, example, variation, and type for the problem.
  • the entry in problem body table 310 also includes code that identifies information that is common to both practice problems and example problems.
  • the code within each entry of problem body table 310 identifies a set of random choices rl to rn and a set of answers al to am, where n is the number of choices and m is the number of inputs required for generation and completion of a random problem.
  • Each entry in problem body table 310 generally has one or more corresponding entries in each of example body table 320 and practice body table 330 .
  • the entries in example body tables 320 provide information that is specific to presentation of an example problem
  • the entries in practice body tables 330 provide information that is specific to presentation of a practice problem.
  • Equation 1 gives a general form for a procedure corresponding to a step, i.e., an instruction or a question that is part of a presentation of an example or practice problem.
  • a single-step example or practice problem may only require one presentation P, but more generally an N-step example or practice problem may include N presentations PI to PN, with each being associated with a corresponding step.
  • presentation P represents the question or instruction body with a format F for the display, sound, or other media functions of presentation P. Random choices ri and answers ai are used with fixed content of a presentation at locations identified in format F.
  • format F can indicate the locations of cells on a display screen, and the presentation can fill each cell with fixed content or an object that depends on a random choices ri or answers ai.
  • fixed content in the presentation may include fixed sentences, mathematical expressions, a set of available choices, pull down selections, multiple dimension geometric graphics, multiple dimension coordinates, still images, moving pictures, sound, and a mixture of these.
  • P F ( r 1 , r 2, . . . , rn, a 1 , a 2, . . . , am ) Equation 1
  • Random choices rl to rn which are generically referred to herein as choices ri, correspond to variables that may be displayed in the question or instruction body.
  • Each choice ri can represent a number in any format, a set of one or more letters, one or more words, a multiple dimension geometric graphic, a set of coordinates, a still image, a moving image, or a sound.
  • Random problem generator 210 can randomly select each choice ri from a number array, a table, or a call from a file depending on an identifying code for the choice ri. Some choices ri may alternatively be calculated from or dependent upon other choices ri. Choices ri are generally selected once for a problem and may be displayed in the presentations of multiple steps.
  • Answers al to am which are generically referred to herein as answers ai, are inputs in practice mode and can be determined using formulas from tables 220 in example mode. Each answer ai would normally be received from the user when the random problem is used for practice but alternatively may be calculated from choices ri and associated formulas when the random problem is an example.
  • Format F defines the locations of calculated answers ai in an instruction body or cues for the user's input in a question body.
  • answers ai can be numbers in any format, letters, words, multiple-choice selections, pull down selections, multiple dimension geometric drawing.
  • a presentation of a step may provide a text box, a menu, radio buttons, or similar graphic used as a marker or queue that is part of the interface for user assignment of the answer ai.
  • the user can then input answers ai using a keyboard, a mouse button (or mouse movements on a coordinate system defined on the screen), a number or symbol picker, or a voice input.
  • a symbol picker the system displays a pad containing items such as number or other symbols adjacent to the marker in response to the student moving a curser to the marker. The student can then use a mouse to select items on the pad and identify the answer being entered.
  • Multi-step practice problem presentation block 212 generally uses a question format F from practice body table 330 and choices ri from random problem generator 210 for presentation of each question associated with a practice problem.
  • multi-step example presentation block 214 can use instruction formats from example body table 320 for presentation of the instructions in one or more steps of an example problem.
  • a presentation P may include multiple steps that a user activates upon completion of a previous step, and the next step in a presentation can very depending on user's selections or answers in previous steps. For example, if the user provides an answer ai in a prior step, the answer or values derived from the answer may be automatically displayed in presentations of following steps before the user provides additional input. The repeating or use of prior answers ai can make the connection between steps more natural and obvious to the user.
  • the user may is not limited to a specific method or order of steps for solving a problem, and the presentation may give the user a choice among more than one way of solving certain problems.
  • the multi-step presentation has several advantages over a single process where a user is simply presented with a problem and provides only the final answer.
  • the multi-step presentation allows the user to focus on and learn each step needed to solve the problem.
  • the multi-step logic also teaches a logical thought process for solving the problem. Further, multiple answers from the user permits diagnosis or analysis that can accurately locate the source of a user's error.
  • FIG. 4 illustrates an example of a teaching process 400 were the number of steps presented for solving a specific type of practice problem changes according to the user's performance in solving problems having the same type code.
  • step 410 the user begins practice of a particular type or class of problems
  • a presentation step 420 presents a full set of steps/questions for the user to answer.
  • a decision step 430 observes the user's answers to determine whether the user has reached a target level of mastery.
  • step 430 If the user's performance suggests further practice with the full multi-step presentation, e.g., if the number of errors is above a target level, process 400 branches from step 430 back to step 420 . If step 430 recognizes that the user has achieve the target level of mastery, step 440 provides a presentation version having a reduced number of steps, for example, in which some of the initially presented steps are combined. Combining steps in this manner tests whether the user can perform the combined steps without prompting. A decision step 450 determines whether the user has reached a level of mastery suitable for further reduction in the number of presented steps. Step 460 presents a practice problem using the smallest number of steps, e.g., a one-step presentation.
  • Step 470 then leaves the user to practice at the speed providing the fewest steps to demonstrate the user's ability to solve the problem without prompting.
  • Process 400 is complete when step 470 identifies that the user has reached the required level of mastery of that specific type of problem. Process 400 can then be used for problems of a new type.
  • FIG. 5 illustrates a process 500 in which a user selects the steps for solving a practice problem.
  • Process 500 begins with a step 510 presenting the user with options for an initial step in the solution of the problem.
  • step 510 may be skipped.
  • Decision step 520 then directs process 500 to a presentation 530 X, 530 Y, or 530 Z of a question according to the user's selection, and the user in step 540 provides answers to the presented question.
  • Decision step 550 determines whether the solution of the problem is complete, and if not branches back to step 510 .
  • Each question presentation in step 510 will in general provide a different number of choices with the choices being dependent on the user's prior selections.
  • Another feature of teaching system 200 of FIG. 2 is the ability of the user to switch from a practice mode to an example mode.
  • the user can switch to example mode.
  • Example presentation unit 214 can then present an instruction for the current step (or the example problem from its beginning) since both practice presentation unit 212 and example presentation unit 214 use the same random choices ri.
  • the user can also be allowed to step backward in a multi-step example or practice problem if desired.
  • Practice presentation unit 212 or example presentation unit 214 can also provide an alternative help mode by displaying explanatory texts, graphs and/or voice that will explain the calculation, logic, or reasoning for a particular step or the relation of a particular choice ri or answer ai to the desired solution of the problem.
  • the dynamic position corresponding to the feature of a presented question can be activated by highlighting or otherwise altering the appearance, for example, of certain choices ri and answers ai.
  • the choices ri and answers ai that are the elements involved in explanatory texts, graphs, and/or voice can be similarly altered or highlighted, and the explanations can be displayed over or near the feature explained.
  • System 200 includes an architecture that has mapping tables 220 that can be linked to each other as shown in FIG. 3 .
  • problem body table 310 contains executable code for generating portions of the problem that apply both to practice and example problems.
  • the code in problem body table 310 can be linked and called out by random problem generator 210 , for example, for or during selection of random choices ri or as part of the presentation of a problem.
  • each entry in problem body table 310 also identifies random choices ri and answers ai for a corresponding problem.
  • the example body table 320 similarly contains code for the example body, and example presentation unit 214 can link to and called out the code from table 320 .
  • Each entry of example body table 320 includes a mapping or link to a corresponding entry of problem body table 310
  • each entry of problem body table 310 includes a mapping or link back to a corresponding entry of example body table 320 .
  • Practice body table 330 similarly contains the practice problem bodies that practice problem presentation unit 212 can call out for presentation of practice problems.
  • Entries of example body table 330 include mappings or links to corresponding entries of problem body table 310 and/or example body table 320 and visa versa for use of information and/or executable code from tables 310 and 330 when system 200 operates in practice mode.
  • entries from tables 310 and 320 can be called out bi-directionally for use of information or code from example body table 320 and/or practice body table 330 , for example, when switching between a practice mode where practice problems are presented to an example mode where example problems are presented.
  • Diagnosis engine block 230 can access practice body table 330 , diagnosis code tables 340 , and user progress tables 350 when evaluating which answers ai are correct and tabulating the right and wrong answers in specific skill areas.
  • Diagnosis code tables 340 are a group of tables contain diagnosis codes used by diagnosis engine 230 . Each of the diagnosis code tables 340 is linked to a corresponding entry in problem body table 310 and therefore linked to a specific problem. The entries in each of the diagnosis code tables 340 correspond to the answers ai associated with the corresponding problem, and each entry is a diagnosis code that identifies a subject or skill associated with answer ai.
  • diagnosis engine 340 When evaluating a user's response during a practice problem, diagnosis engine 340 as described further below identifies whether an answer ai is correct or incorrect, accesses an entry corresponding to the answer ai in the diagnosis code table associated with the evaluated problem, and tabulates the correct or incorrect answer in a corresponding entry of user progress table 350 .
  • User progress tables 350 are used to indicate the user/progress status information such as: a current level, a current work on point, an overall grade, detail grades, an overall diagnosis result, detail diagnosis results, concepts or lessons needing review, problems that need revisiting, and speed up/speed down parameters.
  • the status entries in tables 350 are linked to corresponding the entries in tables 310 , 330 , and 340 .
  • Diagnosis engine 230 evaluates answers ai to practice problems to identify specific weaknesses in problem solving skills and/or overall weaknesses in completing a class of problems.
  • diagnosis engine 230 contains logic and equations that are linked with diagnosis code tables 340 for performing the diagnostic functions.
  • Four primary diagnostic functions of diagnosis engine 230 are real time weakness detection, overall weaknesses analysis, detection of subjects to be revisited, and adjustment of the learning speed during practice problem presentations.
  • diagnosis engine 230 activates error checking as soon as the user finished/answered the question or problem.
  • the error checking compares the user's answers to calculated answers.
  • the answers can be calculated using the random choices ri for the problem and the formulae from problem body table 310 .
  • Diagnosis code tables 340 can also assign each answer ai to a specific problem type and/or to a specific skill, and diagnosis engine 230 can then update user progress tables 350 to reflect correct and incorrect answers for the identified skills or problem types.
  • diagnosis engine 230 tabulates every answer ai as a right or wrong answer for the corresponding category or categories and for each incorrect answer gives the correct answer with example calculations or logical reasons for why or how the correct answer is found.
  • Diagnosis engine 230 can also activate teaching insertion unit 235 , which can then provide a fixed presentation to explain the concepts or techniques that diagnosis engine 230 identified as needing to be improved. Diagnosis engine 230 preferably presents the correct answers, the reasoning needed to arrive at the correct answer, and the list of the areas or skills requiring study on or near the user's error on a work sheet that practice presentation unit 212 provided.
  • diagnosis engine 230 When the user finishes a section of practice problems, for example corresponding to a lesson, unit, chapter, grade, . . . etc), diagnosis engine 230 performs an overall weaknesses analysis based on error accumulations in user progress tables 350 . Such analysis can grade the overall correct/level percentage for this section of practice problems and grade each concept, technique, or skill used in the section of practices problems. As precursor to the overall grading or analysis, diagnosis engine 230 can calculate a percentage of right answers or a weighted average CRT for each diagnosis code ci as indicated in Equation 2.
  • a teaching insertion unit 235 can be called to present explanations of the identified concepts or skills having a low percentage of right answers.
  • FIG. 6 illustrates the relationship of teaching insertion block 235 to diagnosis engine 230 and practice problem presentation unit 212 in an exemplary embodiment of the invention.
  • diagnosis engine 230 uses a diagnosis code corresponding to a question or skill to identify an entry in teaching body table 360 .
  • the identified entry which corresponds to a diagnosis code identified for teaching insertion, can contain explanations of the concepts or skills corresponding to the diagnosis code and a detailed explanation of how to solve a corresponding problem.
  • Teaching insertion unit 235 includes a routine 610 for presenting a review of the required concepts or skills and a routine 620 for presenting the detailed explanation of how to solve the problem.
  • Teaching insertion unit 235 can provide a fixed detailed explanation in greater detail than does the example problems that example presentation unit 214 presents or alternatively may use a random example problem that example presentation unit 214 presents. Teaching insertion unit 235 can also activate practice problem presentation unit 212 to present a set of random problems corresponding to the identified diagnosis code. Diagnosis engine 230 can then evaluate this set of random problems to monitor the user's improvement.
  • Diagnosis engine 230 can also activate speed adjustment unit 250 to adjust the speeds for specific types of problems based on the overall error percentages.
  • User progress tables 350 can also be updated as part of the overall evaluation of a section to show completion of the section and overall grades or evaluations of the section.
  • Diagnosis engine 230 can thus make the overall grades/evaluations as well as the specific percentages of right/wrong answers for specific questions or skills available to report summary unit 260 for generation of reports for teachers, parents, or other users.
  • the practice revisit detection function of diagnosis engine 230 can also be initiated when the user finishes a section (e.g., a lesson, unit, chapter, or grade) of practices problems.
  • the revisit detection generally includes an overall analysis based on the percentage of correct answers for a particular problem type. If the percentage of correct answers for a particular type of problem or skill is below the fixed or floating threshold, that type of problem or skill can be target for review and/or revisiting.
  • Revisit frequency adjustment unit 245 can then set a frequency for revisiting the identified problem type or skill, so that review insertion unit 240 will be activated to present the user with targeted review problems.
  • revisit frequency adjustment unit 245 updates user progress table 350 to list each diagnosis code that diagnosis engine 230 identifies for a revisit and to provide a revisit frequency that can be calculated from the percentage of correct answers associated with the diagnosis code.
  • Review insertion unit 240 under direction of management unit 270 can start a revisit function that inserts a practice problem of the type or types that diagnostic engine 230 identified for revisiting. Each revisited problem is generally inserted among a current series of practice problems corresponding to a subsequent lesson, for example, when a student is practicing a lesson following the lesson in which a weakness was detected.
  • management unit 270 can set a progress break that allows review insertion unit 240 to perform the revisit function at the frequency set by revisit frequency adjustment block 245 .
  • revisit frequency adjustment block 245 determines a revisit frequency RF m for each diagnosis code m that diagnosis engine 230 identifies for revisiting.
  • Equation 3 provides an example of a suitable formula for revisit frequency RF m .
  • CRV m is the weighted percentage of correct answers for diagnosis code m
  • T CRV is threshold corresponding to a pass or fail level and is greater than weighted percentage CRV m since diagnosis code m is selected for revisiting
  • s m is a proportional adaptive parameter for diagnosis code m.
  • Parameter s m can be stored in a user progress table 350 and may depend on the student's over performance, the grade level, the problem difficulty, and the error importance.
  • Parameter s m can thus scale the revisit frequency depending on the significance of the diagnosis code m. For example, parameter s m for a problem type m that is relatively unimportant for the student's current grade level may be small to reduce the revisit frequency for the less important problems.
  • RF m s m ( T CRV ⁇ CRV m ) Equation 3
  • Diagnosis engine 230 also provides the needed input for learning speed adjustment.
  • the learning speed adjustment can be based on the student's recent performance in completing problems.
  • diagnosis engine 230 can run on the overall analysis to get weighted percentage CRV ci of correct answers for diagnosis code ci during the most recent practice session.
  • Equation 4 gives a suitable formula for weighted percentage CRV.
  • np is the total number of evaluated practice problems with diagnosis code ci in the last practice section
  • j is an index identifying the particular practice problems
  • PR ij is 1 if answer ai in problem j is correct and is 0 otherwise
  • q j is a weight assigned to each problem j.
  • the practice speed is set for all problems in the next session of practice according to an average of the percentage of correct answers for practice problems in the last practice session.
  • Equation 5 gives a formula for a learning speed parameter LSP the weight percentage CRV m correct answers, the number nd of diagnosis codes associated with the last practice session, and weighting factors w m for the diagnosis codes.
  • Weighting factors w m can indicate an importance factor for each diagnosis codes. For example, if a diagnosis code i corresponds to important subject matter, then weighting factor w i can be set to 1, otherwise weighting factor w i could be less than 1.
  • Progress clock 265 in an exemplary embodiment of the invention controls or selects the main lessons that system 200 of FIG. 2 presents. Unless the user directly selects particular subject matter, system 200 can present lessons sequentially in a linear or non- linear manner as desired for completion of series of lessons within a selected time span. Additionally, as noted above, revisited problems can be inserted into the series of lessons to review concepts that were originally more difficult for the student.
  • the time span for a series of lessons could be a calendar or school year, semester, a quarter, a month, or any desired time period set for learning the subject matter contained in the series of lessons.
  • progress clock 265 will be synchronized to the present calendar time movement.
  • progress clock 265 can be for comparison of the student's progress as noted in user progress tables 350 to a schedule for completing the lessons in the allotted time.
  • Progress clock can also provide a time parameter to management unit 270 .
  • Management unit 270 can then perform functions such as sending a warning signal to the user stating his/her progress is behind the schedule and then intensify the learning/practice efforts to let the user catch up.
  • management unit 270 can send a congratulation signal to the user indicating that the student's progress is ahead of schedule.
  • Central management unit 270 or the user can then choose to do more difficult practice problems, do extra learning/practice for any previously detected weaknesses, or make up some learning/practice/subject which was skipped, for example, due to a late sign-up for use of system 200 .
  • Management unit 270 more generally has interfaces to all of the other units in system 200 and can generally control operation of system 200 , for example, to implement the teaching process of FIG. 1 .
  • Report summary unit 260 processes the data stored in user progress tables 350 to generate reports. Such reports can be provided in a variety of formats such as raw data, a table, or a graph and may include a user-selected combination of information about the student's progress.
  • Some of the information about a student's progress that may be included in a report includes: the student's current level, the current lesson, overall performance, detail performance categorized by any or all of the subjects having diagnosis codes, related performance indicating the student's strengths and needs, overall diagnosis result listing the student's overall weaknesses, detail diagnosis result listing specifics or details regarding the weaknesses, concept or lessons that the student needs to review, questions/problems that need revisiting, the speed up/speed down parameter, the student's overall or detailed progress in comparison to the master calendar, the time efficiency of the student's overall progress, and the students' time efficiency for recent progress.
  • Such reports can be printed or transmitted electronically via a network such as a local area network (LAN) or a wide area network such as the Internet to a computer used by a teacher.
  • LAN local area network
  • the Internet a computer used by a teacher.
  • teaching system 200 is implemented on a standalone computer that the student uses for at home or at school.
  • a group of students working on separate computers at a central location such as a school or at separate locations can use copies of system 200 as described above.
  • An advantage of the random problem generation described above is that each student in a class can be provided with different problems (e.g., with different numeric answers) that teach the same subject. This makes cheating or sharing of answers more difficult and does not increase the difficulty of grading or evaluating the students' answers.
  • a teacher with access to the network can monitor the performance and progress of all of the students and quickly have student evaluations that would take a great deal of time to generate by hand.
  • FIG. 7 illustrates an exemplary embodiment of a learning system 700 in which a web site 710 provides a teaching process or system similar to those described above in regard to FIGS. 1 and 2 .
  • Students 720 use web site 700 by logging in and viewing examples or solving practice problems.
  • a teacher 730 can similarly log on to web site 700 to communicate with students 710 and/or parents 740 , to assign lessons to students 710 , or to access reports on any of the students 710 that web site 710 identifies as being in the class of teacher 730 .
  • Parents or guardians 740 can also log on to web site 710 and have access to information from teacher 730 or reports on the performance or needs of their child 710 .

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Physics & Mathematics (AREA)
  • Educational Administration (AREA)
  • Educational Technology (AREA)
  • General Physics & Mathematics (AREA)
  • Electrically Operated Instructional Devices (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

A teaching system provides explanatory materials, examples, and problems on an interactive basis and adapts to the student. The system can generate random questions for presentation to the student as multi-step examples or practice problems. The student's answers during each step of a multi-step practice problem can be evaluated to identify weaknesses. Identified weaknesses can be remedied by presenting additional explanations, providing additional practice problems, and revisiting the weaknesses by inserting practice problems of the identified type into the student's subsequent lessons. The system can also adapt adjusting the speed or number of steps presented according to the student's performance. The system can further provide detailed analysis or diagnosis of the student's performance that can be communicated to teachers and/or parents for use in other teaching environments. The teaching system when embodied in a web site further provides communication among students, teachers, and parents.

Description

    BACKGROUND
  • Economic realities generally limit the available resources that can be dedicated to education of children. Accordingly, educational tools are sought that can improve the efficiency of the teaching process. In particular, since the time available for teachers to interact with students either one-on-one or in classes is limited, teachers generally desire to employ the most efficient teaching systems and methods. Efficiency in teaching generally requires that the teachers appreciate which concepts are difficult for students and which concepts the students understand well. Testing can identify concepts that students have failed to master, but testing and thorough evaluations of tests can be time consuming.
  • One-on-one instruction is a particularly desirable teaching technique because the teacher gains real-time incite to the concepts that are giving a student difficulties, permitting the teacher to tailor instruction to the needs of that student. Unfortunately, extensive one-on-one instruction is economically impractical in most primary schools. Parents can play a significant role in one-on-one teaching of their children, but parents often lack teaching skills and also experience the difficulty of identifying the specific areas where their children may be having difficulties. Further, communication barriers often make it difficult for teachers and parents to co-ordinate teaching efforts and communicate regarding the specific areas where students are having difficulties.
  • Students can also teach themselves through self-study and homework, but homework and self-study generally do not give the student immediate feedback that could guide the student to use their time most efficiently. Additionally, students may not know where to begin when attempting to master new material or concepts they do not understand. Books and other printed materials generally only provide a fixed message that is unable to adapt to the students needs, so that the self-study using printed material may not be targeted at the right concepts.
  • Educational tools are thus needed that will aid teachers and parents in evaluating the strengths and weaknesses of their students to enable efficient targeted instruction. Tools are also desired that will give a student self-study options that automatically adapt to the needs of the student and give the student immediate feedback that can be shared with teachers and parents.
  • SUMMARY
  • In accordance with an aspect of the invention, an automated or computer-aided teaching system enhances the traditional education by rapidly providing detailed analysis of a student's performance at the level of basic concepts and skills. The evaluation can be achieved using practice problems that are divided into multiple steps associated with basic concepts and skills, so that the teaching system allows students to give step-by-step answers when solving the practice problems. Analysis of the step-by-step answers identifies specific concepts and skills that the student has yet to master. Conventional education can then be targeted at the specific concepts and skills for the greatest benefit to the students. The teaching system can further use analysis of a student performance for automatic real-time selection of subsequent practice problems that are targeted to the student's needs. In accordance with another aspect of the invention, the teaching system can periodically provide refresher problems stressing concepts or skills that were historically difficult for the student.
  • In accordance with another aspect of the invention, a computer-aided teaching system includes a random problem generation system allows intensive drills of the targeted subject matter without repeating or exhausting the available problems. Real time diagnosis can precisely locate the errors and weaknesses and provide real-time teaching insertion through problem explanation and demonstration of problem solutions. The system can therefore adapt to the student's learning behavior and speed. When concepts and skills that are identified as difficult for a student, the system can automatically generate examples, practice problems, and fixed material targeted at those concepts and skills. The learning system can revisit the difficult concepts and skills at varied intervals during a learning session. The frequency of revisiting a concept or skill can change (e.g., decrease) as the student's answers to problems mark improvement or show that the student has achieved and maintained mastery of the concept.
  • In accordance with yet another aspect of the invention, computers executing a teaching system are networked to permit teachers to efficiently manage and monitor the students learning progress and analyze the results on a real-time basis for accurate adjustment of the next teaching step or strategy. The computer network can be linked within a campus or to off campus computers, for example, through the Internet. The teaching system when networked among teachers, parents, and students can provide efficient communications of detailed information regarding a student's studies at school or home. This interactive, bi-directional communications can improve the efficiency and quality of education.
  • One specific embodiment of the invention is an automated teaching process. The process begins with selecting a problem template from among a set of problem template, and randomly generating values for choices identified in the selected problem template. The random choices can be numbers, words, or identifiers for objects such as images that are part of a presentation of the problem. The process additionally presents a sequence of one or more questions that lead to a solution of the problem. The presentation of each question generally depends on the randomly generated values and may depend on the student's previous answers.
  • The student's answers to the questions can be compared to answers derived from the randomly generated values and formulas from the selected problem template. Diagnosis codes that correspond to the answers can identify performance values that are updated to reflect whether the answers from the student were correct. These performance values can indicate weaknesses and control whether the process provides additional teaching or review problems and the frequency of review or the speed of question presentation. In particular, the selected problem template can identify two or more sequence of questions leading to the solution with one sequence being shorter than the other, so that the process can select shorter or longer sequence of questions depending on the evaluation of the student's answers.
  • Another specific embodiment of the invention is an automated teaching process that includes: selecting a problem template from among a set of problem templates; randomly generating values for respective choices identified in the selected problem template; and receiving a selection of either a practice mode or an example mode. In response to the student selecting the practice mode, the process presents a sequence of one or more questions that depend on the randomly generated values and lead to a solution of the problem. In response to the student selecting the example mode, the process presents a sequence of one or more instructions that similarly depend on the randomly generated values and lead to the solution of the problem. The process can enable the student to switch between the practice mode and the example mode, for example, when the student is unsure on how to answer a question.
  • Another specific embodiment of the invention is a teaching system including a problem body table, a practice body table, an example body table, a practice presentation unit, and an example presentation unit. The problem body table contains entries corresponding to problems, and each entry identifies random choices and answers. The practice body table contains entries corresponding to the entries in the problem body table, and each entry in the practice body table includes information representing at least one sequence of questions leading to a solution of the problem associated with the corresponding entry in the problem body table. The example body table contains entries corresponding to the entries in the problem body table, and each entry in the example body table includes information representing a sequence of instructions leading to the solution of the problem associated with the corresponding entry in the problem body table. The practice presentation unit uses information from corresponding entries of the practice body table and the problem body table for presentation of the sequence of questions, and the example presentation unit uses information from corresponding entries of the example body table and the problem body table for presentation of the sequence of instructions.
  • Another specific embodiment of the invention is a teaching system that includes a practice problem presentation system, a user progress table, a diagnosis engine, and a management unit. The practice problem presentation system presents practice problems from a sequence of lessons. The diagnosis engine evaluates answers received from a student during presentation of the practice problems based and then updates the user progress table to indicate which of the lessons are complete. The management unit can access the user progress table and a schedule to determine whether the student is on schedule for progress through the lessons
  • Another specific embodiment of the invention is an automated teaching process that includes: setting a progress clock for completion of a sequence of lessons; presenting lessons to a user at a user controlled rate; comparing the user's progress through the lessons to the progress clock; and determining whether the user is ahead or behind schedule. Depending on whether the user is ahead or behind schedule, the process can then adjust teaching parameters such as the number of practice problems provided per lesson, the difficulty of practice problems, and the number or rate of review problems inserted into lessons.
  • Another specific embodiment of the invention is an automated teaching process including: presenting a current question from a sequence containing multiple questions leading to a solution for the problem; receiving an answer to the current question from a student; and repeating the presenting and receiving steps until the student has reached the solution. The student can control whether consecutive presentations of questions go on to a next question in an order of the sequence. The process can also include highlighting an area of a presentation of a question and presenting an explanation of how content of the area relates to answering the question. In addition or alternatively, the process can include: evaluating each answers to the questions in the sequence; locating diagnosis codes respectively corresponding to the answers; and updating progress categories corresponding to the diagnosis codes to reflect which of the answers where correct.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 shows flow diagram of an educational practice process in accordance with an embodiment of the invention.
  • FIG. 2 shows block diagram of a teaching system in accordance with an embodiment of the invention.
  • FIG. 3 is a block diagram of tables suitable for use in the teaching system of FIG. 2.
  • FIG. 4 is a flow diagram of a teaching process in accordance with an embodiment of the invention that reduces the number of presented steps during practice of a specific type of problem.
  • FIG. 5 is a flow diagram of a teaching process in accordance with an embodiment of the invention where the user selects the logical flow of a multi-step solution.
  • FIG. 6 is a block diagram illustrating the relationship of a teaching insertion unit and a diagnosis engine.
  • FIG. 7 is a block diagram illustrating an embodiment of the invention in which an interactive web site provides a learning system that is accessible to students, teachers, and parents.
  • Use of the same reference symbols in different figures indicates similar or identical items.
  • DETAILED DESCRIPTION
  • A teaching system in accordance with an embodiment of the invention provides explanatory materials, examples, and problems on an interactive basis and adapts to the needs of a student. Additionally, the teaching system can provide detailed analysis or diagnosis of the student's performance that can be communicated to teachers and/or parents for use in other teaching environments.
  • FIG. 1 is a flow diagram illustrating a teaching process 100 used by a student. Teaching process 100 is generally a computer-aided process in which either dedicated hardware or software implements portions of process 100 as described further below. Teaching process 100 is generally adaptable to teaching a variety of subjects at many different educational levels. In exemplary embodiment of the invention, teaching process 100 is directed to primary (K-12) education and the teaching of mathematics, science, or a similar subject in which the solving of problems may be divided into a multi-step logical process. Subjects such as mathematics and science are particularly suited for computer-aided learning because problems in these subjects are often numerically dependent and evaluated.
  • A subject matter and level is selected at the start of process 100. In one implementation of process 100, a person such as a teacher, a parent, or a student in step 110 simply selects a subject, area, and level for process 100. For example, a teacher may give a class assignment that identifies a particular lesson for students. Such lessons can be categorized using a standardized teaching curriculum that educational authorities commonly create. For example, lessons can be categorized using the academic standards of the California State Board of Education, as set forth for the CAT/6 standardized tests for the K-12 students. Each student then selects in step 110 the assigned subject, area, and level. Alternatively, the student may take a placement test in step 120 and an automated assignment in step 125 selects the subject, area, and/or level for the student using teaching process 100.
  • Step 130 generates a random problem based on the selected or assigned subject, area, and level. Selection of the random problem may further be based on the student's prior performance as described further below. In an exemplary embodiment of the invention, the generated problem is randomized first in selection from among problem templates in a class associated with the subject, area, and level of process 100. The selected problem template is further randomized through selection of items required for construction of a complete problem. The selections of items for a particular problem template may include, for example, numeric values that can be selected randomly from a target set or range so that numeric solutions or answers are seldom repeated. Further, items in the content, wording, graphics, or other presentation features associated with the template can be randomly selected from associated lists and provide variety in the random problems that will ideally help hold the students interest in solving practice problems.
  • Each problem template will generally be associated with a multi-step problem solving process. For example, the general problem of adding fractions may be associated with the steps of finding a common denominator from the fractions being added, multiplying the numerator and denominators of the fractions by appropriate factors, adding the resulting numerators of the fractions after multiplication, finding common factors in the numerator and denominator of the result, and simplifying the result accordingly. Associated with each step of a template are an instruction, a question, and one or more formulas for solution of that step. Each answer in that the student provides during the multi-step process may have a diagnosis code identifying a skill. For the example of adding fractions, the diagnosis codes may correspond to skills such as finding common denominators, multiplying numerator and denominator, simplifying fractions. The instructions associated with a template are generally for use when the teaching system presents the random problem as an example of problem solving technique. Process 100 uses the question associated with a step of the problem when the student is solving practice problems.
  • The student in step 140 chooses whether to see an example solution of a problem, attempt to solve a problem, or exit process 100. If the student chooses to see an example solution of the problem, step 150 presents a multi-step solution of the random problem. For each step of the problem, process 100 presents the instruction associated with the step as completed using the randomly selected items and numeric values. Teaching process 100 can thus lead the student step by step through the solution of the randomly generated problem.
  • Process 100 executes a practice routine 160 if the student in step 140 decides to practice solving problems. In the embodiment of FIG. 1, practice routine 160 includes presenting a student with a question associated with a current step of the random problem. The question will typically include the random numbers or items selected in step 130. The question may seek one or more numeric answers or may ask the student select from a number of choices on how to proceed logically in solving the problem posed. The student answers the question in step 164, before the teaching system in step 166 decides whether practice routine 160 is finished. If practice routine 160 for the current random problem is not complete, step 168 selects the next step associated with the problem. Selection of the next step may be based on the student's answer to the last question and/or the student's past performance on similar problems. For example, if the student answered the last question incorrectly, step 168 may ask the student to try again. If the student answered correctly and prior performance indicates the student has mastered the concept associated with the next question, step 168 may choose a question that combines two or more steps of the problem solution.
  • An evaluation routine 170 evaluates the student's performance and can be begun when practice routine 160 is complete or alternatively after each question. Evaluation routine 170 in step 172 identifies and tabulates the errors that the student made during practice routine 160. More specifically, the identified errors can be classified by the concept or concepts associated with the incorrect answer, and a table can record the number of right and wrong answers associated with each concept. Step 174 then determines whether a student's mastery of a particular concept is weak. A large number of errors associated with a particular concept or skill would indicate a weakness in the student's mastery of that concept or skill.
  • If step 174 detects a weakness, a review routine 180 for the detected weakness can be initiated. Review routine 180 may include a concept review step 182 in which the teaching system provides a description or explanation of the concept to the student. One or more solved examples that use the concept can then be presented in step 184, and targeted practice problems are given in step 186. Upon completion of review process 180, teaching process 100 can return to step 130 for generation of another random problem or return to evaluation routine 170 to complete further evaluation of the student's performance. The review process and routine 180 in general can be executed at anytime after a weakness is detected. In particular, routine 180 may be delayed while the students finishes a block of the problems, so that the student is provided with one of more review processes at the end of a lesson, a section, or a chapter.
  • If step 174 of evaluation routine 170 does not detect a particular weakness, step 176 calculates a speed for the practice problems of the evaluated type. For example, if a student masters the concepts or skills being presented by a specific type of practice problem, the number of practice problems using the mastered concepts or skills can be reduced. The practice problems can also be presented using fewer steps, and the calculated speed based on the student's mastery of the concepts or skills may control the number of steps used in the multi-step practice problems.
  • Step 178 calculates a frequency for revisiting previously identified weaknesses. In an exemplary embodiment of the invention, the frequency for the revisiting an identified weakness is proportional to the student's error rate for questions corresponding to the identified weakness. For example, if the error rate is 100% for problems corresponding to the weakness, the revisit frequency can be set at a highest level, e.g., a revisit index 9. If the error rate is 90%, the revisit frequency can be set somewhat lower, e.g., a revisit index 8. If error rate is below a target level (e.g., 10%), the revisit frequency can be set to zero. After each cycle of practice, e.g., each problem or lesson, the revisit index can be incremented, and a review process can be started when the revisit index reaches a target level, e.g., a revisit index 10.
  • Step 190, which in FIG. 1 is executed after evaluation routine 170, determines whether to revisit any of the concepts that have given the student particular difficulties. If step 190 determines it is time for a revisit, the next practice problem for the student will be a random problem that specifically uses the concept targeted for the revisit. The revisited practice problems can be the same type of problems identified and used in concept review routine 180. If step 190, determines it is not yet time to revisit any concepts, teaching process 100 branches back to step 130 to generate another random problem of the type that the student is currently studying.
  • Process 100 as described above can be implemented in hardware or software. Such software may be embodiment as machine-readable instructions in a storage medium such as a floppy disk, a CD ROM, a DVD ROM, or a memory stick to name a few. In one exemplary, embodiment process 100 is implemented in software for a standalone computer, one or more computers connected to a local area network, or computers interconnected through a wide area network or the Internet. For example, software that implements process 100 can be accessible and used through a web site.
  • FIG. 2 illustrates the functional blocks of a system 200 for implementing a teaching process in accordance with an embodiment of the invention. System 200 includes a random problem generator 210, presentation units 212 and 214, mapping tables 220, a diagnostic engine 230, a teaching insertion unit 235, a review insertion unit 240, a revisit frequency adjustment unit 245, a practice speed adjustment unit 250, a report summary unit 260, a progress clock 265, and a management unit 270.
  • In the illustrated embodiment, random problem generator 210 uses mapping tables 220 to generate and present random problems for a specific grade, chapter, section, subject, lesson, or level. As shown in FIG. 3, mapping tables 220 can include a problem body table 310, an example body table 320, a practice body table 330, a diagnosis code table 340, user progress tables 350, and a teaching body table 360. Random problem generator 210 uses the student's level and/or the assigned lesson to generate a pointer that locates related entries in problem body table 310 and then selects one of the entries in problem body table 310. Each entry generally includes a label identifying information such as a grade, chapter, section, subject, lesson, level, example, variation, and type for the problem. The entry in problem body table 310 also includes code that identifies information that is common to both practice problems and example problems. In an exemplary embodiment of the invention, the code within each entry of problem body table 310 identifies a set of random choices rl to rn and a set of answers al to am, where n is the number of choices and m is the number of inputs required for generation and completion of a random problem.
  • Each entry in problem body table 310 generally has one or more corresponding entries in each of example body table 320 and practice body table 330. The entries in example body tables 320 provide information that is specific to presentation of an example problem, and the entries in practice body tables 330 provide information that is specific to presentation of a practice problem.
  • Equation 1 gives a general form for a procedure corresponding to a step, i.e., an instruction or a question that is part of a presentation of an example or practice problem. A single-step example or practice problem may only require one presentation P, but more generally an N-step example or practice problem may include N presentations PI to PN, with each being associated with a corresponding step. In Equation 1, presentation P represents the question or instruction body with a format F for the display, sound, or other media functions of presentation P. Random choices ri and answers ai are used with fixed content of a presentation at locations identified in format F. In particular, format F can indicate the locations of cells on a display screen, and the presentation can fill each cell with fixed content or an object that depends on a random choices ri or answers ai. Some examples of fixed content in the presentation may include fixed sentences, mathematical expressions, a set of available choices, pull down selections, multiple dimension geometric graphics, multiple dimension coordinates, still images, moving pictures, sound, and a mixture of these.
    P=F(r1, r2, . . . , rn, a1, a2, . . . , am)   Equation 1
  • Random choices rl to rn, which are generically referred to herein as choices ri, correspond to variables that may be displayed in the question or instruction body. Each choice ri can represent a number in any format, a set of one or more letters, one or more words, a multiple dimension geometric graphic, a set of coordinates, a still image, a moving image, or a sound. Random problem generator 210 can randomly select each choice ri from a number array, a table, or a call from a file depending on an identifying code for the choice ri. Some choices ri may alternatively be calculated from or dependent upon other choices ri. Choices ri are generally selected once for a problem and may be displayed in the presentations of multiple steps.
  • Answers al to am, which are generically referred to herein as answers ai, are inputs in practice mode and can be determined using formulas from tables 220 in example mode. Each answer ai would normally be received from the user when the random problem is used for practice but alternatively may be calculated from choices ri and associated formulas when the random problem is an example. Format F defines the locations of calculated answers ai in an instruction body or cues for the user's input in a question body. Depending on the context, answers ai can be numbers in any format, letters, words, multiple-choice selections, pull down selections, multiple dimension geometric drawing. In a practice problem, a presentation of a step may provide a text box, a menu, radio buttons, or similar graphic used as a marker or queue that is part of the interface for user assignment of the answer ai. The user can then input answers ai using a keyboard, a mouse button (or mouse movements on a coordinate system defined on the screen), a number or symbol picker, or a voice input. With a symbol picker, the system displays a pad containing items such as number or other symbols adjacent to the marker in response to the student moving a curser to the marker. The student can then use a mouse to select items on the pad and identify the answer being entered.
  • Multi-step practice problem presentation block 212 generally uses a question format F from practice body table 330 and choices ri from random problem generator 210 for presentation of each question associated with a practice problem. Similarly, multi-step example presentation block 214 can use instruction formats from example body table 320 for presentation of the instructions in one or more steps of an example problem.
  • User activation generally controls the multi-step presentation during presentation of an example or a practice problem. As indicated above, a presentation P may include multiple steps that a user activates upon completion of a previous step, and the next step in a presentation can very depending on user's selections or answers in previous steps. For example, if the user provides an answer ai in a prior step, the answer or values derived from the answer may be automatically displayed in presentations of following steps before the user provides additional input. The repeating or use of prior answers ai can make the connection between steps more natural and obvious to the user. In the general case, the user may is not limited to a specific method or order of steps for solving a problem, and the presentation may give the user a choice among more than one way of solving certain problems.
  • The multi-step presentation has several advantages over a single process where a user is simply presented with a problem and provides only the final answer. In particular, the multi-step presentation allows the user to focus on and learn each step needed to solve the problem. The multi-step logic also teaches a logical thought process for solving the problem. Further, multiple answers from the user permits diagnosis or analysis that can accurately locate the source of a user's error.
  • The multi-step operation of presentation units 212 and 214 of FIG. 2 can be modeled into presenting different combinations of steps depending on the practice results from the user. For example, FIG. 4 illustrates an example of a teaching process 400 were the number of steps presented for solving a specific type of practice problem changes according to the user's performance in solving problems having the same type code. In step 410, the user begins practice of a particular type or class of problems, a presentation step 420 presents a full set of steps/questions for the user to answer. A decision step 430 observes the user's answers to determine whether the user has reached a target level of mastery. If the user's performance suggests further practice with the full multi-step presentation, e.g., if the number of errors is above a target level, process 400 branches from step 430 back to step 420. If step 430 recognizes that the user has achieve the target level of mastery, step 440 provides a presentation version having a reduced number of steps, for example, in which some of the initially presented steps are combined. Combining steps in this manner tests whether the user can perform the combined steps without prompting. A decision step 450 determines whether the user has reached a level of mastery suitable for further reduction in the number of presented steps. Step 460 presents a practice problem using the smallest number of steps, e.g., a one-step presentation. Decision step 470 then leaves the user to practice at the speed providing the fewest steps to demonstrate the user's ability to solve the problem without prompting. Process 400 is complete when step 470 identifies that the user has reached the required level of mastery of that specific type of problem. Process 400 can then be used for problems of a new type.
  • Practice problem generation unit 212 can also use the multi-step concept to track a user's preferred logical process for solving problems. FIG. 5, for example, illustrates a process 500 in which a user selects the steps for solving a practice problem. Process 500 begins with a step 510 presenting the user with options for an initial step in the solution of the problem. In FIG. 5, three possible selections are available, but more generally any number of alternatives may be provided depending on the problem. However, if only one selection is available, step 510 may be skipped. Decision step 520 then directs process 500 to a presentation 530X, 530Y, or 530Z of a question according to the user's selection, and the user in step 540 provides answers to the presented question. Decision step 550 then determines whether the solution of the problem is complete, and if not branches back to step 510. Each question presentation in step 510 will in general provide a different number of choices with the choices being dependent on the user's prior selections.
  • Another feature of teaching system 200 of FIG. 2 is the ability of the user to switch from a practice mode to an example mode. In particular, if the user unable to answer a question from practice problem presentation unit 212, the user can switch to example mode. Example presentation unit 214 can then present an instruction for the current step (or the example problem from its beginning) since both practice presentation unit 212 and example presentation unit 214 use the same random choices ri. The user can also be allowed to step backward in a multi-step example or practice problem if desired.
  • Practice presentation unit 212 or example presentation unit 214 can also provide an alternative help mode by displaying explanatory texts, graphs and/or voice that will explain the calculation, logic, or reasoning for a particular step or the relation of a particular choice ri or answer ai to the desired solution of the problem. The dynamic position corresponding to the feature of a presented question can be activated by highlighting or otherwise altering the appearance, for example, of certain choices ri and answers ai. In particular, the choices ri and answers ai that are the elements involved in explanatory texts, graphs, and/or voice can be similarly altered or highlighted, and the explanations can be displayed over or near the feature explained.
  • System 200 includes an architecture that has mapping tables 220 that can be linked to each other as shown in FIG. 3. In an exemplary embodiment of the invention, problem body table 310 contains executable code for generating portions of the problem that apply both to practice and example problems. The code in problem body table 310 can be linked and called out by random problem generator 210, for example, for or during selection of random choices ri or as part of the presentation of a problem. As noted above, each entry in problem body table 310 also identifies random choices ri and answers ai for a corresponding problem. The example body table 320 similarly contains code for the example body, and example presentation unit 214 can link to and called out the code from table 320. Each entry of example body table 320 includes a mapping or link to a corresponding entry of problem body table 310, and each entry of problem body table 310 includes a mapping or link back to a corresponding entry of example body table 320.
  • Practice body table 330 similarly contains the practice problem bodies that practice problem presentation unit 212 can call out for presentation of practice problems. Entries of example body table 330 include mappings or links to corresponding entries of problem body table 310 and/or example body table 320 and visa versa for use of information and/or executable code from tables 310 and 330 when system 200 operates in practice mode. With this architecture, entries from tables 310 and 320 can be called out bi-directionally for use of information or code from example body table 320 and/or practice body table 330, for example, when switching between a practice mode where practice problems are presented to an example mode where example problems are presented.
  • Diagnosis engine block 230 can access practice body table 330, diagnosis code tables 340, and user progress tables 350 when evaluating which answers ai are correct and tabulating the right and wrong answers in specific skill areas. Diagnosis code tables 340 are a group of tables contain diagnosis codes used by diagnosis engine 230. Each of the diagnosis code tables 340 is linked to a corresponding entry in problem body table 310 and therefore linked to a specific problem. The entries in each of the diagnosis code tables 340 correspond to the answers ai associated with the corresponding problem, and each entry is a diagnosis code that identifies a subject or skill associated with answer ai. When evaluating a user's response during a practice problem, diagnosis engine 340 as described further below identifies whether an answer ai is correct or incorrect, accesses an entry corresponding to the answer ai in the diagnosis code table associated with the evaluated problem, and tabulates the correct or incorrect answer in a corresponding entry of user progress table 350.
  • User progress tables 350 are used to indicate the user/progress status information such as: a current level, a current work on point, an overall grade, detail grades, an overall diagnosis result, detail diagnosis results, concepts or lessons needing review, problems that need revisiting, and speed up/speed down parameters. Generally, the status entries in tables 350 are linked to corresponding the entries in tables 310, 330, and 340.
  • Diagnosis engine 230 evaluates answers ai to practice problems to identify specific weaknesses in problem solving skills and/or overall weaknesses in completing a class of problems. In an exemplary embodiment of the invention, diagnosis engine 230 contains logic and equations that are linked with diagnosis code tables 340 for performing the diagnostic functions. Four primary diagnostic functions of diagnosis engine 230 are real time weakness detection, overall weaknesses analysis, detection of subjects to be revisited, and adjustment of the learning speed during practice problem presentations.
  • For real time weakness detection, diagnosis engine 230 activates error checking as soon as the user finished/answered the question or problem. The error checking compares the user's answers to calculated answers. The answers can be calculated using the random choices ri for the problem and the formulae from problem body table 310. Diagnosis code tables 340 can also assign each answer ai to a specific problem type and/or to a specific skill, and diagnosis engine 230 can then update user progress tables 350 to reflect correct and incorrect answers for the identified skills or problem types. In an exemplary embodiment, diagnosis engine 230 tabulates every answer ai as a right or wrong answer for the corresponding category or categories and for each incorrect answer gives the correct answer with example calculations or logical reasons for why or how the correct answer is found. Diagnosis engine 230 can also activate teaching insertion unit 235, which can then provide a fixed presentation to explain the concepts or techniques that diagnosis engine 230 identified as needing to be improved. Diagnosis engine 230 preferably presents the correct answers, the reasoning needed to arrive at the correct answer, and the list of the areas or skills requiring study on or near the user's error on a work sheet that practice presentation unit 212 provided.
  • When the user finishes a section of practice problems, for example corresponding to a lesson, unit, chapter, grade, . . . etc), diagnosis engine 230 performs an overall weaknesses analysis based on error accumulations in user progress tables 350. Such analysis can grade the overall correct/level percentage for this section of practice problems and grade each concept, technique, or skill used in the section of practices problems. As precursor to the overall grading or analysis, diagnosis engine 230 can calculate a percentage of right answers or a weighted average CRT for each diagnosis code ci as indicated in Equation 2. In Equation 2, n is the total number of evaluated practice problems having diagnosis code ci, j is an index identifying the particular practice problems, ARij is 1 if answer ai in problem j is correct and is 0 otherwise, and kj is a weight assigned to each problem j. Weights kj for calculation of the weighted average can be from a fixed table (not shown) and can generally have values that depend on the grade level and difficulty of problem j. The percentage of right answers or CRT can then be used in creating a list the concepts or techniques that may need further study after the completed practice section. Equation 2 : CRT ci = j = 1 n k j AR ij j = 1 n k j
  • A teaching insertion unit 235 can be called to present explanations of the identified concepts or skills having a low percentage of right answers. FIG. 6 illustrates the relationship of teaching insertion block 235 to diagnosis engine 230 and practice problem presentation unit 212 in an exemplary embodiment of the invention. As illustrated in FIG. 6, diagnosis engine 230 uses a diagnosis code corresponding to a question or skill to identify an entry in teaching body table 360. The identified entry, which corresponds to a diagnosis code identified for teaching insertion, can contain explanations of the concepts or skills corresponding to the diagnosis code and a detailed explanation of how to solve a corresponding problem. Teaching insertion unit 235 includes a routine 610 for presenting a review of the required concepts or skills and a routine 620 for presenting the detailed explanation of how to solve the problem. Teaching insertion unit 235 can provide a fixed detailed explanation in greater detail than does the example problems that example presentation unit 214 presents or alternatively may use a random example problem that example presentation unit 214 presents. Teaching insertion unit 235 can also activate practice problem presentation unit 212 to present a set of random problems corresponding to the identified diagnosis code. Diagnosis engine 230 can then evaluate this set of random problems to monitor the user's improvement.
  • Diagnosis engine 230 can also activate speed adjustment unit 250 to adjust the speeds for specific types of problems based on the overall error percentages. User progress tables 350 can also be updated as part of the overall evaluation of a section to show completion of the section and overall grades or evaluations of the section. Diagnosis engine 230 can thus make the overall grades/evaluations as well as the specific percentages of right/wrong answers for specific questions or skills available to report summary unit 260 for generation of reports for teachers, parents, or other users.
  • The practice revisit detection function of diagnosis engine 230 can also be initiated when the user finishes a section (e.g., a lesson, unit, chapter, or grade) of practices problems. The revisit detection generally includes an overall analysis based on the percentage of correct answers for a particular problem type. If the percentage of correct answers for a particular type of problem or skill is below the fixed or floating threshold, that type of problem or skill can be target for review and/or revisiting. Revisit frequency adjustment unit 245 can then set a frequency for revisiting the identified problem type or skill, so that review insertion unit 240 will be activated to present the user with targeted review problems. In an exemplary embodiment of the invention, revisit frequency adjustment unit 245 updates user progress table 350 to list each diagnosis code that diagnosis engine 230 identifies for a revisit and to provide a revisit frequency that can be calculated from the percentage of correct answers associated with the diagnosis code.
  • Review insertion unit 240 under direction of management unit 270 can start a revisit function that inserts a practice problem of the type or types that diagnostic engine 230 identified for revisiting. Each revisited problem is generally inserted among a current series of practice problems corresponding to a subsequent lesson, for example, when a student is practicing a lesson following the lesson in which a weakness was detected. For the revisit function, management unit 270 can set a progress break that allows review insertion unit 240 to perform the revisit function at the frequency set by revisit frequency adjustment block 245.
  • In an exemplary embodiment of the invention, revisit frequency adjustment block 245 determines a revisit frequency RFm for each diagnosis code m that diagnosis engine 230 identifies for revisiting. Equation 3 provides an example of a suitable formula for revisit frequency RFm. In Equation 3, CRVm is the weighted percentage of correct answers for diagnosis code m, TCRV is threshold corresponding to a pass or fail level and is greater than weighted percentage CRVm since diagnosis code m is selected for revisiting, and sm is a proportional adaptive parameter for diagnosis code m. Parameter sm can be stored in a user progress table 350 and may depend on the student's over performance, the grade level, the problem difficulty, and the error importance. Parameter sm can thus scale the revisit frequency depending on the significance of the diagnosis code m. For example, parameter sm for a problem type m that is relatively unimportant for the student's current grade level may be small to reduce the revisit frequency for the less important problems.
    RF m =s m(T CRV −CRV m)   Equation 3
  • Diagnosis engine 230 also provides the needed input for learning speed adjustment. The learning speed adjustment can be based on the student's recent performance in completing problems. For this purpose, diagnosis engine 230 can run on the overall analysis to get weighted percentage CRVci of correct answers for diagnosis code ci during the most recent practice session. Equation 4 gives a suitable formula for weighted percentage CRV. In Equation 4, np is the total number of evaluated practice problems with diagnosis code ci in the last practice section, j is an index identifying the particular practice problems, PRij is 1 if answer ai in problem j is correct and is 0 otherwise, and qj is a weight assigned to each problem j. Weightings qj may be the same as weightings kj or alternatively may differ to provide an overall evaluation specific to the current section. If the weighted average CRVci of correct answers is above a fixed or variable threshold, the practice speed for the type of problem corresponding to diagnosis code ci can be increased. Equation 4 : CRV ci = j = 1 np q j PR ij j = 1 np q j
  • In an exemplary embodiment of the invention, the practice speed is set for all problems in the next session of practice according to an average of the percentage of correct answers for practice problems in the last practice session. Equation 5 gives a formula for a learning speed parameter LSP the weight percentage CRVm correct answers, the number nd of diagnosis codes associated with the last practice session, and weighting factors wm for the diagnosis codes. Weighting factors wm can indicate an importance factor for each diagnosis codes. For example, if a diagnosis code i corresponds to important subject matter, then weighting factor wi can be set to 1, otherwise weighting factor wi could be less than 1. Practice problem presentation unit 212 can use the learning speed parameter LSP to select the number of step presented for a practice problem. Equation 5 : LSP = j = 1 nd w j CRV j j = 1 nd w j
  • Progress clock 265 in an exemplary embodiment of the invention controls or selects the main lessons that system 200 of FIG. 2 presents. Unless the user directly selects particular subject matter, system 200 can present lessons sequentially in a linear or non- linear manner as desired for completion of series of lessons within a selected time span. Additionally, as noted above, revisited problems can be inserted into the series of lessons to review concepts that were originally more difficult for the student. The time span for a series of lessons could be a calendar or school year, semester, a quarter, a month, or any desired time period set for learning the subject matter contained in the series of lessons. Typically, progress clock 265 will be synchronized to the present calendar time movement.
  • One purpose of progress clock 265 can be for comparison of the student's progress as noted in user progress tables 350 to a schedule for completing the lessons in the allotted time. Progress clock can also provide a time parameter to management unit 270. Management unit 270 can then perform functions such as sending a warning signal to the user stating his/her progress is behind the schedule and then intensify the learning/practice efforts to let the user catch up. Alternatively, management unit 270 can send a congratulation signal to the user indicating that the student's progress is ahead of schedule. Central management unit 270 or the user can then choose to do more difficult practice problems, do extra learning/practice for any previously detected weaknesses, or make up some learning/practice/subject which was skipped, for example, due to a late sign-up for use of system 200.
  • Management unit 270 more generally has interfaces to all of the other units in system 200 and can generally control operation of system 200, for example, to implement the teaching process of FIG. 1.
  • Much of the above concentrated on the teaching functions of system 200 and emphasized operation where the user is the student. System 200 can also be used as an evaluation tool reporting the student's progress to teachers or parents. Report summary unit 260 processes the data stored in user progress tables 350 to generate reports. Such reports can be provided in a variety of formats such as raw data, a table, or a graph and may include a user-selected combination of information about the student's progress. Some of the information about a student's progress that may be included in a report includes: the student's current level, the current lesson, overall performance, detail performance categorized by any or all of the subjects having diagnosis codes, related performance indicating the student's strengths and needs, overall diagnosis result listing the student's overall weaknesses, detail diagnosis result listing specifics or details regarding the weaknesses, concept or lessons that the student needs to review, questions/problems that need revisiting, the speed up/speed down parameter, the student's overall or detailed progress in comparison to the master calendar, the time efficiency of the student's overall progress, and the students' time efficiency for recent progress. Such reports can be printed or transmitted electronically via a network such as a local area network (LAN) or a wide area network such as the Internet to a computer used by a teacher.
  • The teaching system and process described above can be implemented and used in a variety of environments or situations. For example, in one embodiment, teaching system 200 is implemented on a standalone computer that the student uses for at home or at school.
  • A group of students working on separate computers at a central location such as a school or at separate locations can use copies of system 200 as described above. An advantage of the random problem generation described above is that each student in a class can be provided with different problems (e.g., with different numeric answers) that teach the same subject. This makes cheating or sharing of answers more difficult and does not increase the difficulty of grading or evaluating the students' answers. A teacher with access to the network can monitor the performance and progress of all of the students and quickly have student evaluations that would take a great deal of time to generate by hand.
  • FIG. 7 illustrates an exemplary embodiment of a learning system 700 in which a web site 710 provides a teaching process or system similar to those described above in regard to FIGS. 1 and 2. Students 720 use web site 700 by logging in and viewing examples or solving practice problems. A teacher 730 can similarly log on to web site 700 to communicate with students 710 and/or parents 740, to assign lessons to students 710, or to access reports on any of the students 710 that web site 710 identifies as being in the class of teacher 730. Parents or guardians 740 can also log on to web site 710 and have access to information from teacher 730 or reports on the performance or needs of their child 710.
  • Although the invention has been described with reference to particular embodiments, the description is only an example of the invention's application and should not be taken as a limitation. Various adaptations and combinations of features of the embodiments disclosed are within the scope of the invention as defined by the following claims.

Claims (27)

1. An automated teaching process comprising:
selecting a problem template from among a set of problem templates;
randomly generating values for respective choices identified in the selected problem template;
presenting a sequence of one or more questions to a student, wherein the question presented depends on the randomly generated values;
receiving answers from the student; and
comparing the answers from the student to answers derived from the randomly generated values and formulas from the selected problem template.
2. The process of claim 1, wherein at least one of the random values is a numeric value.
3. The process of claim 1, wherein at least one of the random values identifies an image to be displayed when presenting the question.
4. The process of claim 1, wherein the selected template identifies locations for presentation of the randomly selected values in the question.
5. The process of claim 1, wherein the selected template provides locations for a plurality of cells, wherein items identified by the randomly generated values are located in respective cells when presenting a question.
6. The process of claim 1, further comprising presenting a sequence of questions that the selected problem template identifies as leading to a final solution.
7. The process of claim 6, wherein content of presentation of one of the sequence of question contains content that depends on one or more of the answers from the student.
8. The process of claim 1, wherein receiving answers from the student comprises:
displaying a marker indicating a location for an answer;
displaying a pad in adjacent to the marker in response to the student moving a curser to the marker; and
identifying the answer based on the student selecting items on the pad.
9. The process of claim 1, wherein the selected problem template identifies a first sequence of questions leading to a final solution and a second sequence of questions leading to the final solution, the second sequence including fewer questions that are in the first sequence, wherein the process further comprises evaluating answers the student gave for prior problems to determine whether to present the first or second sequence of questions.
10. The process of claim 1, further comprising:
identifying a diagnosis code that corresponds to at least one of the answers from the student; and
updating a performance value corresponding to the diagnosis code to reflect whether the answer from the student was correct.
11. The process of claim 10, further comprising:
using the performance value to determine whether review of a concept associated with the diagnosis code requires review; and
presenting an explanation of the concept associated with the diagnosis code.
12. The process of claim 10, further comprising:
using the performance value to determine whether a concept associated with the diagnosis code should be revisited; and
inserting a practice problem corresponding to the diagnosis code among a subsequent series of practice problems addressing concepts other than the concept associated with the diagnosis code.
13. The process of claim 12, further comprising using the performance value to determine a frequency for inserting practice problems corresponding to the diagnosis code among the subsequent series of practice problems.
14. The process of claim 10, further comprising using the performance value to determine a speed for presentation of subsequent practice problems.
15. An automated teaching process comprising:
selecting a problem template from among a set of problem templates;
randomly generating values for respective choices identified in the selected problem template;
receiving a selection of either a first mode or a second mode;
in response to the student selecting the first mode, presenting a sequence of one or more questions, wherein the questions presented depend on the randomly generated values and lead to a final solution; and
in response to the student selecting the second mode, presenting a sequence of one or more instructions, wherein the instructions presented depend on the randomly generated values and lead to the final solution.
16. The process of claim 15, wherein receiving the selection of the second mode occurs after presentation of at least one question from the sequence of one or more questions, and wherein presenting the sequence of one or more instructions comprises presenting an instruction that indicates how to answer a corresponding one of the presented questions.
17. The process of claim 15, further comprising switching between stepping forward through the sequence being presented and stepping backward through the sequence being presented in response to a user command.
18. A teaching system comprising:
a problem body table containing a plurality of entries, each entry corresponding to a problem and identifying a set of random choices and a set of answers;
a practice body table containing a plurality of entries, each entry of the practice body table corresponding to one of the entries in the problem body table and including information representing a sequence of questions leading to a solution of the problem associated with the corresponding entry in the problem body table;
an example body table containing a plurality of entries, each entry of the example body table corresponding to one of the entries in the problem body table and including information representing a sequence of instructions leading to the solution of the problem associated with the corresponding entry in the problem body table;
a practice presentation unit that uses information from corresponding entries of the practice body table and the problem body table for presentation of the sequence of questions; and
an example presentation unit that uses information from corresponding entries of the example body table and the problem body table for presentation of the sequence of instructions.
19. A teaching system comprising:
a practice problem presentation system that presents practice problems from a sequence of lessons;
a user progress table;
a diagnosis engine that evaluates answers received from a student during presentation of the practice problems based, the diagnosis engine updating the user progress table to indicate which of the lessons are complete; and
a management unit capable of accessing the user progress table and a schedule to determine whether the student is on schedule for progress through the lessons
20. An automated teaching process comprising:
setting a progress clock for completion of a sequence of lessons;
presenting lessons to a user at a user controlled rate;
comparing user progress through the lessons to the progress clock; and
determining whether the user is ahead or behind schedule.
21. The process of claim 20, further comprising adjusting a number of practice problems provided per lesson according to whether the user is ahead or behind schedule.
22. The process of claim 20, further comprising adjusting the difficulty of practice problems according to whether the user is ahead or behind schedule.
23. The process of claim 20, further comprising inserting review problems into lessons at a rate that depends whether the user is ahead or behind schedule.
24. An automated teaching process comprising:
selecting a problem;
presenting a current question from a sequence to a student, wherein the sequence contains multiple questions leading to a solution for the problem;
receiving an answer to the current question from the student; and
repeating the presenting and receiving steps until the student has reached the solution.
25. The process of claim 24, wherein the student controls whether consecutive repetitions of presenting the questions go on to a next question in an order of the sequence.
26. The process of claim 24, comprising:
highlighting an area of a presentation of a question; and
presenting an explanation of how contents of the area relates to answering the question.
27. The process of claim 24, further comprising:
evaluating each answers to the questions in the sequence;
locating diagnosis codes respectively corresponding to the answers; and
updating progress categories corresponding to the diagnosis codes to reflect which of the answers where correct.
US10/982,439 2004-11-05 2004-11-05 Computerized teaching, practice, and diagnosis system Abandoned US20060099563A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US10/982,439 US20060099563A1 (en) 2004-11-05 2004-11-05 Computerized teaching, practice, and diagnosis system
TW094112864A TWI270023B (en) 2004-11-05 2005-04-22 Computerized teaching, practice, and diagnosis system
CNA2005100783943A CN1770222A (en) 2004-11-05 2005-06-20 Computerized teaching, practice, and diagnosis system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US10/982,439 US20060099563A1 (en) 2004-11-05 2004-11-05 Computerized teaching, practice, and diagnosis system

Publications (1)

Publication Number Publication Date
US20060099563A1 true US20060099563A1 (en) 2006-05-11

Family

ID=36316741

Family Applications (1)

Application Number Title Priority Date Filing Date
US10/982,439 Abandoned US20060099563A1 (en) 2004-11-05 2004-11-05 Computerized teaching, practice, and diagnosis system

Country Status (3)

Country Link
US (1) US20060099563A1 (en)
CN (1) CN1770222A (en)
TW (1) TWI270023B (en)

Cited By (42)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060223040A1 (en) * 2005-03-30 2006-10-05 Edward Brown Interactive computer-assisted method of instruction and system for implementation
US20060234200A1 (en) * 2005-04-18 2006-10-19 Alain Rochedy Computer based method for self-learning and auto-certification
US20060286538A1 (en) * 2005-06-20 2006-12-21 Scalone Alan R Interactive distributed processing learning system and method
US20080038708A1 (en) * 2006-07-14 2008-02-14 Slivka Benjamin W System and method for adapting lessons to student needs
US20080038705A1 (en) * 2006-07-14 2008-02-14 Kerns Daniel R System and method for assessing student progress and delivering appropriate content
US20080057480A1 (en) * 2006-09-01 2008-03-06 K12 Inc. Multimedia system and method for teaching basal math and science
US20080090221A1 (en) * 2006-10-11 2008-04-17 Ashby Cynthia Elnora Interactive method for diagnosing test-taking errors
US20080113327A1 (en) * 2006-11-10 2008-05-15 Microsoft Corporation Interactive system for teaching and learning algorithms through discovery
US20080166686A1 (en) * 2007-01-04 2008-07-10 Cristopher Cook Dashboard for monitoring a child's interaction with a network-based educational system
US20080254437A1 (en) * 2005-07-15 2008-10-16 Neil T Heffernan Global Computer Network Tutoring System
US20090012789A1 (en) * 2006-10-18 2009-01-08 Teresa Ruth Gaudet Method and process for performing category-based analysis, evaluation, and prescriptive practice creation upon stenographically written and voice-written text files
WO2009049353A1 (en) * 2007-10-15 2009-04-23 Vocam Pty Ltd An apparatus for education and assessment rather than and apparatus for education and assessment
WO2009063469A2 (en) * 2007-11-13 2009-05-22 Blossom E-Learning Ltd. Adaptive electronic learning system and method
US20090253114A1 (en) * 2008-04-02 2009-10-08 Sinapse Print Simulators Automatic trace analysis and comparison system for interactive learning and training systems
US20090325140A1 (en) * 2008-06-30 2009-12-31 Lou Gray Method and system to adapt computer-based instruction based on heuristics
WO2010002395A1 (en) * 2008-06-30 2010-01-07 Dreambox Learning Inc. Method and system to adapt computer-based instruction based on heuristics
US20100068687A1 (en) * 2008-03-18 2010-03-18 Jones International, Ltd. Assessment-driven cognition system
US20100209896A1 (en) * 2009-01-22 2010-08-19 Mickelle Weary Virtual manipulatives to facilitate learning
US20100279265A1 (en) * 2007-10-31 2010-11-04 Worcester Polytechnic Institute Computer Method and System for Increasing the Quality of Student Learning
US20100285441A1 (en) * 2007-03-28 2010-11-11 Hefferman Neil T Global Computer Network Self-Tutoring System
US20110033834A1 (en) * 2007-12-28 2011-02-10 Peter Krumhauer Eteacher - electronic teacher for conveying complex content for learning
US20110076654A1 (en) * 2009-09-30 2011-03-31 Green Nigel J Methods and systems to generate personalised e-content
US20110097698A1 (en) * 2009-10-27 2011-04-28 Henderson Charles A System and method for interactive learning
US20110143328A1 (en) * 2009-12-14 2011-06-16 Gerald Alfred Brusher Method and Apparatus for Enhancing an Academic Environment
US20110195389A1 (en) * 2010-02-08 2011-08-11 Xerox Corporation System and method for tracking progression through an educational curriculum
US20120045744A1 (en) * 2010-08-23 2012-02-23 Daniel Nickolai Collaborative University Placement Exam
US20130177892A1 (en) * 2007-02-21 2013-07-11 University Of Central Florida Research Foundation, Inc. Computing device providing electronic book data having selectable content layers with different difficultly levels and related methods
US20130260351A1 (en) * 2012-03-29 2013-10-03 Dreambox Learning Inc. Calendar-driven sequencing of academic lessons
US20130260359A1 (en) * 2010-10-29 2013-10-03 Sk Telecom Co., Ltd. Apparatus and method for diagnosing learning ability
US20130266916A1 (en) * 2011-11-23 2013-10-10 Rodney A. Weems Systems and Methods Using Mathematical Reasoning Blocks
US8606170B2 (en) 2012-01-20 2013-12-10 Northrop Grumman Systems Corporation Method and apparatus for interactive, computer-based, automatically adaptable learning
US8696365B1 (en) * 2012-05-18 2014-04-15 Align, Assess, Achieve, LLC System for defining, tracking, and analyzing student growth over time
US20140335499A1 (en) * 2001-05-07 2014-11-13 Frank R. Miele Method and apparatus for evaluating educational performance
US20150111191A1 (en) * 2012-02-20 2015-04-23 Knowre Korea Inc. Method, system, and computer-readable recording medium for providing education service based on knowledge units
US20160225273A1 (en) * 2015-01-29 2016-08-04 Xerox Corporation Methods and systems for evaluating user
WO2017011910A1 (en) * 2015-07-21 2017-01-26 Varafy Corporation Method and system for templated content generation and assessment
US9858828B1 (en) * 2013-03-15 2018-01-02 Querium Corporation Expert systems and methods for dynamic assessment and assessment authoring
CN109559580A (en) * 2018-12-29 2019-04-02 武汉易测云网络科技有限公司 A kind of on-line study system
US20210383710A1 (en) * 2006-10-11 2021-12-09 Cynthia Elnora ASHBY Interactive system and method for diagnosing test-taking errors based on blooms taxonomy
US11386798B2 (en) * 2017-12-13 2022-07-12 Caveon, Llc Systems and methods for testing skills capability using technologically-enhanced questions in a computerized environment
US20220293006A1 (en) * 2021-03-15 2022-09-15 Casio Computer Co., Ltd. Information processing device, controlling method and recording medium
US11600196B2 (en) * 2017-03-13 2023-03-07 Vitruv Inc. Method and system for supporting learning, and non-transitory computer-readable recording medium

Families Citing this family (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI452547B (en) * 2008-05-06 2014-09-11 Daka Studio Inc A rating learning system and method of the same
TWI393088B (en) * 2009-01-23 2013-04-11 Jung Kuan Kao English level positioning method
TWI385610B (en) * 2009-08-06 2013-02-11 Univ Ishou Two - way Evaluation Method of Course Teaching and Its System and Computer - readable Recording Media
US20120122067A1 (en) * 2010-11-15 2012-05-17 Age Of Learning, Inc. Online educational system with mentor guided learning
US9324240B2 (en) 2010-12-08 2016-04-26 Age Of Learning, Inc. Vertically integrated mobile educational system
WO2013029242A1 (en) * 2011-08-30 2013-03-07 西门子公司 Medical training method and system thereof
CN103218927B (en) * 2012-01-18 2015-08-19 财团法人中文数位化技术推广基金会 Interactive learning monitoring, diagnosis and recommendation system and method
JP5803861B2 (en) * 2012-03-19 2015-11-04 カシオ計算機株式会社 Learning apparatus and learning program
CN102663659A (en) * 2012-03-27 2012-09-12 上海爱友科技有限公司 Education system based on academic achievement development index
TWI569222B (en) * 2012-07-27 2017-02-01 Tian Xin Learning to help prescribe the teaching system
WO2014134278A1 (en) * 2013-02-27 2014-09-04 Janua Educational Services, LLC Underlying student test error detection system and method
JP5864028B2 (en) * 2013-03-29 2016-02-17 三菱電機株式会社 Information processing apparatus and information processing system
CN105810028A (en) * 2015-12-22 2016-07-27 宁波睿易教育科技股份有限公司 Electronic learning and examination system mixed with multiple answering modes
CN105679108A (en) * 2016-01-04 2016-06-15 广东小天才科技有限公司 Calculation method and system for mathematical vertical operation
US10930169B2 (en) * 2017-05-04 2021-02-23 International Business Machines Corporation Computationally derived assessment in childhood education systems
CN111782720A (en) * 2020-06-05 2020-10-16 上海高绩数据科技有限公司 Graphical method, system and terminal for rapidly detecting and positioning transaction effect
CN111785104B (en) * 2020-07-16 2022-03-04 北京字节跳动网络技术有限公司 Information processing method and device and electronic equipment

Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5797753A (en) * 1995-03-22 1998-08-25 William M. Bancroft Method and system for computerized learning response, and evaluation
US5827066A (en) * 1995-03-10 1998-10-27 Henter; Ted Methods of teaching mathematics to disabled students
US5863208A (en) * 1996-07-02 1999-01-26 Ho; Chi Fai Learning system and method based on review
US5934909A (en) * 1996-03-19 1999-08-10 Ho; Chi Fai Methods and apparatus to assess and enhance a student's understanding in a subject
US6058385A (en) * 1988-05-20 2000-05-02 Koza; John R. Simultaneous evolution of the architecture of a multi-part program while solving a problem using architecture altering operations
US6112051A (en) * 1996-11-22 2000-08-29 Fogcutter, Llc Random problem generator
US6259890B1 (en) * 1997-03-27 2001-07-10 Educational Testing Service System and method for computer based test creation
US20020102522A1 (en) * 2001-01-31 2002-08-01 Koichi Sugimoto Method and system for perfoming adaptive test
US6431875B1 (en) * 1999-08-12 2002-08-13 Test And Evaluation Software Technologies Method for developing and administering tests over a network
US20020127533A1 (en) * 2000-09-21 2002-09-12 Grant Charles Alexander Method and apparatus for delivery of educational content
US20020197594A1 (en) * 2000-08-08 2002-12-26 Dickmeyer Jon Steven Systems and methods for searching for and delivering solutions to specific problems and problem types
US20030203342A1 (en) * 2001-11-13 2003-10-30 Prometric, Inc. Method and system for computer based testing using customizable templates
US20040018479A1 (en) * 2001-12-21 2004-01-29 Pritchard David E. Computer implemented tutoring system
US20040191746A1 (en) * 2003-03-27 2004-09-30 Mel Maron Process for computerized grading of formula-based multi-step problems via a web-interface
US20040234936A1 (en) * 2003-05-22 2004-11-25 Ullman Jeffrey D. System and method for generating and providing educational exercises
US6990519B2 (en) * 2001-11-08 2006-01-24 Texas Instruments Incorporated Use of a directed acyclic organization structure for selection and execution of consistent subsets of rewrite rules

Patent Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6058385A (en) * 1988-05-20 2000-05-02 Koza; John R. Simultaneous evolution of the architecture of a multi-part program while solving a problem using architecture altering operations
US5827066A (en) * 1995-03-10 1998-10-27 Henter; Ted Methods of teaching mathematics to disabled students
US5797753A (en) * 1995-03-22 1998-08-25 William M. Bancroft Method and system for computerized learning response, and evaluation
US5934909A (en) * 1996-03-19 1999-08-10 Ho; Chi Fai Methods and apparatus to assess and enhance a student's understanding in a subject
US5863208A (en) * 1996-07-02 1999-01-26 Ho; Chi Fai Learning system and method based on review
US6112051A (en) * 1996-11-22 2000-08-29 Fogcutter, Llc Random problem generator
US6259890B1 (en) * 1997-03-27 2001-07-10 Educational Testing Service System and method for computer based test creation
US6431875B1 (en) * 1999-08-12 2002-08-13 Test And Evaluation Software Technologies Method for developing and administering tests over a network
US20020197594A1 (en) * 2000-08-08 2002-12-26 Dickmeyer Jon Steven Systems and methods for searching for and delivering solutions to specific problems and problem types
US20020127533A1 (en) * 2000-09-21 2002-09-12 Grant Charles Alexander Method and apparatus for delivery of educational content
US20020102522A1 (en) * 2001-01-31 2002-08-01 Koichi Sugimoto Method and system for perfoming adaptive test
US6990519B2 (en) * 2001-11-08 2006-01-24 Texas Instruments Incorporated Use of a directed acyclic organization structure for selection and execution of consistent subsets of rewrite rules
US20030203342A1 (en) * 2001-11-13 2003-10-30 Prometric, Inc. Method and system for computer based testing using customizable templates
US20040018479A1 (en) * 2001-12-21 2004-01-29 Pritchard David E. Computer implemented tutoring system
US20040191746A1 (en) * 2003-03-27 2004-09-30 Mel Maron Process for computerized grading of formula-based multi-step problems via a web-interface
US20040234936A1 (en) * 2003-05-22 2004-11-25 Ullman Jeffrey D. System and method for generating and providing educational exercises

Cited By (63)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140335499A1 (en) * 2001-05-07 2014-11-13 Frank R. Miele Method and apparatus for evaluating educational performance
US20060223040A1 (en) * 2005-03-30 2006-10-05 Edward Brown Interactive computer-assisted method of instruction and system for implementation
US20060234200A1 (en) * 2005-04-18 2006-10-19 Alain Rochedy Computer based method for self-learning and auto-certification
US20060286538A1 (en) * 2005-06-20 2006-12-21 Scalone Alan R Interactive distributed processing learning system and method
US20080254437A1 (en) * 2005-07-15 2008-10-16 Neil T Heffernan Global Computer Network Tutoring System
US20080038705A1 (en) * 2006-07-14 2008-02-14 Kerns Daniel R System and method for assessing student progress and delivering appropriate content
US11462119B2 (en) * 2006-07-14 2022-10-04 Dreambox Learning, Inc. System and methods for adapting lessons to student needs
US10347148B2 (en) 2006-07-14 2019-07-09 Dreambox Learning, Inc. System and method for adapting lessons to student needs
US20080038708A1 (en) * 2006-07-14 2008-02-14 Slivka Benjamin W System and method for adapting lessons to student needs
US20080057480A1 (en) * 2006-09-01 2008-03-06 K12 Inc. Multimedia system and method for teaching basal math and science
US20080090221A1 (en) * 2006-10-11 2008-04-17 Ashby Cynthia Elnora Interactive method for diagnosing test-taking errors
US10037707B2 (en) * 2006-10-11 2018-07-31 Cynthia Elnora ASHBY Interactive method for diagnosing test-taking errors
US11741848B2 (en) * 2006-10-11 2023-08-29 Cynthia Elnora ASHBY Interactive system and method for diagnosing test-taking errors based on blooms taxonomy
US11037458B2 (en) * 2006-10-11 2021-06-15 Cynthia Elnora ASHBY Interactive method for diagnosing test-taking errors
US20210383710A1 (en) * 2006-10-11 2021-12-09 Cynthia Elnora ASHBY Interactive system and method for diagnosing test-taking errors based on blooms taxonomy
US20090012789A1 (en) * 2006-10-18 2009-01-08 Teresa Ruth Gaudet Method and process for performing category-based analysis, evaluation, and prescriptive practice creation upon stenographically written and voice-written text files
US8321197B2 (en) * 2006-10-18 2012-11-27 Teresa Ruth Gaudet Method and process for performing category-based analysis, evaluation, and prescriptive practice creation upon stenographically written and voice-written text files
US20080113327A1 (en) * 2006-11-10 2008-05-15 Microsoft Corporation Interactive system for teaching and learning algorithms through discovery
US20080166686A1 (en) * 2007-01-04 2008-07-10 Cristopher Cook Dashboard for monitoring a child's interaction with a network-based educational system
US20130177892A1 (en) * 2007-02-21 2013-07-11 University Of Central Florida Research Foundation, Inc. Computing device providing electronic book data having selectable content layers with different difficultly levels and related methods
US9368038B2 (en) * 2007-02-21 2016-06-14 University Of Central Florida Research Foundation, Inc. Computing device providing electronic book data with configurable problems and changeable seed values and related methods
US20130179823A1 (en) * 2007-02-21 2013-07-11 University Of Central Florida Research Foundation, Inc. Computing device providing electronic book data with configurable problems and changeable seed values and related methods
US9443442B2 (en) * 2007-02-21 2016-09-13 University Of Central Florida Research Foundation, Inc. Computing device providing electronic book data having selectable content layers with different difficulty levels and related methods
US20100285441A1 (en) * 2007-03-28 2010-11-11 Hefferman Neil T Global Computer Network Self-Tutoring System
WO2009049353A1 (en) * 2007-10-15 2009-04-23 Vocam Pty Ltd An apparatus for education and assessment rather than and apparatus for education and assessment
US20100279265A1 (en) * 2007-10-31 2010-11-04 Worcester Polytechnic Institute Computer Method and System for Increasing the Quality of Student Learning
US20100221694A1 (en) * 2007-11-13 2010-09-02 Blossom E-Learning Ltd. Adaptive electronic learning system and method
WO2009063469A3 (en) * 2007-11-13 2010-03-11 Blossom E-Learning Ltd. Adaptive electronic learning system and method
WO2009063469A2 (en) * 2007-11-13 2009-05-22 Blossom E-Learning Ltd. Adaptive electronic learning system and method
US20110033834A1 (en) * 2007-12-28 2011-02-10 Peter Krumhauer Eteacher - electronic teacher for conveying complex content for learning
US8385812B2 (en) 2008-03-18 2013-02-26 Jones International, Ltd. Assessment-driven cognition system
US20100068687A1 (en) * 2008-03-18 2010-03-18 Jones International, Ltd. Assessment-driven cognition system
US20090253114A1 (en) * 2008-04-02 2009-10-08 Sinapse Print Simulators Automatic trace analysis and comparison system for interactive learning and training systems
US8714981B2 (en) * 2008-04-02 2014-05-06 Sinapse Print Simulators Automatic trace analysis and comparison system for interactive learning and training systems
US20090325140A1 (en) * 2008-06-30 2009-12-31 Lou Gray Method and system to adapt computer-based instruction based on heuristics
WO2010002395A1 (en) * 2008-06-30 2010-01-07 Dreambox Learning Inc. Method and system to adapt computer-based instruction based on heuristics
US20100209896A1 (en) * 2009-01-22 2010-08-19 Mickelle Weary Virtual manipulatives to facilitate learning
US20110076654A1 (en) * 2009-09-30 2011-03-31 Green Nigel J Methods and systems to generate personalised e-content
US20110097698A1 (en) * 2009-10-27 2011-04-28 Henderson Charles A System and method for interactive learning
US20110143328A1 (en) * 2009-12-14 2011-06-16 Gerald Alfred Brusher Method and Apparatus for Enhancing an Academic Environment
US20110195389A1 (en) * 2010-02-08 2011-08-11 Xerox Corporation System and method for tracking progression through an educational curriculum
US20120045744A1 (en) * 2010-08-23 2012-02-23 Daniel Nickolai Collaborative University Placement Exam
US8684746B2 (en) * 2010-08-23 2014-04-01 Saint Louis University Collaborative university placement exam
US20130260359A1 (en) * 2010-10-29 2013-10-03 Sk Telecom Co., Ltd. Apparatus and method for diagnosing learning ability
US20130266916A1 (en) * 2011-11-23 2013-10-10 Rodney A. Weems Systems and Methods Using Mathematical Reasoning Blocks
US10460615B2 (en) * 2011-11-23 2019-10-29 Rodney A. Weems Systems and methods using mathematical reasoning blocks
US8606170B2 (en) 2012-01-20 2013-12-10 Northrop Grumman Systems Corporation Method and apparatus for interactive, computer-based, automatically adaptable learning
US20150111191A1 (en) * 2012-02-20 2015-04-23 Knowre Korea Inc. Method, system, and computer-readable recording medium for providing education service based on knowledge units
US11605305B2 (en) * 2012-02-20 2023-03-14 Knowre Korea Inc. Method, system, and computer-readable recording medium for providing education service based on knowledge units
US10937330B2 (en) * 2012-02-20 2021-03-02 Knowre Korea Inc. Method, system, and computer-readable recording medium for providing education service based on knowledge units
US20210125514A1 (en) * 2012-02-20 2021-04-29 Knowre Korea Inc. Method, system, and computer-readable recording medium for providing education service based on knowledge units
US20130260351A1 (en) * 2012-03-29 2013-10-03 Dreambox Learning Inc. Calendar-driven sequencing of academic lessons
US8696365B1 (en) * 2012-05-18 2014-04-15 Align, Assess, Achieve, LLC System for defining, tracking, and analyzing student growth over time
US9858828B1 (en) * 2013-03-15 2018-01-02 Querium Corporation Expert systems and methods for dynamic assessment and assessment authoring
US10467919B2 (en) 2013-03-15 2019-11-05 Querium Corporation Systems and methods for AI-based student tutoring
US20160225273A1 (en) * 2015-01-29 2016-08-04 Xerox Corporation Methods and systems for evaluating user
WO2017011910A1 (en) * 2015-07-21 2017-01-26 Varafy Corporation Method and system for templated content generation and assessment
US11600196B2 (en) * 2017-03-13 2023-03-07 Vitruv Inc. Method and system for supporting learning, and non-transitory computer-readable recording medium
US20220301450A1 (en) * 2017-12-13 2022-09-22 Caveon, Llc Systems and Methods for Testing Skills Capability Using Technologically-Enhanced Questions in a Computerized Environment
US11386798B2 (en) * 2017-12-13 2022-07-12 Caveon, Llc Systems and methods for testing skills capability using technologically-enhanced questions in a computerized environment
US11961416B2 (en) * 2017-12-13 2024-04-16 Caveon, Llc Systems and methods for testing skills capability using technologically-enhanced questions in a computerized environment
CN109559580A (en) * 2018-12-29 2019-04-02 武汉易测云网络科技有限公司 A kind of on-line study system
US20220293006A1 (en) * 2021-03-15 2022-09-15 Casio Computer Co., Ltd. Information processing device, controlling method and recording medium

Also Published As

Publication number Publication date
TWI270023B (en) 2007-01-01
CN1770222A (en) 2006-05-10
TW200615867A (en) 2006-05-16

Similar Documents

Publication Publication Date Title
US20060099563A1 (en) Computerized teaching, practice, and diagnosis system
Kurt Intelligent tutoring systems for continuous, embedded assessment
Foshay et al. Principles for teaching problem solving
US20040018479A1 (en) Computer implemented tutoring system
US20040072130A1 (en) Computer-based educational learning
Jeremic et al. Evaluating an intelligent tutoring system for design patterns: The DEPTHS experience
RU2010104996A (en) DEVICE, SYSTEM AND METHOD OF ADAPTIVE TEACHING AND TRAINING
Lepak et al. Capturing and characterizing students’ strategic algebraic reasoning through cognitively demanding tasks with focus on representations
Roth et al. Evaluating student response to WeBWorK, a web-based homework delivery and grading system
US8545232B1 (en) Computer-based student testing with dynamic problem assignment
Morrison et al. Advancing evidence-based practice through program evaluation: A practical guide for school-based professionals
Lopes et al. Improving students skills to solve elementary equations in k-12 programs using an intelligent tutoring system
Ball et al. Recognising Equivalent Algebraic Expressions: An Important Component of Algebraic Expectation for Working with CAS.
Mohamed et al. Developing adaptive intelligent tutoring system based on item response theory and metrics
Rudzewitz et al. Enhancing a Web-based Language Tutoring System with Learning Analytics.
Pisan et al. Submit! a web-based system for automatic program critiquing
Alfrey et al. Developing a rubric to assess critical thinking in assignments with an open ended component
Foshay et al. Principles for Teaching Problem Solving. Technical Paper.
Lenczewski Using Just-in-Time Teaching To Engage Rural Students in Small Enrollment Organic Chemistry Classes
Li et al. Assess: promoting learner reflection in student self-assessment
Barrett Facilitating variable-length computerized classification testing via automatic racing calibration heuristics
Paul Improving education through computer-based alternative assessment methods
Nirmalakhandan Computer-aided tutorials and tests for use in distance learning
Shute et al. Diagnostic assessment in mathematics problem solving
Wen et al. Adaptive Assessment in Web-based learning

Legal Events

Date Code Title Description
AS Assignment

Owner name: MULTIPI, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LIU, ZHENYU LAWRENCE;CHEN, XIN;CHEN, LANG SHIE;REEL/FRAME:015969/0794

Effective date: 20041105

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION