WO2021095561A1 - Information processing device, information processing method, and program - Google Patents

Information processing device, information processing method, and program Download PDF

Info

Publication number
WO2021095561A1
WO2021095561A1 PCT/JP2020/040771 JP2020040771W WO2021095561A1 WO 2021095561 A1 WO2021095561 A1 WO 2021095561A1 JP 2020040771 W JP2020040771 W JP 2020040771W WO 2021095561 A1 WO2021095561 A1 WO 2021095561A1
Authority
WO
WIPO (PCT)
Prior art keywords
learning
user
environment
state
information
Prior art date
Application number
PCT/JP2020/040771
Other languages
French (fr)
Japanese (ja)
Inventor
拓郎 川合
Original Assignee
ソニーグループ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニーグループ株式会社 filed Critical ソニーグループ株式会社
Publication of WO2021095561A1 publication Critical patent/WO2021095561A1/en

Links

Images

Classifications

    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B5/00Electrically-operated educational appliances
    • G09B5/08Electrically-operated educational appliances providing for individual presentation of information to a plurality of student stations
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B7/00Electrically-operated teaching apparatus or devices working with questions and answers
    • G09B7/06Electrically-operated teaching apparatus or devices working with questions and answers of the multiple-choice answer-type, i.e. where a given question is provided with a series of answers and a choice has to be made from the answers
    • G09B7/07Electrically-operated teaching apparatus or devices working with questions and answers of the multiple-choice answer-type, i.e. where a given question is provided with a series of answers and a choice has to be made from the answers providing for individual presentation of questions to a plurality of student stations

Definitions

  • This technology relates to an information processing device, an information processing method, and a program, and particularly to an information processing device, an information processing method, and a program for improving the learning efficiency of a user.
  • Patent Document 1 discloses a technique for changing the difficulty level of a problem presented to a user according to the learning level or the like of the user.
  • Patent Document 2 discloses that a stress index is measured from a user's brain wave or the like, and white noise is generated based on the stress index to relieve stress.
  • Patent Document 3 discloses a technique for estimating a user's emotional state from the state of an electronic pen.
  • This technology was made in view of such a situation, and aims to improve the learning efficiency of the user.
  • the information processing device or program of one aspect of the present technology has a processing unit that calculates the changed contents for the learning environment of the user and the learning state is improved based on the learning state of the user.
  • the processing unit of the information processing device including the processing unit is a change content with respect to the learning environment of the user based on the learning state of the user, and the learning state is improved. This is an information processing method for calculating changes.
  • the change contents for the learning environment of the user and the change contents for which the learning state is improved are calculated based on the learning state of the user. Will be done.
  • FIG. 1 It is a block diagram which showed the structural example of one Embodiment of the information processing apparatus to which this technique is applied. It is a functional block diagram explaining the function of the information processing apparatus of FIG. It is a figure which illustrated the type of the sensor which the user state sensing part can use for sensing the user state, and the information (the purpose of sensing) obtained by the sensor. It is a figure which illustrated the element of the learning environment sensed by the learning environment sensing part, and the type of a sensor. It is a figure which illustrated the element of the learning environment controlled by the environment control unit, and the type of the environment control device used for controlling each element. It is a flowchart explaining the processing example performed by the information processing apparatus of FIG.
  • FIG. 1 is a block diagram showing a configuration example of an embodiment of an information processing device to which the present technology is applied.
  • the information processing device 11 includes an information processing unit 12, various sensors 13, and various environmental control devices 14.
  • the information processing unit 12 includes a computer, and may be, for example, a personal computer, a smartphone, a notepad, a mobile phone, or the like.
  • Various sensors 13 include one or more types of sensors.
  • the various sensors 13 include sensors that sense the user's state such as the position and behavior of the user, and sensors that sense the user's learning environment such as sound and temperature.
  • the various sensors 13 are connected to the communication unit 27 or the connection port 28 described later of the information processing unit 12 and exchange information with the information processing unit 12.
  • the environmental control device 14 includes one or a plurality of types of devices that change the sound, temperature, etc. of the user's learning environment.
  • the environmental control device 14 is connected to the communication unit 27 or the connection port 28 described later of the information processing unit 12 and exchanges information with the information processing unit 12.
  • the information processing unit 12 includes, for example, a CPU (Central Processing Unit) 21, a ROM (Read Only Memory) 22, a RAM (Random Access Memory) 23, an input unit 24, an output unit 25, a storage unit 26, a communication unit 27, and a connection port. It has 28 and a drive 29.
  • a CPU Central Processing Unit
  • ROM Read Only Memory
  • RAM Random Access Memory
  • the CPU 21 performs all or part of the operation of each component of the information processing unit 12 via the bus 31 and the input / output interface 32 based on various programs recorded in the ROM 22, the RAM 23, the storage unit 26, or the removable media 30. Control.
  • the ROM 22 stores a program read into the CPU 21, data used for calculation, and the like.
  • the RAM 23 temporarily stores a program read into the CPU 21 and various parameters that change as appropriate when the program is executed.
  • the input unit 24 is a device for a user to input information, and may be, for example, a mouse, a keyboard, a touch panel, a microphone, a button, a switch, or the like.
  • the output unit 25 is a device that visually or audibly notifies the user of information, and may be, for example, a display device, an audio output device such as a speaker and headphones, a printer, a facsimile, or the like.
  • the storage unit 26 is a device for storing various types of data, and may be, for example, a magnetic storage device such as a hard disk drive, a semiconductor storage device, an optical storage device, an optical magnetic storage device, or the like.
  • the communication unit 27 is a communication device for connecting to a network, and may be, for example, a wired LAN or a wireless LAN, Bluetooth (registered trademark), or the like.
  • connection port 28 is a port for connecting an externally connected device, and may be, for example, a USB port, an IEEE1394 port, SCSI, an optical audio terminal, or the like.
  • the drive 29 is a device that reads or writes information to a removable medium 30 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory.
  • the information processing unit 12 can install the program in the storage unit 26 via the input / output interface 32 by mounting the removable media 30 storing the program executed by the CPU 21 in the drive 29.
  • the program may be received by the communication unit 27 and installed in the storage unit 26 via a wired or wireless transmission medium, or may be installed in the ROM 22 or the storage unit 26 in advance. Further, the program executed by the CPU 21 may be a program that is processed in chronological order in the order described in this specification, or may be a program that is processed in parallel or at a necessary timing such as when a call is made. It may be a program in which processing is performed.
  • FIG. 2 is a functional block diagram illustrating the function of the information processing device 11 of FIG.
  • the information processing device 11 has a learning content control unit 41 and a learning environment control unit 42.
  • the learning content control unit 41 controls to provide the user with a problem according to the degree of understanding of the user's learning.
  • the learning environment control unit 42 controls the environment (learning environment) of the learning space in which the user learns in order to improve the learning state such as the degree of concentration of the user on learning.
  • the learning content control unit 41 includes a user interface unit 61, a user state sensing unit 62, a learning data analysis unit 63, and a problem generation unit 64.
  • the user interface unit 61 includes the input unit 24 and the output unit 25 of FIG. 1, presents information to the user, and receives information from the user.
  • the user interface unit 61 presents the problem q from the problem generation unit 64 to the user by the output unit 25. Further, when the user inputs information such as an answer to the question q from the input unit 24, the user interface unit 61 supplies the answer information R including the answer and the answer time information to the learning data analysis unit 63.
  • the user solves the problem q from the problem generation unit 64 presented by the user interface unit 61 in the learning environment E'. Then, the user inputs the answer from the user interface unit 61.
  • the answer time is the time required for the user to solve the problem q, and the user may input the answer time from the user interface unit 61, or the user interface unit 61 presents the problem q to the user.
  • the learning data analysis unit 63 may calculate based on the time until the answer is input from the user interface unit 61.
  • the user state sensing unit 62 includes some of the sensors 13 of the various sensors 13 shown in FIG.
  • the user state sensing unit 62 senses the state (user state G) of the user who is learning using the user interface unit 61, and causes the learning data analysis unit 63 and the environment data analysis unit 82 of the learning environment control unit 42 to sense the state (user state G). Supply.
  • the user state sensed by the user state sensing unit 62 with respect to the actual user state G'of the user is represented by the user state G in consideration of the measurement error of each sensor.
  • the user state sensing unit 62 may include the arithmetic processing function of the CPU 21, and may acquire the information obtained by the CPU 21 performing predetermined signal processing on the signals directly obtained from the various sensors 13 as the user state G.
  • FIG. 3 is a diagram illustrating the types of sensors that the user state sensing unit 62 can use for sensing the user state and the information (purpose of sensing) obtained by the sensors.
  • GPS Global Positioning System
  • a camera a motion sensor
  • a microphone a biometric information sensor
  • a depth sensor a depth sensor
  • an acceleration sensor a angular speed
  • sensors etc.
  • GPS senses the user's position when the user carries or wears GPS. By sensing the user's position, the user's behavior such as whether the user is in the same position or moved can be grasped in addition to the place where the user is learning (home or outside, etc.).
  • the GPS may be mounted on a mobile terminal such as a smartphone.
  • the camera is one or more cameras that capture the user's learning space.
  • the user state sensing unit 62 senses the position of the user in the learning space based on the image obtained from the camera.
  • the user's behavior can be grasped by sensing the user's position.
  • the user's minute behavior such as facial movement can be grasped from the image obtained from the camera.
  • the motion sensor is installed in the learning space and senses the position of the user in the learning space using infrared rays or the like. In addition, the user's behavior can be grasped by sensing the user's position.
  • the microphone is installed in the learning space and senses the user's voice. By sensing the user's voice, the state of fatigue of the user can be grasped.
  • the biological information sensor senses a biological state such as a user's pulse, sweating, brain wave, touch, smell, or taste. By sensing the user's pulse, sweating, and brain waves, the degree of concentration of the user on learning can be grasped.
  • sensing the user's sense of touch, smell, or taste means sensing how much the user's sense of touch, smell, or taste is working. By sensing the user's sense of touch, smell, or taste, the degree of concentration of the user on learning can be grasped.
  • the depth sensor senses depth information (three-dimensional position including the depth direction) in the user's learning space. By sensing the depth information, the user's three-dimensional position and behavior can be grasped.
  • the acceleration sensor senses the user's acceleration when the user carries or wears the acceleration sensor. By sensing the acceleration of the user, it is possible to grasp minute actions (movements) such as a change in posture that does not accompany the movement of the position in addition to the movement of the position of the user.
  • the acceleration sensor may be mounted on a mobile terminal such as a smartphone carried by the user.
  • the angular velocity sensor senses the user's angular velocity when the user carries or wears the angular velocity sensor. By sensing the user's angular velocity, it is possible to grasp minute actions (movements) that change the direction of the user.
  • the user state sensing unit 62 does not have to have all the types of sensors shown in FIG. 3, and may have other types of sensors as long as it is a sensor that senses the user state. Good. Further, the user state to be detected may be any one or more of the user's position, behavior, orientation, pulse, sweating, brain wave, touch, smell, and taste.
  • the learning data analysis unit 63 is a functional block realized by the arithmetic processing of the CPU 21 of FIG. 1, and includes the answer information R from the user interface unit 61 and the user state G from the user state sensing unit 62. Based on this, the learning state (good quality for learning) such as the user's concentration on learning, comprehension (learning degree), and learning speed is analyzed.
  • the learning data analysis unit 63 supplies the analysis result AI representing the learning state of the analyzed user to the problem generation unit 64.
  • the degree of concentration of the user on learning means an index of the user's concentration.
  • the learning data analysis unit 63 can obtain the degree of concentration based on the user state G from the user state sensing unit 62, for example, and may particularly obtain it from the user's ecological information. Further, the learning data analysis unit 63 may obtain the degree of concentration from the time since the user starts learning, the time, the correct answer rate for the problem, the transition of the time required for the answer, and the like.
  • the user's understanding of learning means an index of the user's understanding of a predetermined learning area.
  • the learning area refers to the range of learning consisting of learning units such as grades, subjects, and units.
  • the learning data analysis unit 63 obtains the degree of understanding based on the correct answer rate of the question, the time required for the answer, and the questionnaire result for the question.
  • the learning speed for the user's learning means an index of the speed of understanding the problem once learned by the user. Whether or not the learning data analysis unit 63 was able to solve the wrong problem once (if it was possible to solve it, the time required for the answer), and whether or not it was able to solve a similar problem related to the correct answer (if it was possible, the time required for the answer). ), Find the learning speed based on the correct answer rate when answering the wrong question again and the time required to answer.
  • the learning data analysis unit 63 supplies a part or all of the analyzed learning state (analysis result AI) as learning information C to the environment data analysis unit 82 of the learning environment control unit 42.
  • the learning data analysis unit 63 supplies the analyzed user's concentration, comprehension, and concentration and comprehension of the learning speed to the environment data analysis unit 82 as learning information C.
  • the learning data analysis unit 63 supplies all the information of the degree of concentration, the degree of understanding, and the acquisition speed, or any one or two pieces of information as learning information C to the environmental data analysis unit 82. May be good.
  • the learning information C supplied from the learning data analysis unit 63 to the environment data analysis unit 82 is information representing the user's current learning state (good quality for learning), the degree of concentration, the degree of understanding, and the degree of understanding, and Information other than the learning speed may be used.
  • the problem generation unit 64 generates a problem according to the learning state of the user, for example, a problem having a difficulty level according to the learning state of the user, based on the analysis result AI from the learning data analysis unit 63, and the user interface unit 61. Supply to.
  • Patent Document 1 International Publication No. 2016/0884663
  • the technique described in Patent Document 1 may be applied to the learning content control unit 41.
  • Patent Document 1 does not have the learning environment control unit 42 of FIG. Therefore, the technique of Patent Document 1 does not improve the learning efficiency by controlling the learning environment as in the information processing device 11 to which the present technique is applied.
  • the learning environment control unit 42 acquires the learning information C including the learning state of the user from the learning content control unit 41, and the temperature of the learning environment is improved so that the learning state is improved. It is possible to improve the learning efficiency appropriately because such changes are made.
  • the learning environment control unit 42 includes a learning environment sensing unit 81, an environmental data analysis unit 82, and an environment control unit 83.
  • the learning environment sensing unit 81 includes some of the sensors 13 of the various sensors 13 shown in FIG.
  • the learning environment sensing unit 81 senses the learning environment of the user and supplies the environment information E representing the current state of the learning environment to the environment data analysis unit 82.
  • the learning environment sensing unit 81 may include the arithmetic processing function of the CPU 21 and acquire the information obtained by the CPU 21 performing predetermined signal processing on the signals directly obtained from the various sensors 13 as the environment information E.
  • FIG. 4 is a diagram illustrating the elements of the learning environment sensed by the learning environment sensing unit 81 and the types of sensors.
  • FIG. 4 as elements of the learning environment sensed by the learning environment sensing unit 81, sound, image, illuminance, temperature, humidity, atmospheric pressure, open / closed state of windows and doors, clutter of the room, presence / absence of others, weather, and , Time (learning time, time) is shown.
  • the loudness of sound such as noise in the learning space is detected by a microphone installed in the learning space.
  • video sensing for example, information on whether or not video is being displayed is acquired from a video display device (display, etc.) or from a video output device that supplies video to the video display device, and is deployed in the learning space. It is detected whether or not the image is displayed on the displayed image display device.
  • the height of illuminance in the learning space is detected by an illuminance sensor installed in the learning space.
  • the height of the illuminance in the learning space may be detected by acquiring the information of the set value of the illuminance in the lighting device that illuminates the learning space.
  • the high illuminance of the learning space may be detected by analyzing the image from the camera that captures the learning space.
  • the high temperature of the learning space is detected by the temperature sensor built into the air conditioning equipment in the learning space or the temperature sensor installed in the learning space separately from the air conditioning equipment.
  • the high humidity of the learning space is detected by the humidity sensor built into the air conditioning equipment in the learning space or the humidity sensor installed in the learning space separately from the air conditioning equipment.
  • the height of atmospheric pressure in the learning space is detected by the atmospheric pressure sensor built into the air conditioning equipment in the learning space or the atmospheric pressure sensor installed in the learning space separately from the air conditioning equipment.
  • the open / closed state of the window or door that shields the learning space is detected based on the image from the camera that captures the learning space or by the open / close sensor installed on the window or door. The door.
  • the clutter of the room is detected based on the image from the camera that captures the space of the learning environment.
  • the presence or absence of others is detected by analyzing whether or not there are multiple persons in the learning space based on the image from the camera that captures the learning space.
  • the presence or absence of others may be detected by a motion sensor or a depth sensor installed in the learning space.
  • weather sensing it is detected whether the weather is sunny, cloudy, or rainy based on the information from the humidity sensor, temperature sensor, and illuminance sensor.
  • the weather information may be obtained from an internet site or the like.
  • time information is acquired from the clock function built in the information processing unit 12 or a specific server on the Internet, and the learning time (elapsed time from the start of learning) and the current time are detected.
  • the learning environment element sensed by the learning environment sensing unit 81 may be any one or a plurality of elements of the learning environment shown in FIG.
  • the environment data analysis unit 82 is a functional block realized by the arithmetic processing of the CPU 21 of FIG. 1, and includes the environment information E from the learning environment sensing unit 81, the user state G from the user state sensing unit 62, and the user state G. Based on the learning information C from the learning data analysis unit 63, the influence of the learning environment on the learning state of the user is analyzed.
  • the environment data analysis unit 82 calculates the change contents of the learning environment based on the analysis result, in which the learning state of the user is improved from the present.
  • the environment data analysis unit 82 supplies the control content (change content of the learning environment) for the learning environment for changing to the learning environment in which the learning state of the user is improved to the environment control unit 83 as the analysis result Ae.
  • the environmental control unit 83 includes the arithmetic processing function of the CPU 21 shown in FIG. 1 and various environmental control devices 14.
  • the environment control unit 83 controls the various environment control devices 14 shown in FIG. 2 based on the analysis result Ae from the environment data analysis unit 82, and controls the learning environment E'.
  • FIG. 5 is a diagram illustrating the elements of the learning environment controlled by the environment control unit 83 and the types of environment control devices used to control each element.
  • FIG. 5 illustrates noise, music, illuminance, temperature, humidity, information presentation (video), communication, and room clutter as elements of the learning environment controlled by the environment control unit 83.
  • Noise and music are both elements related to the sound of the learning environment, but they are separate elements in the control of the learning environment.
  • noise control whether or not noise cancellation is performed is controlled by an audio device connected to a speaker installed in the learning space or a headphone worn by the user.
  • the audio device supplies a sound having a phase opposite to the noise to the speaker or headphones.
  • music control whether or not to output music and the volume of music are controlled by the speakers installed in the learning space and the audio equipment connected to the headphones worn by the user.
  • music selection may be controlled based on the genre of music or the like.
  • the height of illuminance in the learning space is controlled by the lighting equipment installed in the learning space.
  • the temperature of the learning space is controlled by the air conditioner.
  • the humidity level of the learning space is controlled by the air conditioner.
  • the video display device In the control of information presentation (video), whether or not to display the video is controlled by the video display device (device that outputs the video to the display) installed in the learning space.
  • the video may be information for deepening the user's understanding, or may be information for increasing the concentration.
  • FIG. 6 is a flowchart illustrating a processing example performed by the information processing apparatus 11 of FIG.
  • steps S11 to S14 show a processing example of learning content control performed by the learning content control unit 41
  • steps S15 to S17 show a processing example of learning environment control performed by the learning environment control unit 42.
  • step S11 when the user solves the question q in the learning environment E'and inputs the answer (answer), the user interface unit 61 accepts the user's answer. Then, the user interface unit 61 supplies the answer information R including the user's answer and the answer time to the learning data analysis unit 63. The process proceeds from step S11 to step S12.
  • step S12 the user state sensing unit 62 senses the user state G'of the learning user and acquires the user state G. Then, the user state sensing unit 62 supplies the acquired user state G to the learning data analysis unit 63 of the learning content control unit 41 and the environment data analysis unit 82 of the learning environment control unit 42. The process proceeds from step S12 to step S13.
  • step S13 the learning data analysis unit 63 uses the answer information R supplied from the user interface unit 61 in step S11 and the user state G supplied from the user state sensing unit 62 in step S12 to concentrate the users.
  • the learning state such as degree, comprehension degree (learning degree), and learning speed is analyzed, and the analysis result AI representing the learning state of the user is calculated.
  • the learning data analysis unit 63 supplies the calculated analysis result AI to the problem generation unit 64.
  • the learning data analysis unit 63 supplies, for example, the degree of concentration and the degree of understanding as learning information C to the environment data analysis unit 82 among the learning states obtained by the analysis. The process proceeds from step S13 to step S14.
  • step S14 the problem generation unit 64 generates the problem q according to the learning state of the user based on the analysis result AI supplied from the learning data analysis unit 63 in step S13.
  • the problem q generated by the problem generation unit 64 is supplied to the output unit 25 (see FIG. 1) in the user interface unit 61.
  • the process returns to step S11 after step S14, and repeats steps S11 to S14.
  • step S15 the learning environment sensing unit 81 of the learning environment control unit 42 senses the environment (learning environment) of the learning space in which the user is learning, and acquires the environment information E.
  • the learning environment sensing unit 81 supplies the acquired environmental information E to the environmental data analysis unit 82. The process proceeds from step S15 to step S16.
  • the environment data analysis unit 82 includes the user state G supplied from the user state sensing unit 62 in step S12, the learning information C supplied from the learning data analysis unit 63 in step S13, and the learning environment in step S15. Using the environmental information E supplied from the sensing unit 81, it is analyzed whether or not the degree of concentration this time is improved as compared with the degree of concentration indicated by the learning information C supplied from the learning data analysis unit 63 last time.
  • the learning data analysis unit 63 calculates the next change content (control content) for the learning environment as the analysis result Ae, which improves the degree of concentration.
  • the learning data analysis unit 63 analyzes whether the learning state is improved in consideration of not only the improvement of the degree of concentration but also the degree of understanding of the user, and analyzes the changed contents of the learning environment in which the learning state is improved. It may be calculated as. The process proceeds from step S16 to step S17.
  • step S17 the environment control unit 83 controls the learning environment so that the learning environment E'is suitable for learning for the user, based on the analysis result Ae supplied from the environment data analysis unit 82 in step S16.
  • the process returns to step S15 after step S17, and repeats steps S15 to S17.
  • the learning environment is changed to the user because the learning environment control unit 42 changes the learning environment so that the learning state of the user is improved based on the learning information C from the learning content control unit 41.
  • the learning environment will be appropriately changed to be suitable for learning, and the learning efficiency will be improved appropriately.
  • FIG. 7 is a functional block diagram illustrating details of a processing example of learning environment control in the learning environment control unit 42.
  • FIG. 7 shows the environmental data analysis unit 82 and the environmental control unit 83 of FIG. 2, and the environmental control database 101 (not shown) of FIG.
  • the environment control database 101 is a functional block realized by the storage unit 26 of FIG. 1, and stores data and the like referred to by the environment data analysis unit 82.
  • FIG. 7 shows a configuration example of the environment control unit 83, and the environment control unit 83 includes a control signal generation unit 91 and devices 92A to 92N.
  • the control signal generation unit 91 is a functional block realized by the arithmetic processing of the CPU 21 of FIG.
  • the control signal generation unit 91 generates a control signal for controlling each of the devices 92A to 92N based on the analysis result Ae from the environment data analysis unit 82.
  • the analysis result Ae from the environment data analysis unit 82 includes, for example, information on the change contents (control contents) for each element of the learning environment shown in FIG.
  • each of the devices 92A to 92N is an environmental control device used for controlling each element of the learning environment.
  • the devices 92A to 92N are associated with elements of the learning environment that can be changed by each.
  • the control signal generation unit 91 is associated with each element of the learning environment 92A so that each element of the learning environment is changed according to the control content indicated by the analysis result Ae from the environment data analysis unit 82. To generate a control signal for the device 92N. Then, the control signal generation unit 91 supplies the generated control signal to each device 92A to 92N to control each device 92A to 92N.
  • the device 92A is an environment control device used for controlling noise in the learning environment shown in FIG. Further, the device 92A is, for example, an audio device that cancels noise by outputting a sound having a phase opposite to that of noise to a speaker installed in a learning space or a headphone worn by a user.
  • the control signal generation unit 91 controls the device 92A according to the control content included in the analysis result Ae and whether or not noise cancellation is performed.
  • the device 92B is an environmental control device used for controlling music in the learning environment shown in FIG.
  • the device 92B is, for example, an audio device that outputs music to a speaker installed in a learning space or a headphone worn by a user.
  • the control signal generation unit 91 controls the device 29B according to the control content included in the analysis result Ae and whether or not to output music.
  • the analysis result Ae may include a control content for raising or lowering the volume of the music.
  • the control signal generation unit 91 controls the device 92B to raise or lower the volume of the music by a certain amount. Lower.
  • the analysis result Ae may include a control content for changing a song (music genre, etc.).
  • the control signal generation unit 91 controls the device 92B to change the song.
  • the device 92C is an environmental control device used for controlling the illuminance of the learning environment shown in FIG.
  • the device 92C is, for example, a lighting device installed in a learning space.
  • the control signal generation unit 91 controls the device 29C according to the control content included in the analysis result Ae to increase, decrease, or maintain the height of the illuminance, and the high illuminance of the learning environment. The height is increased, decreased, or maintained by a certain amount.
  • the device 92D is an environmental control device used for controlling the temperature of the learning environment shown in FIG. 5, and is, for example, an air-conditioning device installed in the learning space.
  • the control signal generation unit 91 controls the device 29D according to the control content included in the analysis result Ae to raise, lower, or maintain the temperature height, and the temperature of the learning environment is high. The temperature is increased, decreased, or maintained by a certain amount.
  • the device 92E is an environmental control device used for controlling the information presentation (video) shown in FIG. 5, and is, for example, a video display device installed in a learning space.
  • the control signal generation unit 91 controls the device 29E according to the control content included in the analysis result Ae and whether or not to display the image, and controls whether or not to display the image in the learning space. ..
  • the content of the information (video) to be displayed includes, for example, information for deepening the user's understanding.
  • the device 92N is an environmental control device used for controlling factors that cause the user to lose concentration, and is, for example, a communication terminal used for controlling communication shown in FIG.
  • the control signal generation unit 91 controls the device 92N according to the control content included in the analysis result Ae and whether or not to cut off the communication.
  • control signal generation unit 91 is registered in advance by the device 92N whether or not to block the contact from others by e-mail, telephone, etc. to the user, or not to contact the user. Controls whether or not to announce to others.
  • a device electric key that controls the key of the entrance door to the learning space may be used.
  • the control signal generation unit 91 controls the device 92N according to the control content included in the analysis result Ae and whether or not to block the communication, and controls whether or not to lock the door. ..
  • the devices 92A to 92N shown in FIG. 7 are examples, and the environmental control unit 83 may have any one or more of the devices 92A to 92N, or the device 92A. Or may have a device other than the device 92N.
  • the environmental control unit 83 has a device for controlling the humidity of the learning environment (for example, an air conditioner) and a device for controlling the degree of clutter in the room (robot, etc.) as the environmental control device. You may be.
  • a device for controlling the humidity of the learning environment for example, an air conditioner
  • a device for controlling the degree of clutter in the room robot, etc.
  • control signal generation unit 91 is a device that controls the temperature of the learning environment according to the control content included in the analysis result Ae, which is to increase, decrease, or maintain the height of the humidity. To raise, lower, or maintain a certain amount of temperature in the learning environment.
  • control signal generation unit 91 controls the device that controls the clutter of the room according to the control content included in the analysis result Ae to reduce the clutter of the room, and cleans the room. Controls whether or not to keep things tidy.
  • the environment control unit 83 may control the scent of the learning environment by, for example, an aroma diffuser according to the analysis result Ae, or may control the lock of the TV, the game, or the smartphone to reduce the temptation. ..
  • the environmental control unit 83 may set the fixed telephone to the answering machine mode (answering machine) according to the analysis result Ae, or may prevent the intercom from being turned on. Further, the environment control unit 83 may operate the espresso machine according to the analysis result Ae to urge the user to take a break, or may control the humanoid robot or the animal robot to support the user. Good.
  • the environment data analysis unit 82 uses the environment information E from the learning environment sensing unit 81, the user state G from the user state sensing unit 62, and the learning information C from the learning data analysis unit 63 as input data, and each of the environments.
  • the analysis result Ae indicating the control content (change content) of the element is calculated as output data and supplied to the control signal generation unit 91.
  • the environmental information E includes sound, video, illuminance, temperature, humidity, atmospheric pressure, open / closed state of windows and doors, clutter of the room, presence / absence of others other than the user, and weather as shown in FIG. , And there is information about time etc. Any one or more of these pieces of information are given to the environmental data analysis unit 82 as environmental information E.
  • the user state G As the user state G, as shown in FIG. 3, there are states related to the user's position, behavior, orientation, pulse, sweating, brain wave, touch, smell, and taste. Information on one or more of these states is given to the environment data analysis unit 82 as the user state G.
  • the learning information C includes information on the learning state such as the user's concentration level, comprehension level, and learning speed for learning.
  • the information on the learning state such as the degree of concentration, the degree of understanding, and the learning speed
  • the information on the degree of concentration and the degree of understanding is given to the environmental data analysis unit 82 as the learning information C.
  • any one or more information on the learning state such as the degree of concentration, the degree of understanding, and the learning speed of the user is given to the environmental data analysis unit 82 as the learning information C. ..
  • the environmental data analysis unit 82 compares each of the environmental information E, the user state G, and the learning information C with the past values. For example, it is calculated how much the user state G and the learning information C have changed with respect to the amount of change in the environment information E, and which element of the learning environment has influenced the change in the user state G and the learning information C. Perform an analysis.
  • the environment data analysis unit 82 calculates the change contents (change contents for each element of the learning environment) for the learning environment for improving (increasing) the learning state such as the concentration degree and the comprehension degree of the user as a result of the analysis. As an analysis result Ae, it is supplied to the control signal generation unit 91.
  • the environment data analysis unit 82 becomes an environmental control device that controls the elements of the learning environment that have a large influence on the learning state.
  • the control policy (control content) different from the previous one may be changed. For example, when the learning state of the user deteriorates as a result of outputting music to the device 92B, the environmental data analysis unit 82 does not stop the output of the music to the device 92B, but outputs the music. You may change the genre.
  • the environment data analysis unit 82 maintains the state of the current learning environment or each element of the current learning environment. Set a control policy to change within a predetermined range. The environment data analysis unit 82 outputs the set control policy as the analysis result Ae to the control signal generation unit 91, aiming to further improve the learning state of the user.
  • the environment data analysis unit 82 stores the control policy for the environment information E, the user state G, and the learning information C when the user is learning, that is, the analysis result Ae in the environment control database 101. Create a database.
  • the environment data analysis unit 82 promptly learns the user regardless of the pattern of the environment information E and the user state G. It becomes possible to transition the state to a good state.
  • the environmental data analysis unit 82 may use any one of the concentration level, the comprehension level, the learning speed, and the like representing the learning state as an evaluation value representing the learning state of the user, or an element representing the learning state.
  • the weighted average of (concentration ratio, comprehension ratio, learning speed, etc.) may be used as the evaluation value.
  • the environmental data analysis unit 82 can determine that the larger the evaluation value, the better the learning state.
  • the environment data analysis unit 82 may change the state for each element of the learning environment that can be changed by the environment control unit 83 to set the state in which the evaluation value is maximized.
  • the environment data analysis unit 82 uses deep learning or other machine learning to learn the calculation process of the change content (analysis result Ae, which is the control policy) of the learning environment.
  • Learning model may be used.
  • the learning method in machine learning may be supervised learning or reinforcement learning. In the following, the process for calculating the analysis result Ae using the learning model will be described.
  • First processing example of the environmental data analysis unit 82 As a first processing example of the environmental data analysis unit 82, a case where a learning model (DNN) trained (trained) by supervised learning using machine learning (deep learning) is used will be described.
  • DNN learning model
  • the environment data analysis unit 82 inputs the environment information E supplied from the learning environment sensing unit 81, the user state G supplied from the user state sensing unit 62, and the learning information C supplied from the learning data analysis unit 63.
  • the analysis result Ae is calculated using a learning model that outputs output data indicating an appropriate control policy for the learning environment (changes in the learning environment that improve the learning state of the user).
  • control policy represents the change content (control content) for all the elements of the learning environment to be changed (controlled), and there is a control policy for the number of all combinations of the control content that can be adopted for each element. To do.
  • the learning model has output nodes corresponding to each of all control policies, and each output node outputs, for example, a value in the range of 0 to 1. Then, the output value from the output node corresponding to each control policy represents the appropriateness of each control policy.
  • the environmental data analysis unit 82 determines the control policy that maximizes the aptitude output from the output node of the learning model as the control policy that improves the learning state (for example, the degree of concentration) of the user. Further, the environmental data analysis unit 82 supplies the determined control policy as the analysis result Ae to the control signal generation unit 91 of the environmental control unit 83.
  • the input data to the learning model may be any one or more of the environment information E, the user state G, and the learning information C.
  • the environment data analysis unit 82 is a learned learning model (learning for learning data collection) stored in advance in the environment control database 101 until a user-specific learning model is generated by machine learning (deep learning). The training data is collected using the model).
  • the learning model for collecting learning data may be a learning model unrelated to the user (for example, an unlearned learning model), or a learning model corresponding to the user's tendency toward learning.
  • the environment data analysis unit 82 inputs to the user a condition for improving the learning state (concentration ratio, etc.) for each element of the learning environment to be controlled, for example. It is input from Fig. 1) and acquired as user information.
  • the conditions for improving the learning state for each element of the controlled environment are, for example, whether or not the person with noise can concentrate on learning, whether or not the person who is playing music can concentrate on learning, and what is the temperature. It is a condition for the user to judge that the learning state is improved for each element of the learning environment, such as whether the degree can concentrate on learning.
  • the environment control database 101 stores a learning model that calculates a substantially appropriate control policy for each of the same or similar user information.
  • a learning model that calculates a substantially appropriate control policy for each of the same or similar user information.
  • the environment data analysis unit 82 reads a learning model corresponding to the same or similar user information as the user information acquired from the user from the environment control database 101 and uses it as a learning model for collecting learning data.
  • the environmental data analysis unit 82 evaluates the goodness of the learning state of the user when collecting the learning data using the learning model for collecting the learning data.
  • the environmental data analysis unit 82 calculates the evaluation value Z based on the learning information C from the learning data analysis unit 63.
  • the environmental data analysis unit 82 includes a plurality of information (elements), for example, when the learning information C includes the concentration level and the comprehension level, or when the learning information C includes information such as the acquisition speed in addition to the concentration level and the comprehension level.
  • the weighted average of those plurality of elements is set as the evaluation value Z.
  • the evaluation value Z may be a weighted average when the weights other than one element are set to 0 among the plurality of elements of the learning information C. In this case, the evaluation value Z is the learning information C. It is the value of any one element.
  • the evaluation value Z indicates a higher value as the user's learning state is better (the higher the concentration and understanding), and the larger the increase in the evaluation value Z due to the control of the learning environment based on the analysis result Ae, the higher the value. It shows that the environmental control based on the analysis result Ae was an appropriate control content for the user to improve the learning state.
  • the environment data analysis unit 82 supplies the analysis result Ae to the environment control unit 83 and changes (changes) the learning environment, the environment information E from the learning environment sensing unit 81 and the user state sensing unit 62 The user state G and the learning information C from the learning data analysis unit 63 are acquired.
  • the environment data analysis unit 82 calculates the next control policy using the learning model based on the acquisition of the newly acquired environment information E, the user state G, and the learning information C. Then, it is supplied to the environment control unit 83 as the analysis result Ae.
  • the environment data analysis unit 82 calculates the evaluation value Z based on the newly acquired learning information C every time the learning environment is changed.
  • the time t is represented by the number of time steps, with the time from the change of the learning environment at a certain time by the analysis result Ae to the next change of the learning environment as one hour step. It is assumed that the one-hour step is a time longer than the time from when the environment control unit 83 changes the learning environment based on the analysis result Ae until the change in the learning environment appears as an effect on the learning state of the user.
  • the environmental information E, the user state G, and the learning information C acquired by the environmental data analysis unit 82 are E [t] and G, respectively. It is represented by [t] and C [t], and the evaluation value Z calculated based on the learning information C [t] is represented by Z [t]. Further, the analysis result Ae calculated based on the environmental information E [t], the user state G [t], and the learning information C [t] is represented by Ae [t].
  • the environmental data analysis unit 82 obtains an increase amount ⁇ Z [t + 1] of the evaluation value Z [t + 1] at the time t + 1 with respect to the evaluation value Z [t] at the time t, and the increase amount ⁇ Z [t + 1].
  • a predetermined threshold value ⁇ Zs it is determined that the control of the learning environment by the analysis result Ae [t] at the time t is an appropriate control content.
  • the environment data analysis unit 82 determines that the control of the learning environment by the analysis result Ae [t] has appropriate control contents, the environment information E [t], the user state G [t], and the learning information.
  • C [t] is stored in the environment control database 101 as input data in the training data.
  • the environmental data analysis unit 82 uses the learning model when the environmental information E [t], the user state G [t], and the learning information C [t] are input to the learning model for collecting the learning data as input data.
  • the output data of is stored in the environment control database 101 as teacher data in the training data.
  • the teacher data is output data from the learning model when the environment information E [t], the user state G [t], and the learning information C [t] are input to the learning model for collecting the learning data as input data.
  • the output data may be adjusted.
  • the output value from the output node of the learning model corresponding to the control policy set as the analysis result Ae [t] is adjusted to a value close to 1, and from other output nodes.
  • the data obtained by adjusting the output value of the above to a value close to 0 may be used as the training data.
  • the environmental data analysis unit 82 acquires the environmental information E, the user state G, and the learning information C, calculates the control policy (analysis result Ae) based on the environmental information E, the user state G, and the learning information C, and By repeating the calculation of the evaluation value Z, the learning data is collected and stored in the environment control database 101.
  • the environmental data analysis unit 82 uses the learning data stored in the environment control database 101 when a predetermined number or more of the learning data is stored in the environment control database 101. Learn the learning model for learning.
  • the learning model for learning may be a learning model for collecting learning data, or may be a learning model with initial values of parameters (weights and biases). Further, any of batch learning, mini-batch learning, and online learning may be adopted as a learning method of a learning model for learning using learning data.
  • the environment data analysis unit 82 stores the learned learning model generated by training the learning model for learning in the environment control database 101 as a learning model specialized for the user.
  • the environmental data analysis unit 82 reads the learning model specialized for the user from the environment control database 101 from the next learning of the user, and uses it for calculating the control policy (analysis result Ae).
  • the environment data analysis unit 82 collects the learning data in the same manner as when the learning model for collecting the learning data is used.
  • the number of accumulated learning data in the control database 101 may be increased.
  • the environmental data analysis unit 82 may perform learning of the learning model using all the learning data accumulated in the environment control database 101, or the increase from the previous learning of the learning model.
  • the learning model trained by the training data up to the previous time may be further trained using the training data.
  • FIG. 8 is a flowchart illustrating the learning model generation process performed by the environmental data analysis unit 82 in the first processing example.
  • step S31 when the user first learns using the information processing device 11 (when a learning model specialized for the user is not generated), the environment data analysis unit 82 learns from the environment control database 101. Read the learning model for data collection. The process proceeds from step S31 to step S32.
  • step S32 the environmental data analysis unit 82 sets the time t to 0. The process proceeds from step S32 to step S33.
  • step S33 the environment data analysis unit 82 acquires the environment information E [t] from the learning environment sensing unit 81, acquires the user state G [t] from the user state sensing unit 62, and learns from the learning data analysis unit 63. Acquire information C [t]. The process proceeds from step S33 to step S34.
  • step S34 the environment data analysis unit 82 calculates an evaluation value Z [t] representing the goodness of the learning state of the user based on the learning information C [t] acquired in step S33. The process proceeds from step S34 to step S35.
  • step S35 the environment data analysis unit 82 uses the environment information E [t], the user state G [t], and the learning information C [t] acquired in step S33 as input data, and the learning data read in step S31. Using the learning model for collection, the analysis result Ae [t], which is the control policy for the learning environment, is calculated. The process proceeds from step S35 to step S36.
  • step S36 the environment data analysis unit 82 supplies the analysis result Ae [t] calculated in step S35 to the control signal generation unit 91 of the environment control unit 83, and changes the learning environment according to the analysis result Ae [t]. ..
  • the process proceeds from step S36 to step S37.
  • the environmental data analysis unit 82 skips step S37 and step S38 and proceeds to step S39.
  • step S37 If it is determined in step S37 that the increase amount ⁇ Z [t] of the evaluation value is not equal to or greater than the predetermined threshold value ⁇ Zs, the process skips step S38 and proceeds to step S39.
  • step S37 If it is determined in step S37 that the increase amount ⁇ Z [t] of the evaluation value is equal to or greater than the predetermined threshold ⁇ Zs, the process proceeds to step S38, and the environmental data analysis unit 82 performs the environmental information E [1 hour before the step. t-1], the user state G [t-1], and the learning information C [t-1] are stored in the environment control database 101 as learning data. Further, the output of the learning model when the environmental information E [t-1], the user state G [t-1], and the learning information C [t-1] are input as input data to the learning model for collecting training data. The data or the adjusted data of the output data is stored in the environment control database 101 as learning data (teacher data). The process proceeds from step S38 to step S39.
  • step S39 the environment data analysis unit 82 determines whether or not the user's learning has been completed.
  • step S39 If it is determined in step S39 that the user's learning has not been completed, the process proceeds from step S39 to step S40.
  • step S40 the environmental data analysis unit 82 waits until the time for one hour step elapses from the time when the process of step S33 is started. Also, the time t represented by the number of time steps is incremented by 1. The process returns from step S40 to step S33, and steps S33 to S40 are repeated.
  • step S39 determines whether the user's learning has been completed. If it is determined in step S39 that the user's learning has been completed, the process proceeds from step S39 to step S41.
  • step S41 the environmental data analysis unit 82 trains the learning model for learning using the learning data stored in the environment control database 101, generates a learning model specialized for the user, and generates the learning model specialized for the user, and the environment control database. Store in 101.
  • the environment data analysis unit 82 will use the learning model for learning data collection even at the next learning of the user. Is used to perform the processes of steps S31 to S40 to collect learning data. Then, when the number of learning data stored in the environment control database 101 exceeds a predetermined number determined in advance, the environment data analysis unit 82 performs the process in step S41 to specialize in the user. Generate a learning model.
  • step S41 When the process of step S41 is completed, the process of this flowchart is completed.
  • FIG. 9 is a flowchart illustrating a process when the environment data analysis unit 82 calculates the analysis result Ae using the learned learning model specialized for the user in the first processing example.
  • step S51 when a user-specific learning model is generated, the environment data analysis unit 82 reads the user-specific learned learning model from the environment control database 101. The process proceeds from step S51 to step S52.
  • step S52 the environmental data analysis unit 82 sets the time t to 0. The process proceeds from step S52 to step S53.
  • step S53 the environment data analysis unit 82 acquires the environment information E [t] from the learning environment sensing unit 81, acquires the user state G [t] from the user state sensing unit 62, and learns from the learning data analysis unit 63. Acquire information C [t]. The process proceeds from step S53 to step S54.
  • step S54 the environment data analysis unit 82 takes the environment information E [t], the user state G [t], and the learning information C [t] acquired in step S53 as input data, and reads the learning model in step S51. Is used to calculate the analysis result Ae [t], which is the control policy for the learning environment. The process proceeds from step S54 to step S55.
  • step S55 the environment data analysis unit 82 supplies the analysis result Ae [t] calculated in step S54 to the control signal generation unit 91 of the environment control unit 83, and changes the learning environment according to the analysis result Ae [t]. ..
  • the process proceeds from step S55 to step S56.
  • step S56 the environment data analysis unit 82 determines whether or not the user's learning has been completed.
  • step S56 If it is determined in step S56 that the user's learning has not been completed, the process proceeds to step S57.
  • step S57 the environmental data analysis unit 82 waits until the time for one hour step elapses from the time when the process of step S53 is started. Also, the time t represented by the number of time steps is incremented by 1. The process returns from step S57 to step S53, and steps S53 to S57 are repeated.
  • step S56 determines whether the user's learning has been completed. If it is determined in step S56 that the user's learning has been completed, the process skips step S57 and the process of this flowchart ends.
  • the environmental data analysis unit 82 supplies the environmental information E supplied from the learning environment sensing unit 81 and the user state G supplied from the user state sensing unit 62. And the learning information C supplied from the learning data analysis unit 63, a control policy for improving the learning state of the user is calculated using the learning model, without the user's trouble, and the user's preference and personality, etc. Appropriate control of the learning environment will be performed in consideration of.
  • the environment data analysis unit 82 inputs the environment information E supplied from the learning environment sensing unit 81, the user state G supplied from the user state sensing unit 62, and the learning information C supplied from the learning data analysis unit 63.
  • the analysis result Ae is calculated using the DNN that outputs the value Q of all the control policies a as output data as a learning model.
  • the input data to the learning model may be any one or more of the environment information E, the user state G, and the learning information C.
  • control policy a has the same meaning as the control policy explained in the first processing example of the environmental data analysis unit 82, the description thereof is omitted here.
  • the learning model is gradually updated to a learning model specialized for the user by reinforcement learning, but the initial learning model used first by the environmental data analysis unit 82 is a learning model unrelated to the user (for example, not yet). It may be a learning model of learning), or it may be a learning model corresponding to a user's tendency toward learning. Since the initial learning model is the same as the learning model for collecting learning data in the case of the first processing example of the environmental data analysis unit 82, the description thereof will be omitted.
  • the value Q of the control policy a represents the goodness of the control policy a, and is calculated based on the reward V when the learning environment is changed according to the control policy a.
  • the reward V represents the degree of goodness of the learning state of the user as in the evaluation value Z in the case of the first processing example of the environmental data analysis unit 82.
  • the environmental data analysis unit 82 calculates the reward V based on the learning information C from the learning data analysis unit 63.
  • the environmental data analysis unit 82 includes a plurality of information (elements), for example, when the learning information C includes the concentration level and the comprehension level, or when the learning information C includes information such as the acquisition speed in addition to the concentration level and the comprehension level.
  • the weighted average of those plurality of elements is defined as the reward V.
  • the reward V may be a weighted average when the weights other than one element are set to 0 among the plurality of elements of the learning information C. In this case, the reward V is any one of the learning information C. It is the value of one element.
  • the reward V shows a higher value as the user's learning state is better (the higher the degree of concentration and understanding).
  • the value Q of the control policy a is the sum of the discount rewards V in the learning environment after the learning environment is changed according to the control policy a and the reward V that will be obtained thereafter (so-called Bellman equation). ..
  • the time t is a time represented by the number of time steps, as in the case of the first processing example of the environmental data analysis unit 82.
  • the value Q [t] represents the value Q of the control policy a when the learning environment is changed according to the control policy a with respect to the learning environment at a certain time t.
  • the control policy a at time t is hereinafter represented by a [t].
  • the reward V [t + 1] represents the reward V in the learning environment after the learning environment is changed according to the control policy a [t] with respect to the learning environment at time t.
  • Q [t + 1] represents the value Q of the control policy a [t + 1] when the learning environment at time t + 1 is changed according to the control policy a [t + 1], and is calculated by the learning model.
  • Qmax [t + 1] represents the maximum value of each value Q [t + 1] of all control policies a [t + 1] at time t + 1.
  • is a discount rate ( ⁇ is a value of 0 or more and 1 or less), which is a predetermined value.
  • the environment data analysis unit 82 receives the environment information E [t] from the learning environment sensing unit 81 acquired at a certain time t, the user state G [t] from the user state sensing unit 62, and the learning data analysis unit 63.
  • the learning information C [t] is input to the learning model, and the value Q [t] of all the control policies a [t] is calculated as the output data of the learning model.
  • the learning model has output nodes corresponding to each of all control policies a, and each output node outputs, for example, a value in the range of 0 to 1.
  • the output value from each output node of the learning model represents the value Q of the control policy a corresponding to each output node.
  • the environmental data analysis unit 82 determines, for example, the control policy a [t] that maximizes the value Q [t] among all the value Q [t] of the control policy a [t] output from the learning model.
  • the analysis result Ae [t] is supplied to the environment control unit 83.
  • the environment control unit 83 changes the learning environment based on the analysis result Ae [t]
  • the environment data analysis unit 82 receives the environment information E [t + 1] from the learning environment sensing unit 81 acquired at time t + 1.
  • the user state G [t + 1] from the user state sensing unit 62 and the learning information C [t + 1] from the learning data analysis unit 63 are input to the learning model, and all the control policies a which are the output data of the learning model.
  • Each value Q [t + 1] of [t + 1] is calculated.
  • the environmental data analysis unit 82 controls the environment by using, for example, the control policy ac [t + 1] at which the value Q [t + 1] is maximized as the analysis result Ae at the time t + 1 among the value Q [t + 1] output from the learning model. It is supplied to the unit 83.
  • the environmental data analysis unit 82 calculates the reward V [t + 1] based on the learning information C [t + 1]. As a result, the environmental data analysis unit 82 calculates the value Q [t] of the control policy ac [t], which is the analysis result Ae [t] at time t, by the above equation (1), and the calculated value Q [t]. Let be the correct answer value Q'[t] of the value of the control policy ac [t].
  • the control policy ac output from the learning model.
  • the learning model is trained so that the value Q [t] of [t] becomes the correct answer value Q'[t].
  • the environmental data analysis unit 82 acquires the environmental information E, the user state G, and the learning information C, calculates the analysis result Ae based on the environmental information E, the user state G, and the learning information C, and calculates the value Q. While repeating the above steps, the learning model is strengthened and learned. As a result, the learning model is gradually updated to a user-specific learning model.
  • the environment data analysis unit 82 acquires the correct answer value Q'[t] of the output data of the learning model for the input of the environment information E [t], the user state G [t], and the learning information C [t]. It is not necessary to train the learning model for each.
  • the environment data analysis unit 82 has the environment information E [t], the user state G [t], the learning information C [t], and the correct answer value Q'[t] of the output data of the learning model for those inputs. And are stored as learning data in the environment control database 101 and stored.
  • the environmental data analysis unit 82 learns the learning model using the learning data accumulated every time a predetermined number of learning data are accumulated.
  • FIG. 10 is a flowchart illustrating the processing when the environment data analysis unit 82 calculates the analysis result Ae using the learning model in the second processing example.
  • step S71 when the user first learns using the information processing device 11 (when the learning model reinforcement learning is not performed at all), the environment data analysis unit 82 initially starts from the environment control database 101. Read the learning model of. When the reinforcement learning of the learning model has already been performed, the environment data analysis unit 82 reads out the learning model in which the reinforcement learning has been performed from the environment control database 101. The process proceeds from step S71 to step S72.
  • step S72 the environmental data analysis unit 82 sets the time t to 0. The process proceeds from step S72 to step S73.
  • step S73 the environment data analysis unit 82 acquires the environment information E [t] from the learning environment sensing unit 81, acquires the user state G [t] from the user state sensing unit 62, and learns from the learning data analysis unit 63. Acquire information C [t]. The process proceeds from step S73 to step S74.
  • step S74 the environmental data analysis unit 82 calculates a reward V [t] representing the goodness of the learning state of the user based on the learning information C [t] acquired in step S73.
  • the process proceeds from step S74 to step S75.
  • the process skips step S74 and proceeds to step S75.
  • step S75 the environment data analysis unit 82 learns using the learning model with the environment information E [t], the user state G [t], and the learning information C [t] acquired in step S73 as input data.
  • the analysis result Ae [t] which is the control policy ac [t] for the environment, is calculated. The process proceeds from step S75 to step S76.
  • step S76 the environment data analysis unit 82 supplies the analysis result Ae [t] calculated in step S75 to the control signal generation unit 91 of the environment control unit 83, and changes the learning environment according to the analysis result Ae [t]. ..
  • the process proceeds from step S76 to step S77.
  • step S77 the maximum value Qmax [t] of the value Q [t] of each control policy output from the learning model in step S75 and the reward V [t] calculated in step S74 are added.
  • the calculated value Q [t-1] is set as the correct value Q'[t-1] of the value with respect to the control policy ac [t-1].
  • step S78 proceeds from step S77 to step S78.
  • the value Q [t-1] calculated by the process of step S77 is a value calculated by the calculation formula in which the time t is replaced with the time t-1 in the above formula (1).
  • the process skips step S77 and step S78 and proceeds to step S79.
  • step S78 the environment data analysis unit 82 uses the environment information E [t-1], the user state G [t-1], and the learning information C [t-1] one hour before as input data as input data for the learning model.
  • the learning model is trained so that the value Q [t-1] for the control policy ac [t-1] output from the learning model when inputting to is the correct answer value Q'[t-1]. Further, the environment data analysis unit 82 stores the trained learning model in the environment control database 101. The process proceeds from step S78 to step S79.
  • step S79 the environment data analysis unit 82 determines whether or not the user's learning has been completed.
  • step S79 If it is determined in step S79 that the user's learning has not been completed, the process proceeds to step S80.
  • step S80 the environmental data analysis unit 82 waits until the time for one hour step elapses from the time when the process of step S73 is started. Also, the time t represented by the number of time steps is incremented by 1. The process returns from step S80 to step S73, and steps S73 to S80 are repeated.
  • step S79 if it is determined in step S79 that the user's learning has been completed, the processing of this flowchart ends.
  • the environment data analysis unit 82 supplies the environment information E supplied from the learning environment sensing unit 81, the user state G supplied from the user state sensing unit 62, and the user state G. Based on the learning information C supplied from the learning data analysis unit 63, a control policy for improving the learning state of the user is calculated using the learning model, and the user's taste and personality are taken into consideration without the user's trouble. Appropriate control of the learning environment will be performed.
  • the information processing unit 12 is a server device connected to an information terminal (smartphone, personal computer, etc.) deployed in the user's learning space by a communication line such as the Internet. May be good.
  • the input unit and the output unit provided in the information terminal function as a substitute for the input unit 24 and the output unit 25 of FIG.
  • the information terminal functions as a device that mediates the exchange of information between the various sensors 13 and the various environmental control devices 14 and the server device.
  • the present technology can also have the following configurations.
  • the processing unit calculates the change content for the learning environment based on the evaluation value calculated based on any one or more of the concentration level, the comprehension level, and the learning speed.
  • ⁇ 5> The information processing apparatus according to any one of ⁇ 1> to ⁇ 4>, wherein the processing unit calculates the change contents with respect to the learning environment based on the current state of the learning environment.
  • the state of the learning environment is any of sound, image, illuminance, temperature, humidity, atmospheric pressure, open / closed state of windows or doors, clutter of the room, presence / absence of others, weather, and time.
  • the information processing apparatus according to ⁇ 5> which is in one or more states.
  • ⁇ 7> The information processing device according to any one of ⁇ 1> to ⁇ 6>, wherein the processing unit calculates the change contents with respect to the learning environment based on the state of the user.
  • the state of the user is one or more of the states related to the position, behavior, orientation, pulse, sweating, brain wave, touch, smell, and taste of the user ⁇ 7>.
  • ⁇ 9> The information processing apparatus according to any one of ⁇ 1> to ⁇ 8>, wherein the processing unit calculates the changed content with respect to the learning environment using a learning model trained by machine learning.
  • ⁇ 10> The information processing apparatus according to ⁇ 9>, wherein the processing unit learns the learning model based on the learning data collected when the learning environment of the user is changed and the learning state of the user is improved.
  • the processing unit calculates the value of each of the changes that can be taken for the learning environment using the learning model, and determines the changes for the learning environment based on the values.
  • ⁇ 12> The information processing apparatus according to any one of ⁇ 9> to ⁇ 11>, wherein the learning model is a deep neural network.
  • ⁇ 13> The information processing apparatus according to any one of ⁇ 1> to ⁇ 12>, further comprising a problem generation unit that presents a problem corresponding to the learning state of the user to the user.
  • ⁇ 14> The information processing apparatus according to any one of ⁇ 1> to ⁇ 13>, further comprising an environment control unit that changes the learning environment based on the change contents with respect to the learning environment calculated by the processing unit.
  • 11 information processing device 12 information processing unit, 13 various sensors, 14 various environment control devices, 21 CPU, 24 input unit, 25 output unit, 41 learning content control unit, 42 learning environment control unit, 61 user interface unit, 62 users State sensing unit, 63 learning data analysis unit, 64 problem generation unit, 81 learning environment sensing unit, 82 environment data analysis unit, 83 environment control unit, 91 control signal generation unit, 101 environment control database

Abstract

The present technology relates to an information processing device, an information processing method, and a program with which it is possible to improve the learning efficiency of a user. On the basis of the learning state of a user, the contents of a change to the learning environment of the user for improving the learning state are calculated.

Description

情報処理装置、情報処理方法、及び、プログラムInformation processing equipment, information processing methods, and programs
 本技術は、情報処理装置、情報処理方法、及び、プログラムに関し、特に、ユーザの学習効率の向上を図る情報処理装置、情報処理方法、及び、プログラムに関する。 This technology relates to an information processing device, an information processing method, and a program, and particularly to an information processing device, an information processing method, and a program for improving the learning efficiency of a user.
 特許文献1には、ユーザの学習度等に応じてユーザに提示する問題の難易度等を変更する技術が開示されている。 Patent Document 1 discloses a technique for changing the difficulty level of a problem presented to a user according to the learning level or the like of the user.
 特許文献2には、ユーザの脳波などからストレス指数を計り、ストレス指数に基づいてホワイトノイズを発生させてストレス緩和を図ることが開示されている。 Patent Document 2 discloses that a stress index is measured from a user's brain wave or the like, and white noise is generated based on the stress index to relieve stress.
 特許文献3には、電子ペンの状態からユーザの感情状態を推定する技術が開示されている。 Patent Document 3 discloses a technique for estimating a user's emotional state from the state of an electronic pen.
国際公開第2016/088463号International Publication No. 2016/0884663 特表2017-528282号公報Special Table 2017-528282 国際公開第2018/04306号International Publication No. 2018/04306
 ユーザの学習効率の向上を図ることは有益な課題であるが、ユーザごとに嗜好や性格などの個人差があるため、単にホワイトノイズを発生させる等の一律の対応では実現することが難しい。 It is a useful task to improve the learning efficiency of users, but it is difficult to realize it by simply generating white noise because there are individual differences in tastes and personalities of each user.
 本技術は、このような状況に鑑みてなされたものであり、ユーザの学習効率の向上を図るようにする。 This technology was made in view of such a situation, and aims to improve the learning efficiency of the user.
 本技術の一側面の情報処理装置、又は、プログラムは、ユーザの学習状態に基づいて、前記ユーザの学習環境に対する変更内容であって前記学習状態が向上する前記変更内容を算出する処理部を有する情報処理装置、又は、そのような情報処理装置として、コンピュータを機能させるためのプログラムである。 The information processing device or program of one aspect of the present technology has a processing unit that calculates the changed contents for the learning environment of the user and the learning state is improved based on the learning state of the user. An information processing device, or a program for operating a computer as such an information processing device.
 本技術の一側面の情報処理方法は、処理部を含む情報処理装置の前記処理部が、ユーザの学習状態に基づいて、前記ユーザの学習環境に対する変更内容であって前記学習状態が向上する前記変更内容を算出する情報処理方法である。 In the information processing method of one aspect of the present technology, the processing unit of the information processing device including the processing unit is a change content with respect to the learning environment of the user based on the learning state of the user, and the learning state is improved. This is an information processing method for calculating changes.
 本技術の一側面の情報処理装置、情報処理方法、及び、プログラムにおいては、ユーザの学習状態に基づいて、前記ユーザの学習環境に対する変更内容であって前記学習状態が向上する前記変更内容が算出される。 In the information processing device, the information processing method, and the program of one aspect of the present technology, the change contents for the learning environment of the user and the change contents for which the learning state is improved are calculated based on the learning state of the user. Will be done.
本技術を適用した情報処理装置の一実施の形態の構成例を示したブロック図である。It is a block diagram which showed the structural example of one Embodiment of the information processing apparatus to which this technique is applied. 図1の情報処理装置の機能を説明する機能ブロック図である。It is a functional block diagram explaining the function of the information processing apparatus of FIG. ユーザ状態センシング部がユーザ状態のセンシングに使用し得るセンサの種類とセンサにより得られる情報(センシングの目的)とを例示した図である。It is a figure which illustrated the type of the sensor which the user state sensing part can use for sensing the user state, and the information (the purpose of sensing) obtained by the sensor. 学習環境センシング部がセンシングする学習環境の要素とセンサの種類とを例示した図である。It is a figure which illustrated the element of the learning environment sensed by the learning environment sensing part, and the type of a sensor. 環境制御部が制御する学習環境の要素と、各要素の制御に用いられる環境制御機器の種類とを例示した図である。It is a figure which illustrated the element of the learning environment controlled by the environment control unit, and the type of the environment control device used for controlling each element. 図2の情報処理装置が行う処理例を説明するフローチャートである。It is a flowchart explaining the processing example performed by the information processing apparatus of FIG. 学習環境制御部における学習環境制御の処理例の詳細を説明する機能ブロック図である。It is a functional block diagram explaining the details of the processing example of the learning environment control in the learning environment control part. 環境データ解析部が第1の処理例において行う学習モデルの生成処理を説明するフローチャートである。It is a flowchart explaining the generation process of the learning model performed by the environment data analysis part in the 1st process example. 環境データ解析部が第1の処理例においてユーザに特化した学習済みの学習モデルを用いて解析結果を算出する際の処理を説明するフローチャートである。It is a flowchart explaining the process when the environment data analysis part calculates the analysis result using the trained learning model specialized for the user in the 1st process example. 環境データ解析部が第2の処理例において学習モデルを用いて解析結果を算出する際の処理を説明するフローチャートである。It is a flowchart explaining the process when the environmental data analysis part calculates the analysis result using the learning model in the 2nd process example.
 以下、図面を参照しながら本技術の実施の形態について説明する。 Hereinafter, embodiments of the present technology will be described with reference to the drawings.
<<本技術を適用した情報処理装置の一実施の形態>>
 図1は、本技術を適用した情報処理装置の一実施の形態の構成例を示したブロック図である。
<< An embodiment of an information processing device to which this technology is applied >>
FIG. 1 is a block diagram showing a configuration example of an embodiment of an information processing device to which the present technology is applied.
 図1において、情報処理装置11は、情報処理部12、各種センサ13、及び、各種環境制御機器14を有する。 In FIG. 1, the information processing device 11 includes an information processing unit 12, various sensors 13, and various environmental control devices 14.
 情報処理部12は、コンピュータを含み、例えば、パーソナルコンピュータ、スマートフォン、ノートパッド、又は、携帯電話などであってよい。 The information processing unit 12 includes a computer, and may be, for example, a personal computer, a smartphone, a notepad, a mobile phone, or the like.
 各種センサ13は、1又は複数の種類のセンサを含む。各種センサ13には、ユーザの位置や行動等のユーザ状態をセンシングするセンサや、音や温度等のユーザの学習環境をセンシングするセンサが含まれる。各種センサ13は、情報処理部12の後述の通信部27又は接続ポート28に接続されて情報処理部12との間で情報のやり取りを行う。 Various sensors 13 include one or more types of sensors. The various sensors 13 include sensors that sense the user's state such as the position and behavior of the user, and sensors that sense the user's learning environment such as sound and temperature. The various sensors 13 are connected to the communication unit 27 or the connection port 28 described later of the information processing unit 12 and exchange information with the information processing unit 12.
 環境制御機器14は、ユーザの学習環境の音や温度等を変更する1又は複数の種類の機器を含む。環境制御機器14は、情報処理部12の後述の通信部27又は接続ポート28に接続されて情報処理部12との間で情報のやり取りを行う。 The environmental control device 14 includes one or a plurality of types of devices that change the sound, temperature, etc. of the user's learning environment. The environmental control device 14 is connected to the communication unit 27 or the connection port 28 described later of the information processing unit 12 and exchanges information with the information processing unit 12.
 情報処理部12は、例えば、CPU(Central Processing Unit)21、ROM(Read Only Memory)22、RAM(Random Access Memory)23、入力部24、出力部25、記憶部26、通信部27、接続ポート28、及び、ドライブ29を有する。 The information processing unit 12 includes, for example, a CPU (Central Processing Unit) 21, a ROM (Read Only Memory) 22, a RAM (Random Access Memory) 23, an input unit 24, an output unit 25, a storage unit 26, a communication unit 27, and a connection port. It has 28 and a drive 29.
 CPU21は、ROM22、RAM23、記憶部26、又は、リムーバブルメディア30に記録された各種プログラムに基づいて情報処理部12の各構成要素の動作全般又は一部をバス31及び入出力インタフェース32を介して制御する。 The CPU 21 performs all or part of the operation of each component of the information processing unit 12 via the bus 31 and the input / output interface 32 based on various programs recorded in the ROM 22, the RAM 23, the storage unit 26, or the removable media 30. Control.
 ROM22は、CPU21に読み込まれるプログラムや演算に用いるデータ等を格納する。 The ROM 22 stores a program read into the CPU 21, data used for calculation, and the like.
 RAM23は、CPU21に読み込まれるプログラムや、そのプログラムを実行する際に適宜変化する各種パラメータ等を一時的に格納する。 The RAM 23 temporarily stores a program read into the CPU 21 and various parameters that change as appropriate when the program is executed.
 入力部24は、ユーザが情報を入力する装置であり、例えば、マウス、キーボード、タッチパネル、マイクロフォン、ボタン、及び、スイッチ等であってよい。 The input unit 24 is a device for a user to input information, and may be, for example, a mouse, a keyboard, a touch panel, a microphone, a button, a switch, or the like.
 出力部25は、ユーザに情報を視覚的又は聴覚的に通知する装置であり、例えば、ディスプレイ装置、スピーカ及びヘッドフォン等のオーディオ出力装置、プリンタ、並びに、ファクシミリ等であってよい。 The output unit 25 is a device that visually or audibly notifies the user of information, and may be, for example, a display device, an audio output device such as a speaker and headphones, a printer, a facsimile, or the like.
 記憶部26は、各種のデータを格納するための装置であり、例えば、ハードディスクドライブ等の磁気記憶デバイス、半導体記憶デバイス、光記憶デバイス、及び、光磁気記憶デバイス等であってよい。 The storage unit 26 is a device for storing various types of data, and may be, for example, a magnetic storage device such as a hard disk drive, a semiconductor storage device, an optical storage device, an optical magnetic storage device, or the like.
 通信部27は、ネットワークに接続するための通信デバイスであり、例えば、有線LAN又は無線LAN、及び、Bluetooth(登録商標)等であってよい。 The communication unit 27 is a communication device for connecting to a network, and may be, for example, a wired LAN or a wireless LAN, Bluetooth (registered trademark), or the like.
 接続ポート28は、外部接続機器を接続するためのポートであり、例えば、USBポート、IEEE1394ポート、SCSI、及び、光オーディオ端子等であってよい。 The connection port 28 is a port for connecting an externally connected device, and may be, for example, a USB port, an IEEE1394 port, SCSI, an optical audio terminal, or the like.
 ドライブ29は、磁気ディスク、光ディスク、光磁気ディスク、又は、半導体メモリ等のリムーバブルメディア30に対する情報の読み出し、又は、書き込みを行う装置である。 The drive 29 is a device that reads or writes information to a removable medium 30 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory.
 情報処理部12では、CPU21が実行するプログラムを記憶したリムーバブルメディア30をドライブ29に装着することにより、入出力インタフェース32を介して、記憶部26にそのプログラムをインストールすることができる。 The information processing unit 12 can install the program in the storage unit 26 via the input / output interface 32 by mounting the removable media 30 storing the program executed by the CPU 21 in the drive 29.
 なお、プログラムは、有線又は無線の伝送媒体を介して、通信部27で受信し、記憶部26にインストールしてもよいし、ROM22や記憶部26に予めインストールしておいてもよい。また、CPU21が実行するプログラムは、本明細書で説明する順序に沿って時系列に処理が行われるプログラムであっても良いし、並列に、あるいは呼び出しが行われたとき等の必要なタイミングで処理が行われるプログラムであっても良い。 The program may be received by the communication unit 27 and installed in the storage unit 26 via a wired or wireless transmission medium, or may be installed in the ROM 22 or the storage unit 26 in advance. Further, the program executed by the CPU 21 may be a program that is processed in chronological order in the order described in this specification, or may be a program that is processed in parallel or at a necessary timing such as when a call is made. It may be a program in which processing is performed.
<情報処理装置11の機能ブロック図>
 図2は、図1の情報処理装置11の機能を説明する機能ブロック図である。
<Functional block diagram of information processing device 11>
FIG. 2 is a functional block diagram illustrating the function of the information processing device 11 of FIG.
 図2において、情報処理装置11は、学習コンテンツ制御部41と、学習環境制御部42とを有する。 In FIG. 2, the information processing device 11 has a learning content control unit 41 and a learning environment control unit 42.
 学習コンテンツ制御部41は、ユーザの学習の理解度等に応じた問題をユーザに提供するための制御を行う。 The learning content control unit 41 controls to provide the user with a problem according to the degree of understanding of the user's learning.
 学習環境制御部42は、ユーザの学習に対する集中度等の学習状態の向上を図るためにユーザが学習を行う学習空間の環境(学習環境)の制御を行う。 The learning environment control unit 42 controls the environment (learning environment) of the learning space in which the user learns in order to improve the learning state such as the degree of concentration of the user on learning.
(学習コンテンツ制御部41)
 学習コンテンツ制御部41は、ユーザインタフェース部61と、ユーザ状態センシング部62と、学習データ解析部63と、問題生成部64とを有する。
(Learning content control unit 41)
The learning content control unit 41 includes a user interface unit 61, a user state sensing unit 62, a learning data analysis unit 63, and a problem generation unit 64.
 ユーザインタフェース部61は、図1の入力部24及び出力部25を含み、ユーザに情報を提示し、また、ユーザからの情報を受け付ける。 The user interface unit 61 includes the input unit 24 and the output unit 25 of FIG. 1, presents information to the user, and receives information from the user.
 ユーザインタフェース部61は、問題生成部64からの問題qを出力部25によりユーザに提示する。また、ユーザが問題qに対する答案等の情報を入力部24から入力すると、ユーザインタフェース部61は、答案及び解答時間の情報を含む答案情報Rを学習データ解析部63に供給する。 The user interface unit 61 presents the problem q from the problem generation unit 64 to the user by the output unit 25. Further, when the user inputs information such as an answer to the question q from the input unit 24, the user interface unit 61 supplies the answer information R including the answer and the answer time information to the learning data analysis unit 63.
 ユーザは、学習環境E′の中で、ユーザインタフェース部61により提示された問題生成部64からの問題qを解く。そして、ユーザは、答案をユーザインタフェース部61から入力する。 The user solves the problem q from the problem generation unit 64 presented by the user interface unit 61 in the learning environment E'. Then, the user inputs the answer from the user interface unit 61.
 なお、解答時間は、ユーザが問題qを解くために要した時間であり、ユーザが解答時間をユーザインタフェース部61から入力してもよいし、ユーザインタフェース部61によりユーザに問題qが提示されてから答案がユーザインタフェース部61から入力されるまでの時間に基づいて学習データ解析部63が算出してもよい。 The answer time is the time required for the user to solve the problem q, and the user may input the answer time from the user interface unit 61, or the user interface unit 61 presents the problem q to the user. The learning data analysis unit 63 may calculate based on the time until the answer is input from the user interface unit 61.
 ユーザ状態センシング部62は、図1の各種センサ13の一部のセンサを含む。ユーザ状態センシング部62は、ユーザインタフェース部61を使用して学習を行っているユーザの状態(ユーザ状態G)をセンシングして学習データ解析部63及び学習環境制御部42の環境データ解析部82に供給する。 The user state sensing unit 62 includes some of the sensors 13 of the various sensors 13 shown in FIG. The user state sensing unit 62 senses the state (user state G) of the user who is learning using the user interface unit 61, and causes the learning data analysis unit 63 and the environment data analysis unit 82 of the learning environment control unit 42 to sense the state (user state G). Supply.
 なお、ユーザの実際のユーザ状態G′に対して、ユーザ状態センシング部62によりセンシングされるユーザ状態を、各センサの測定誤差などを考慮してユーザ状態Gで表す。 Note that the user state sensed by the user state sensing unit 62 with respect to the actual user state G'of the user is represented by the user state G in consideration of the measurement error of each sensor.
 また、ユーザ状態センシング部62は、CPU21の演算処理機能を含み、各種センサ13から直接得られる信号に対してCPU21が所定の信号処理を施した情報をユーザ状態Gとして取得してもよい。 Further, the user state sensing unit 62 may include the arithmetic processing function of the CPU 21, and may acquire the information obtained by the CPU 21 performing predetermined signal processing on the signals directly obtained from the various sensors 13 as the user state G.
 図3は、ユーザ状態センシング部62がユーザ状態のセンシングに使用し得るセンサの種類とセンサにより得られる情報(センシングの目的)とを例示した図である。 FIG. 3 is a diagram illustrating the types of sensors that the user state sensing unit 62 can use for sensing the user state and the information (purpose of sensing) obtained by the sensors.
 図3において、ユーザ状態センシング部62がユーザ状態のセンシングに使用し得るセンサの種類として、GPS(Global Positioning System)、カメラ、人感センサ、マイク、生体情報センサ、深度センサ、加速度センサ、及び角速度センサ等がある。 In FIG. 3, GPS (Global Positioning System), a camera, a motion sensor, a microphone, a biometric information sensor, a depth sensor, an acceleration sensor, and an angular speed are used as types of sensors that the user state sensing unit 62 can use for user state sensing. There are sensors, etc.
 GPSは、ユーザがGPSを携帯し、または、装着している場合に、ユーザの位置をセンシングする。ユーザの位置をセンシングすることによりユーザが学習をしている場所(自宅又は外等)の他に、ユーザが同じ位置にいるのか、移動したのか等のユーザの行動が把握され得る。なお、GPSは、スマートフォンなどの携帯端末に搭載されているものであってよい。 GPS senses the user's position when the user carries or wears GPS. By sensing the user's position, the user's behavior such as whether the user is in the same position or moved can be grasped in addition to the place where the user is learning (home or outside, etc.). The GPS may be mounted on a mobile terminal such as a smartphone.
 カメラは、ユーザの学習空間を撮影する1台又は複数台のカメラである。ユーザ状態センシング部62は、カメラから得られた映像に基づいて学習空間におけるユーザの位置をセンシングする。また、ユーザの位置をセンシングすることによりユーザの行動が把握され得る。また、カメラから得られた映像から顔の動きなどのユーザの微細な行動も把握され得る。 The camera is one or more cameras that capture the user's learning space. The user state sensing unit 62 senses the position of the user in the learning space based on the image obtained from the camera. In addition, the user's behavior can be grasped by sensing the user's position. In addition, the user's minute behavior such as facial movement can be grasped from the image obtained from the camera.
 人感センサは、学習空間に配備され、赤外線等を用いて学習空間におけるユーザの位置をセンシングする。また、ユーザの位置をセンシングすることによりユーザの行動が把握され得る。 The motion sensor is installed in the learning space and senses the position of the user in the learning space using infrared rays or the like. In addition, the user's behavior can be grasped by sensing the user's position.
 マイクは、学習空間に配備され、ユーザの声をセンシングする。ユーザの声をセンシングすることにより、ユーザの疲労の状態等が把握され得る。 The microphone is installed in the learning space and senses the user's voice. By sensing the user's voice, the state of fatigue of the user can be grasped.
 生体情報センサは、ユーザの脈拍、発汗、脳波、触覚、嗅覚、又は、味覚等の生体的な状態をセンシングする。ユーザの脈拍、発汗、脳波をセンシングすることにより、ユーザの学習への集中度等が把握され得る。 The biological information sensor senses a biological state such as a user's pulse, sweating, brain wave, touch, smell, or taste. By sensing the user's pulse, sweating, and brain waves, the degree of concentration of the user on learning can be grasped.
 また、ユーザの触覚、嗅覚、又は、味覚をセンシングするとは、ユーザの触覚、嗅覚、又は、味覚への意識がどの程度働いているかをセンシングすることを意味する。ユーザの触覚、嗅覚、又は、味覚をセンシングすることにより、ユーザの学習への集中度等が把握され得る。 Also, sensing the user's sense of touch, smell, or taste means sensing how much the user's sense of touch, smell, or taste is working. By sensing the user's sense of touch, smell, or taste, the degree of concentration of the user on learning can be grasped.
 深度センサは、ユーザの学習空間における深度情報(奥行方向を含めた3次元的な位置)をセンシングする。深度情報をセンシングすることにより、ユーザの3次元的な位置や行動が把握され得る。 The depth sensor senses depth information (three-dimensional position including the depth direction) in the user's learning space. By sensing the depth information, the user's three-dimensional position and behavior can be grasped.
 加速度センサは、ユーザが加速度センサを携帯し、または、装着している場合に、ユーザの加速度をセンシングする。ユーザの加速度をセンシングすることによりユーザの位置の移動の他に位置の移動を伴わない姿勢の変化のような微細な行動(動作)が把握され得る。なお、加速度センサはユーザが携帯するスマートフォンなどの携帯端末に搭載されていてもよい。 The acceleration sensor senses the user's acceleration when the user carries or wears the acceleration sensor. By sensing the acceleration of the user, it is possible to grasp minute actions (movements) such as a change in posture that does not accompany the movement of the position in addition to the movement of the position of the user. The acceleration sensor may be mounted on a mobile terminal such as a smartphone carried by the user.
 角速度センサは、ユーザが角速度センサを携帯し、または、装着している場合に、ユーザの角速度をセンシングする。ユーザの角速度をセンシングすることによりユーザの向きを変える微細な行動(動作)も把握され得る。 The angular velocity sensor senses the user's angular velocity when the user carries or wears the angular velocity sensor. By sensing the user's angular velocity, it is possible to grasp minute actions (movements) that change the direction of the user.
 なお、ユーザ状態センシング部62は、図3に示した全ての種類のセンサを有していなくてもよく、また、ユーザ状態をセンシングするセンサであれば他の種類のセンサを有していてもよい。また、検出するユーザ状態は、ユーザの位置、行動、向き、脈拍、発汗、脳波、触覚、嗅覚、及び、味覚に関する状態のうちのいずれか1つ又は複数の状態であってよい。 The user state sensing unit 62 does not have to have all the types of sensors shown in FIG. 3, and may have other types of sensors as long as it is a sensor that senses the user state. Good. Further, the user state to be detected may be any one or more of the user's position, behavior, orientation, pulse, sweating, brain wave, touch, smell, and taste.
 図2において、学習データ解析部63は、図1のCPU21の演算処理により実現される機能ブロックであり、ユーザインタフェース部61からの答案情報Rと、ユーザ状態センシング部62からのユーザ状態Gとに基づいて、ユーザの学習に対する集中度、理解度(学習度)、習得速度等の学習状態(学習に対する質の良さ)を解析する。学習データ解析部63は、解析したユーザの学習状態を表す解析結果AIを問題生成部64に供給する。 In FIG. 2, the learning data analysis unit 63 is a functional block realized by the arithmetic processing of the CPU 21 of FIG. 1, and includes the answer information R from the user interface unit 61 and the user state G from the user state sensing unit 62. Based on this, the learning state (good quality for learning) such as the user's concentration on learning, comprehension (learning degree), and learning speed is analyzed. The learning data analysis unit 63 supplies the analysis result AI representing the learning state of the analyzed user to the problem generation unit 64.
 ユーザの学習に対する集中度は、ユーザの集中力の指標を意味する。学習データ解析部63は、例えば、ユーザ状態センシング部62からのユーザ状態Gに基づいて集中度を求めることができ、特にユーザの生態情報から求めてもよい。また、学習データ解析部63は、ユーザが学習を開始してからの時間や、時刻、問題に対する正解率、解答への所要時間の推移などから集中度を求めてもよい。 The degree of concentration of the user on learning means an index of the user's concentration. The learning data analysis unit 63 can obtain the degree of concentration based on the user state G from the user state sensing unit 62, for example, and may particularly obtain it from the user's ecological information. Further, the learning data analysis unit 63 may obtain the degree of concentration from the time since the user starts learning, the time, the correct answer rate for the problem, the transition of the time required for the answer, and the like.
 ユーザの学習に対する理解度は、ユーザの所定の学習領域に対する理解度の指標を意味する。学習領域とは、学年、科目、単元等の学習単位からなる学習の範囲をいう。学習データ解析部63は、問題の正解率、解答への所要時間、問題に対するアンケート結果に基づいて理解度を求める。 The user's understanding of learning means an index of the user's understanding of a predetermined learning area. The learning area refers to the range of learning consisting of learning units such as grades, subjects, and units. The learning data analysis unit 63 obtains the degree of understanding based on the correct answer rate of the question, the time required for the answer, and the questionnaire result for the question.
 ユーザの学習に対する習得速度は、ユーザが一度学んだ問題に対する理解の速さの指標を意味する。学習データ解析部63は、一度間違った問題を解き直せたかどうか(解き直せた場合はその解答への所要時間)、正解した問題に関する類題を解けたかどうか(解けた場合はその解答への所要時間)、間違った問題を再度解答した時の正解率や解答への所要時間等に基づいて習得速度を求める。 The learning speed for the user's learning means an index of the speed of understanding the problem once learned by the user. Whether or not the learning data analysis unit 63 was able to solve the wrong problem once (if it was possible to solve it, the time required for the answer), and whether or not it was able to solve a similar problem related to the correct answer (if it was possible, the time required for the answer). ), Find the learning speed based on the correct answer rate when answering the wrong question again and the time required to answer.
 また、学習データ解析部63は、解析した学習状態(解析結果AI)の一部又は全ての情報を学習情報Cとして学習環境制御部42の環境データ解析部82に供給する。図2では、学習データ解析部63は、解析したユーザの集中度、理解度、及び、習得速度のうちの集中度及び理解度を学習情報Cとして環境データ解析部82に供給する。 Further, the learning data analysis unit 63 supplies a part or all of the analyzed learning state (analysis result AI) as learning information C to the environment data analysis unit 82 of the learning environment control unit 42. In FIG. 2, the learning data analysis unit 63 supplies the analyzed user's concentration, comprehension, and concentration and comprehension of the learning speed to the environment data analysis unit 82 as learning information C.
 ただし、学習データ解析部63は、集中度、理解度、及び、習得速度のうちの全ての情報、又は、いずれか1つ若しくは2つの情報を学習情報Cとして環境データ解析部82に供給してもよい。 However, the learning data analysis unit 63 supplies all the information of the degree of concentration, the degree of understanding, and the acquisition speed, or any one or two pieces of information as learning information C to the environmental data analysis unit 82. May be good.
 また、学習データ解析部63から環境データ解析部82に供給される学習情報Cは、ユーザの現在の学習状態(学習に対する質の良さ)を表す情報であれば、集中度、理解度、及び、習得速度以外の情報であってもよい。 Further, if the learning information C supplied from the learning data analysis unit 63 to the environment data analysis unit 82 is information representing the user's current learning state (good quality for learning), the degree of concentration, the degree of understanding, and the degree of understanding, and Information other than the learning speed may be used.
 問題生成部64は、学習データ解析部63からの解析結果AIに基づいて、ユーザの学習状態に応じた問題、例えば、ユーザの学習状態に応じた難易度の問題を生成し、ユーザインタフェース部61に供給する。 The problem generation unit 64 generates a problem according to the learning state of the user, for example, a problem having a difficulty level according to the learning state of the user, based on the analysis result AI from the learning data analysis unit 63, and the user interface unit 61. Supply to.
 ここで、学習コンテンツ制御部41には、例えば、特許文献1(国際公開第2016/088463号)に記載された技術を適用してもよい。 Here, for example, the technique described in Patent Document 1 (International Publication No. 2016/0884663) may be applied to the learning content control unit 41.
 ただし、特許文献1の技術は、図2の学習環境制御部42を有してない。そのため、特許文献1の技術では、本技術を適用した情報処理装置11のように学習環境に対する制御により学習効率の向上を図ることは行われない。 However, the technique of Patent Document 1 does not have the learning environment control unit 42 of FIG. Therefore, the technique of Patent Document 1 does not improve the learning efficiency by controlling the learning environment as in the information processing device 11 to which the present technique is applied.
 また、ユーザにとって快適な環境となるように室内の機器を制御しても、必ずしも学習効率が向上するとは限らない。一方、本技術を適用した情報処理装置11では、学習環境制御部42が、学習コンテンツ制御部41からユーザの学習状態を含む学習情報Cを取得し、学習状態が向上するように学習環境の温度などを変更するため学習効率の向上を適切に図ることができる。 Also, controlling the equipment in the room so that the environment is comfortable for the user does not necessarily improve the learning efficiency. On the other hand, in the information processing device 11 to which the present technology is applied, the learning environment control unit 42 acquires the learning information C including the learning state of the user from the learning content control unit 41, and the temperature of the learning environment is improved so that the learning state is improved. It is possible to improve the learning efficiency appropriately because such changes are made.
(学習環境制御部42)
 図2において、学習環境制御部42は、学習環境センシング部81と、環境データ解析部82と、環境制御部83とを有する。
(Learning environment control unit 42)
In FIG. 2, the learning environment control unit 42 includes a learning environment sensing unit 81, an environmental data analysis unit 82, and an environment control unit 83.
 学習環境センシング部81は、図1の各種センサ13の一部のセンサを含む。学習環境センシング部81は、ユーザの学習環境をセンシングして学習環境の現在の状態を表す環境情報Eを環境データ解析部82に供給する。 The learning environment sensing unit 81 includes some of the sensors 13 of the various sensors 13 shown in FIG. The learning environment sensing unit 81 senses the learning environment of the user and supplies the environment information E representing the current state of the learning environment to the environment data analysis unit 82.
 なお、学習環境センシング部81は、CPU21の演算処理機能を含み、各種センサ13から直接得られる信号に対してCPU21が所定の信号処理を施した情報を環境情報Eとして取得してもよい。 The learning environment sensing unit 81 may include the arithmetic processing function of the CPU 21 and acquire the information obtained by the CPU 21 performing predetermined signal processing on the signals directly obtained from the various sensors 13 as the environment information E.
 図4は、学習環境センシング部81がセンシングする学習環境の要素と、センサの種類とを例示した図である。 FIG. 4 is a diagram illustrating the elements of the learning environment sensed by the learning environment sensing unit 81 and the types of sensors.
 図4において、学習環境センシング部81がセンシングする学習環境の要素として、音、映像、照度、温度、湿度、気圧、窓や扉の開閉状態、部屋の散らかり具合、他者の有無、天候、及び、時間(学習時間、時刻)が示されている。 In FIG. 4, as elements of the learning environment sensed by the learning environment sensing unit 81, sound, image, illuminance, temperature, humidity, atmospheric pressure, open / closed state of windows and doors, clutter of the room, presence / absence of others, weather, and , Time (learning time, time) is shown.
 音のセンシングでは、例えば、学習空間に配備されたマイクにより、学習空間のノイズ等の音の大きさが検出される。 In sound sensing, for example, the loudness of sound such as noise in the learning space is detected by a microphone installed in the learning space.
 映像のセンシングでは、例えば、映像表示機器(ディスプレイ等)から、または、映像表示機器に映像を供給する映像出力機器から、映像の表示を行っているか否かの情報が取得され、学習空間に配備された映像表示機器に映像が表示されているか否かが検出される。 In video sensing, for example, information on whether or not video is being displayed is acquired from a video display device (display, etc.) or from a video output device that supplies video to the video display device, and is deployed in the learning space. It is detected whether or not the image is displayed on the displayed image display device.
 照度のセンシングでは、例えば、学習空間に配備された照度センサにより、学習空間の照度の高さが検出される。また、照度センサを用いる代わりに、学習空間を照明する照明機器における照度の設定値の情報を取得することにより、学習空間の照度の高さが検出されてもよい。また、学習空間を撮影するカメラからの映像を解析することにより、学習空間の照度の高さが検出されてもよい。 In illuminance sensing, for example, the height of illuminance in the learning space is detected by an illuminance sensor installed in the learning space. Further, instead of using the illuminance sensor, the height of the illuminance in the learning space may be detected by acquiring the information of the set value of the illuminance in the lighting device that illuminates the learning space. Further, the high illuminance of the learning space may be detected by analyzing the image from the camera that captures the learning space.
 温度のセンシングでは、学習空間の空調機器に内蔵された温度センサ、または、空調機器とは別に学習空間に配備された温度センサにより、学習空間の温度の高さが検出される。 In temperature sensing, the high temperature of the learning space is detected by the temperature sensor built into the air conditioning equipment in the learning space or the temperature sensor installed in the learning space separately from the air conditioning equipment.
 湿度のセンシングでは、学習空間の空調機器に内蔵された湿度センサ、または、空調機器とは別に学習空間に配備された湿度センサにより、学習空間の湿度の高さが検出される。 In humidity sensing, the high humidity of the learning space is detected by the humidity sensor built into the air conditioning equipment in the learning space or the humidity sensor installed in the learning space separately from the air conditioning equipment.
 気圧のセンシングでは、学習空間の空調機器に内蔵された気圧センサ、または、空調機器とは別に学習空間に配備された気圧センサにより、学習空間の気圧の高さが検出される。 In atmospheric pressure sensing, the height of atmospheric pressure in the learning space is detected by the atmospheric pressure sensor built into the air conditioning equipment in the learning space or the atmospheric pressure sensor installed in the learning space separately from the air conditioning equipment.
 窓や扉の開閉状態のセンシングでは、学習空間を撮影するカメラからの映像に基づいて、または、窓や扉に設置された開閉センサにより、学習空間を遮蔽する窓や扉の開閉状態が検出される。 In the sensing of the open / closed state of the window or door, the open / closed state of the window or door that shields the learning space is detected based on the image from the camera that captures the learning space or by the open / close sensor installed on the window or door. The door.
 部屋の散らかり具合のセンシングでは、学習環境の空間を撮影するカメラからの映像に基づいて部屋の散らかり具合が検出される。 In the sensing of the clutter of the room, the clutter of the room is detected based on the image from the camera that captures the space of the learning environment.
 他者の有無のセンシングでは、学習空間を撮影するカメラからの映像に基づいて学習空間に複数の人物が存在するか否かが解析されて他者の有無が検出される。なお、学習空間に配備された人感センサ、または、深度センサにより他者の有無が検出されてもよい。 In the sensing of the presence or absence of others, the presence or absence of others is detected by analyzing whether or not there are multiple persons in the learning space based on the image from the camera that captures the learning space. The presence or absence of others may be detected by a motion sensor or a depth sensor installed in the learning space.
 天候のセンシングでは、湿度センサ、温度センサ、及び、照度センサからの情報に基づいて天候が晴れ、曇り、又は、雨であるかが検出される。なお、天候の情報は、インターネットのサイト等から取得されてもよい。 In weather sensing, it is detected whether the weather is sunny, cloudy, or rainy based on the information from the humidity sensor, temperature sensor, and illuminance sensor. The weather information may be obtained from an internet site or the like.
 時間のセンシングでは、情報処理部12が内蔵する時計機能またはインターネットの特定のサーバから時刻情報が取得され、学習時間(学習開始からの経過時間)や現在時刻が検出される。 In time sensing, time information is acquired from the clock function built in the information processing unit 12 or a specific server on the Internet, and the learning time (elapsed time from the start of learning) and the current time are detected.
 なお、学習環境センシング部81がセンシングする学習環境の要素は、図4の学習環境の要素のうちのいずれか1つ又は複数の要素であってもよい。 The learning environment element sensed by the learning environment sensing unit 81 may be any one or a plurality of elements of the learning environment shown in FIG.
 図2において環境データ解析部82は、図1のCPU21の演算処理により実現される機能ブロックであり、学習環境センシング部81からの環境情報Eと、ユーザ状態センシング部62からのユーザ状態Gと、学習データ解析部63からの学習情報Cとに基づいて、学習環境がユーザの学習状態に与える影響を解析する。 In FIG. 2, the environment data analysis unit 82 is a functional block realized by the arithmetic processing of the CPU 21 of FIG. 1, and includes the environment information E from the learning environment sensing unit 81, the user state G from the user state sensing unit 62, and the user state G. Based on the learning information C from the learning data analysis unit 63, the influence of the learning environment on the learning state of the user is analyzed.
 そして、環境データ解析部82は、解析の結果に基づいて、学習環境の変更内容であって、ユーザの学習状態が現在よりも向上する変更内容を算出する。環境データ解析部82は、ユーザの学習状態が向上する学習環境へと変更するための学習環境に対する制御内容(学習環境の変更内容)を解析結果Aeとして環境制御部83に供給する。 Then, the environment data analysis unit 82 calculates the change contents of the learning environment based on the analysis result, in which the learning state of the user is improved from the present. The environment data analysis unit 82 supplies the control content (change content of the learning environment) for the learning environment for changing to the learning environment in which the learning state of the user is improved to the environment control unit 83 as the analysis result Ae.
 環境制御部83は、図1のCPU21の演算処理機能と、各種環境制御機器14とを含む。環境制御部83は、環境データ解析部82からの解析結果Aeに基づいて図2の各種環境制御機器14を制御し、学習環境E′を制御する。 The environmental control unit 83 includes the arithmetic processing function of the CPU 21 shown in FIG. 1 and various environmental control devices 14. The environment control unit 83 controls the various environment control devices 14 shown in FIG. 2 based on the analysis result Ae from the environment data analysis unit 82, and controls the learning environment E'.
 図5は、環境制御部83が制御する学習環境の要素と、各要素の制御に用いられる環境制御機器の種類とを例示した図である。 FIG. 5 is a diagram illustrating the elements of the learning environment controlled by the environment control unit 83 and the types of environment control devices used to control each element.
 図5には環境制御部83が制御する学習環境の要素として、ノイズ、音楽、照度、温度、湿度、情報提示(映像)、連絡、及び、部屋の散らかり具合が例示されている。ノイズと音楽とは、いずれも学習環境の音に関する要素であるが、学習環境の制御では別の要素としている。 FIG. 5 illustrates noise, music, illuminance, temperature, humidity, information presentation (video), communication, and room clutter as elements of the learning environment controlled by the environment control unit 83. Noise and music are both elements related to the sound of the learning environment, but they are separate elements in the control of the learning environment.
 ノイズの制御では、学習空間に配備されたスピーカやユーザが装着するヘッドフォンに接続されたオーディオ機器により、ノイズキャンセルを行うか否かが制御される。オーディオ機器は、ノイズキャンセルを行う場合にはノイズと逆位相の音をスピーカ又はヘッドフォンに供給する。 In noise control, whether or not noise cancellation is performed is controlled by an audio device connected to a speaker installed in the learning space or a headphone worn by the user. When canceling noise, the audio device supplies a sound having a phase opposite to the noise to the speaker or headphones.
 音楽の制御では、学習空間に配備されたスピーカやユーザが装着するヘッドフォンに接続されたオーディオ機器により、音楽を出力するか否かや、音楽の音量が制御される。なお、音楽の制御では、音楽のジャンルなどに基づいて選曲が制御されてもよい。 In music control, whether or not to output music and the volume of music are controlled by the speakers installed in the learning space and the audio equipment connected to the headphones worn by the user. In music control, music selection may be controlled based on the genre of music or the like.
 照度の制御では、学習空間に配備された照明機器により、学習空間の照度の高さが制御される。 In illuminance control, the height of illuminance in the learning space is controlled by the lighting equipment installed in the learning space.
 温度の制御では、空調機器により、学習空間の温度の高さが制御される。 In temperature control, the temperature of the learning space is controlled by the air conditioner.
 湿度の制御では、空調機器により、学習空間の湿度の高さが制御される。 In humidity control, the humidity level of the learning space is controlled by the air conditioner.
 情報提示(映像)の制御では、学習空間に配備された映像表示機器(ディスプレイに映像を出力する機器)により、映像を表示するか否かが制御される。映像はユーザの理解度を深めるための情報であってもよいし、集中度を高める情報等であってもよい。 In the control of information presentation (video), whether or not to display the video is controlled by the video display device (device that outputs the video to the display) installed in the learning space. The video may be information for deepening the user's understanding, or may be information for increasing the concentration.
 連絡(他者の介入)の制御では、通信端末により、ユーザに対するメールや電話等での他者からの連絡を遮断するか否か、または、ユーザに対して連絡を行わないように事前に登録された他者に対してアナウンスするか否かが制御される。また、連絡の制御では、学習空間への入口の扉の鍵を制御する機器により、扉の鍵を施錠するか解錠するかが制御される。 In the control of contact (intervention of others), whether or not to block the contact from others by e-mail or telephone to the user by the communication terminal, or to register in advance so as not to contact the user. Whether or not to announce to others who have been made is controlled. Further, in the control of communication, whether the door key is locked or unlocked is controlled by the device that controls the door key at the entrance to the learning space.
 部屋の散らかり具合の制御では、掃除を行うロボットや片付けを行うロボットなどにより、学習空間の清掃や整頓を行うか否かが制御される。 In the control of how cluttered the room is, whether or not to clean or tidy up the learning space is controlled by a robot that cleans or a robot that cleans up.
<情報処理装置11の処理例>
 図6は、図2の情報処理装置11が行う処理例を説明するフローチャートである。
<Processing example of information processing device 11>
FIG. 6 is a flowchart illustrating a processing example performed by the information processing apparatus 11 of FIG.
 図6において、ステップS11乃至ステップS14は学習コンテンツ制御部41が行う学習コンテンツ制御の処理例を示し、ステップS15乃至ステップS17は学習環境制御部42が行う学習環境制御の処理例を示す。 In FIG. 6, steps S11 to S14 show a processing example of learning content control performed by the learning content control unit 41, and steps S15 to S17 show a processing example of learning environment control performed by the learning environment control unit 42.
 ステップS11では、ユーザインタフェース部61は、ユーザが学習環境E′の中で問題qを解いて、答案(解答)を入力すると、そのユーザの答案を受け付ける。そして、ユーザインタフェース部61は、ユーザの答案及び解答時間を含む答案情報Rを学習データ解析部63に供給する。処理はステップS11からステップS12に進む。 In step S11, when the user solves the question q in the learning environment E'and inputs the answer (answer), the user interface unit 61 accepts the user's answer. Then, the user interface unit 61 supplies the answer information R including the user's answer and the answer time to the learning data analysis unit 63. The process proceeds from step S11 to step S12.
 ステップS12では、ユーザ状態センシング部62は、学習を行っているユーザのユーザ状態G′をセンシングしてユーザ状態Gを取得する。そして、ユーザ状態センシング部62は、取得したユーザ状態Gを学習コンテンツ制御部41の学習データ解析部63及び学習環境制御部42の環境データ解析部82に供給する。処理はステップS12からステップS13に進む。 In step S12, the user state sensing unit 62 senses the user state G'of the learning user and acquires the user state G. Then, the user state sensing unit 62 supplies the acquired user state G to the learning data analysis unit 63 of the learning content control unit 41 and the environment data analysis unit 82 of the learning environment control unit 42. The process proceeds from step S12 to step S13.
 ステップS13では、学習データ解析部63は、ステップS11でユーザインタフェース部61から供給された答案情報Rと、ステップS12でユーザ状態センシング部62から供給されたユーザ状態Gとを用いて、ユーザの集中度、理解度(学習度)、習得速度等の学習状態を解析してユーザの学習状態を表す解析結果AIを算出する。学習データ解析部63は、算出した解析結果AIを問題生成部64に供給する。 In step S13, the learning data analysis unit 63 uses the answer information R supplied from the user interface unit 61 in step S11 and the user state G supplied from the user state sensing unit 62 in step S12 to concentrate the users. The learning state such as degree, comprehension degree (learning degree), and learning speed is analyzed, and the analysis result AI representing the learning state of the user is calculated. The learning data analysis unit 63 supplies the calculated analysis result AI to the problem generation unit 64.
 また、学習データ解析部63は、解析して得られた学習状態のうち、例えば、集中度及び理解度を学習情報Cとして環境データ解析部82に供給する。処理はステップS13からステップS14に進む。 Further, the learning data analysis unit 63 supplies, for example, the degree of concentration and the degree of understanding as learning information C to the environment data analysis unit 82 among the learning states obtained by the analysis. The process proceeds from step S13 to step S14.
 ステップS14では、問題生成部64は、ステップS13で学習データ解析部63から供給された解析結果AIに基づいて、ユーザの学習状態に応じた問題qを生成する。問題生成部64により生成された問題qはユーザインタフェース部61における出力部25(図1参照)に供給される。処理はステップS14の後、ステップS11に戻り、ステップS11乃至ステップS14を繰り返す。 In step S14, the problem generation unit 64 generates the problem q according to the learning state of the user based on the analysis result AI supplied from the learning data analysis unit 63 in step S13. The problem q generated by the problem generation unit 64 is supplied to the output unit 25 (see FIG. 1) in the user interface unit 61. The process returns to step S11 after step S14, and repeats steps S11 to S14.
 ステップS15では、学習環境制御部42の学習環境センシング部81は、ユーザが学習を行っている学習空間の環境(学習環境)をセンシングし、環境情報Eを取得する。学習環境センシング部81は、取得した環境情報Eを環境データ解析部82に供給する。処理はステップS15からステップS16に進む。 In step S15, the learning environment sensing unit 81 of the learning environment control unit 42 senses the environment (learning environment) of the learning space in which the user is learning, and acquires the environment information E. The learning environment sensing unit 81 supplies the acquired environmental information E to the environmental data analysis unit 82. The process proceeds from step S15 to step S16.
 ステップS16では、環境データ解析部82は、ステップS12でユーザ状態センシング部62から供給されたユーザ状態Gと、ステップS13で学習データ解析部63から供給された学習情報Cと、ステップS15で学習環境センシング部81から供給された環境情報Eを用いて、前回、学習データ解析部63から供給された学習情報Cが示した集中度と比較して今回の集中度が改善されたかを解析する。 In step S16, the environment data analysis unit 82 includes the user state G supplied from the user state sensing unit 62 in step S12, the learning information C supplied from the learning data analysis unit 63 in step S13, and the learning environment in step S15. Using the environmental information E supplied from the sensing unit 81, it is analyzed whether or not the degree of concentration this time is improved as compared with the degree of concentration indicated by the learning information C supplied from the learning data analysis unit 63 last time.
 そして、学習データ解析部63は、学習環境に対する次の変更内容(制御内容)であって集中度が向上する変更内容を解析結果Aeとして算出する。なお、学習データ解析部63は、集中度の改善だけでなく、ユーザの理解度も考慮して学習状態が向上したかを解析して、学習状態が向上する学習環境の変更内容を解析結果Aeとして算出してもよい。処理はステップS16からステップS17に進む。 Then, the learning data analysis unit 63 calculates the next change content (control content) for the learning environment as the analysis result Ae, which improves the degree of concentration. The learning data analysis unit 63 analyzes whether the learning state is improved in consideration of not only the improvement of the degree of concentration but also the degree of understanding of the user, and analyzes the changed contents of the learning environment in which the learning state is improved. It may be calculated as. The process proceeds from step S16 to step S17.
 ステップS17では、環境制御部83は、ステップS16で環境データ解析部82から供給された解析結果Aeに基づいて、ユーザにとって学習に適した学習環境E′となるように学習環境を制御する。処理はステップS17の後、ステップS15に戻り、ステップS15乃至ステップS17を繰り返す。 In step S17, the environment control unit 83 controls the learning environment so that the learning environment E'is suitable for learning for the user, based on the analysis result Ae supplied from the environment data analysis unit 82 in step S16. The process returns to step S15 after step S17, and repeats steps S15 to S17.
(効果)
 以上の情報処理装置11によれば、学習コンテンツ制御部41からの学習情報Cに基づいて、学習環境制御部42がユーザの学習状態が向上するように学習環境を変更するため、学習環境がユーザの学習に適した学習環境に適切に変更され、学習効率の向上が適切に図られる。
(effect)
According to the above information processing device 11, the learning environment is changed to the user because the learning environment control unit 42 changes the learning environment so that the learning state of the user is improved based on the learning information C from the learning content control unit 41. The learning environment will be appropriately changed to be suitable for learning, and the learning efficiency will be improved appropriately.
<学習環境制御の詳細>
 図7は、学習環境制御部42における学習環境制御の処理例の詳細を説明する機能ブロック図である。
<Details of learning environment control>
FIG. 7 is a functional block diagram illustrating details of a processing example of learning environment control in the learning environment control unit 42.
 図7には、図2の環境データ解析部82及び環境制御部83と、図2では不図示の環境制御用データベース101とが示されている。環境制御用データベース101は、図1の記憶部26により実現される機能ブロックであり、環境データ解析部82において参照するデータ等が記憶される。 FIG. 7 shows the environmental data analysis unit 82 and the environmental control unit 83 of FIG. 2, and the environmental control database 101 (not shown) of FIG. The environment control database 101 is a functional block realized by the storage unit 26 of FIG. 1, and stores data and the like referred to by the environment data analysis unit 82.
 さらに、図7には、環境制御部83の構成例が示されており、環境制御部83は、制御信号生成部91と、機器92A乃至機器92Nとを有する。 Further, FIG. 7 shows a configuration example of the environment control unit 83, and the environment control unit 83 includes a control signal generation unit 91 and devices 92A to 92N.
 制御信号生成部91は、図1のCPU21の演算処理により実現される機能ブロックである。制御信号生成部91は、環境データ解析部82からの解析結果Aeに基づいて機器92A乃至機器92Nの各々を制御する制御信号を生成する。環境データ解析部82からの解析結果Aeには、例えば、図5に示した学習環境の各要素についての変更内容(制御内容)の情報が含まれる。 The control signal generation unit 91 is a functional block realized by the arithmetic processing of the CPU 21 of FIG. The control signal generation unit 91 generates a control signal for controlling each of the devices 92A to 92N based on the analysis result Ae from the environment data analysis unit 82. The analysis result Ae from the environment data analysis unit 82 includes, for example, information on the change contents (control contents) for each element of the learning environment shown in FIG.
 また、機器92A乃至機器92Nの各々は、学習環境の各要素の制御に用いられる環境制御機器である。そして、機器92A乃至機器92Nは、各々が変更し得る学習環境の要素に対応付けされている。 Further, each of the devices 92A to 92N is an environmental control device used for controlling each element of the learning environment. The devices 92A to 92N are associated with elements of the learning environment that can be changed by each.
 制御信号生成部91は、環境データ解析部82からの解析結果Aeにより示された制御内容にしたがって学習環境の各要素が変更されるように、学習環境の各要素に対応付けされた各機器92A乃至機器92Nに対する制御信号を生成する。そして、制御信号生成部91は、生成した制御信号を各機器92A乃至機器92Nに供給して、各機器92A乃至機器92Nを制御する。 The control signal generation unit 91 is associated with each element of the learning environment 92A so that each element of the learning environment is changed according to the control content indicated by the analysis result Ae from the environment data analysis unit 82. To generate a control signal for the device 92N. Then, the control signal generation unit 91 supplies the generated control signal to each device 92A to 92N to control each device 92A to 92N.
 図7において例示された機器92A乃至機器92Nのうちの一部の環境制御機器の制御について説明する。 The control of some of the environmental control devices among the devices 92A to 92N exemplified in FIG. 7 will be described.
 機器92Aは、図5に示した学習環境のノイズの制御に用いられる環境制御機器である。また、機器92Aは、例えば、学習空間に配備されたスピーカやユーザが装着するヘッドフォンに対してノイズと逆位相の音を出力してノイズキャンセルを行うオーディオ機器である。制御信号生成部91は、解析結果Aeに含まれる制御内容であって、ノイズキャンセルを行うか否かの制御内容にしたがって機器92Aを制御する。 The device 92A is an environment control device used for controlling noise in the learning environment shown in FIG. Further, the device 92A is, for example, an audio device that cancels noise by outputting a sound having a phase opposite to that of noise to a speaker installed in a learning space or a headphone worn by a user. The control signal generation unit 91 controls the device 92A according to the control content included in the analysis result Ae and whether or not noise cancellation is performed.
 機器92Bは、図5に示した学習環境の音楽の制御に用いられる環境制御機器である。機器92Bは、例えば、学習空間に配備されたスピーカやユーザが装着するヘッドフォンに対して音楽を出力するオーディオ機器である。制御信号生成部91は、解析結果Aeに含まれる制御内容であって、音楽を出力するか否かの制御内容にしたがって機器29Bを制御する。 The device 92B is an environmental control device used for controlling music in the learning environment shown in FIG. The device 92B is, for example, an audio device that outputs music to a speaker installed in a learning space or a headphone worn by a user. The control signal generation unit 91 controls the device 29B according to the control content included in the analysis result Ae and whether or not to output music.
 なお、解析結果Aeとして音楽の音量を上げる又は下げる旨の制御内容が含まれていてもよく、その場合には、制御信号生成部91は機器92Bを制御して音楽の音量を一定量上昇又は下降させる。 The analysis result Ae may include a control content for raising or lowering the volume of the music. In that case, the control signal generation unit 91 controls the device 92B to raise or lower the volume of the music by a certain amount. Lower.
 また、解析結果Aeとして曲(音楽のジャンル等)を変更する制御内容が含まれていてもよく、その場合には、制御信号生成部91は機器92Bを制御して曲を変更する。 Further, the analysis result Ae may include a control content for changing a song (music genre, etc.). In that case, the control signal generation unit 91 controls the device 92B to change the song.
 機器92Cは、図5に示した学習環境の照度の制御に用いられる環境制御機器である。機器92Cは、例えば、学習空間に配備された照明機器である。制御信号生成部91は、解析結果Aeに含まれる制御内容であって、照度の高さを上昇、下降、又は、維持させる旨の制御内容にしたがって機器29Cを制御し、学習環境の照度の高さを一定量上昇若しくは下降、又は、維持させる。 The device 92C is an environmental control device used for controlling the illuminance of the learning environment shown in FIG. The device 92C is, for example, a lighting device installed in a learning space. The control signal generation unit 91 controls the device 29C according to the control content included in the analysis result Ae to increase, decrease, or maintain the height of the illuminance, and the high illuminance of the learning environment. The height is increased, decreased, or maintained by a certain amount.
 機器92Dは、図5に示した学習環境の温度の制御に用いられる環境制御機器であり、例えば、学習空間に配備された空調機器である。制御信号生成部91は、解析結果Aeに含まれる制御内容であって、温度の高さを上昇、下降、又は、維持させる旨の制御内容にしたがって機器29Dを制御し、学習環境の温度の高さを一定量上昇若しくは下降、又は、維持させる。 The device 92D is an environmental control device used for controlling the temperature of the learning environment shown in FIG. 5, and is, for example, an air-conditioning device installed in the learning space. The control signal generation unit 91 controls the device 29D according to the control content included in the analysis result Ae to raise, lower, or maintain the temperature height, and the temperature of the learning environment is high. The temperature is increased, decreased, or maintained by a certain amount.
 機器92Eは、図5に示した情報提示(映像)の制御に用いられる環境制御機器であり、例えば、学習空間に配備された映像表示機器である。制御信号生成部91は、解析結果Aeに含まれる制御内容であって、映像を表示させるか否かの制御内容にしたがって機器29Eを制御し、学習空間に映像を表示させるか否かを制御する。なお、表示する情報(映像)の内容としては、例えば、ユーザの理解度を深めるための情報等がある。 The device 92E is an environmental control device used for controlling the information presentation (video) shown in FIG. 5, and is, for example, a video display device installed in a learning space. The control signal generation unit 91 controls the device 29E according to the control content included in the analysis result Ae and whether or not to display the image, and controls whether or not to display the image in the learning space. .. The content of the information (video) to be displayed includes, for example, information for deepening the user's understanding.
 機器92Nは、ユーザの集中を切らす要因の制御に用いられる環境制御機器であり、例えば、図5に示した連絡の制御に用いられる通信端末である。制御信号生成部91は、解析結果Aeに含まれる制御内容であって、連絡を遮断するか否かの制御内容にしたがって機器92Nを制御する。 The device 92N is an environmental control device used for controlling factors that cause the user to lose concentration, and is, for example, a communication terminal used for controlling communication shown in FIG. The control signal generation unit 91 controls the device 92N according to the control content included in the analysis result Ae and whether or not to cut off the communication.
 例えば、制御信号生成部91は、機器92Nにより、ユーザに対するメールや電話等での他者からの連絡を遮断するか否か、または、ユーザに対して連絡を行わないように事前に登録された他者に対してアナウンスするか否かを制御する。また、ユーザの集中を切らす要因として学習空間への他者の出入りがあり、機器92Nの例として、学習空間への入口の扉の鍵を制御する機器(電気鍵)であってもよい。この場合、制御信号生成部91は、解析結果Aeに含まれる制御内容であって、連絡を遮断するか否かの制御内容にしたがって機器92Nを制御し、扉を施錠するか否かを制御する。 For example, the control signal generation unit 91 is registered in advance by the device 92N whether or not to block the contact from others by e-mail, telephone, etc. to the user, or not to contact the user. Controls whether or not to announce to others. In addition, there is the entry and exit of others into the learning space as a factor that discourages the user's concentration, and as an example of the device 92N, a device (electric key) that controls the key of the entrance door to the learning space may be used. In this case, the control signal generation unit 91 controls the device 92N according to the control content included in the analysis result Ae and whether or not to block the communication, and controls whether or not to lock the door. ..
 なお、図7に示した機器92A乃至92Nは一例であって、環境制御部83は、機器92A乃至機器92Nのうちのいずれか1つ又は複数の機器を有していてもよいし、機器92A乃至機器92N以外の機器を有していてもよい。 The devices 92A to 92N shown in FIG. 7 are examples, and the environmental control unit 83 may have any one or more of the devices 92A to 92N, or the device 92A. Or may have a device other than the device 92N.
 また、図7では省略したが、環境制御部83は、環境制御機器として、学習環境の湿度を制御する機器(例えば空調機器)や、部屋の散らかり具合を制御する機器(ロボット等)を有していてもよい。 Further, although omitted in FIG. 7, the environmental control unit 83 has a device for controlling the humidity of the learning environment (for example, an air conditioner) and a device for controlling the degree of clutter in the room (robot, etc.) as the environmental control device. You may be.
 この場合に、制御信号生成部91は、解析結果Aeに含まれる制御内容であって、湿度の高さを上昇、下降、又は、維持させる旨の制御内容にしたがって学習環境の温度を制御する機器を制御し、学習環境の温度の高さを一定量上昇若しくは下降、又は、維持させる。 In this case, the control signal generation unit 91 is a device that controls the temperature of the learning environment according to the control content included in the analysis result Ae, which is to increase, decrease, or maintain the height of the humidity. To raise, lower, or maintain a certain amount of temperature in the learning environment.
 また、制御信号生成部91は、解析結果Aeに含まれる制御内容であって、部屋の散らかり具合を低減させる旨の制御内容にしたがって部屋の散らかり具合を制御する機器を制御し、部屋の清掃や整頓を行わせるか否かを制御する。 Further, the control signal generation unit 91 controls the device that controls the clutter of the room according to the control content included in the analysis result Ae to reduce the clutter of the room, and cleans the room. Controls whether or not to keep things tidy.
 さらに、環境制御部83は、解析結果Aeにしたがって、学習環境の香りを例えばアロマディフューザにより制御してもよいし、テレビ、ゲーム、又は、スマートフォンのロックを制御して誘惑を低減してもよい。 Further, the environment control unit 83 may control the scent of the learning environment by, for example, an aroma diffuser according to the analysis result Ae, or may control the lock of the TV, the game, or the smartphone to reduce the temptation. ..
 また、環境制御部83は、解析結果Aeにしたがって、固定電話を留守番モード(留守電)に設定してもよいし、インターフォンがならないようにしてもよい。また、環境制御部83は、解析結果Aeにしたがって、エスプレッソマシンを稼働させてユーザに休憩を促してもよし、人型ロボットや動物型ロボットを制御してユーザの応援等を行うようにしてもよい。 Further, the environmental control unit 83 may set the fixed telephone to the answering machine mode (answering machine) according to the analysis result Ae, or may prevent the intercom from being turned on. Further, the environment control unit 83 may operate the espresso machine according to the analysis result Ae to urge the user to take a break, or may control the humanoid robot or the animal robot to support the user. Good.
<環境データ解析部82の処理の詳細>
 環境データ解析部82は、学習環境センシング部81からの環境情報Eと、ユーザ状態センシング部62からのユーザ状態Gと、学習データ解析部63からの学習情報Cとを入力データとし、環境の各要素の制御内容(変更内容)を示す解析結果Aeを出力データとして算出して制御信号生成部91に供給する。
<Details of processing by the environmental data analysis unit 82>
The environment data analysis unit 82 uses the environment information E from the learning environment sensing unit 81, the user state G from the user state sensing unit 62, and the learning information C from the learning data analysis unit 63 as input data, and each of the environments. The analysis result Ae indicating the control content (change content) of the element is calculated as output data and supplied to the control signal generation unit 91.
 環境情報Eとしては、図4に示したように、学習空間における音、映像、照度、温度、湿度、気圧、窓や扉の開閉状態、部屋の散らかり具合、ユーザ以外の他者の有無、天候、及び、時間等に関する情報がある。これらの情報のうちのいずれか1つ又は複数の情報が環境情報Eとして環境データ解析部82に与えられる。 As shown in FIG. 4, the environmental information E includes sound, video, illuminance, temperature, humidity, atmospheric pressure, open / closed state of windows and doors, clutter of the room, presence / absence of others other than the user, and weather as shown in FIG. , And there is information about time etc. Any one or more of these pieces of information are given to the environmental data analysis unit 82 as environmental information E.
 ユーザ状態Gとしては、図3に示したように、ユーザの位置、行動、向き、脈拍、発汗、脳波、触覚、嗅覚、及び、味覚に関する状態がある。これらの状態のうちのいずれか1つ又は複数の状態の情報がユーザ状態Gとして環境データ解析部82に与えられる。 As the user state G, as shown in FIG. 3, there are states related to the user's position, behavior, orientation, pulse, sweating, brain wave, touch, smell, and taste. Information on one or more of these states is given to the environment data analysis unit 82 as the user state G.
 学習情報Cとしては、学習に対するユーザの集中度、理解度、及び、習得速度等の学習状態に関する情報がある。これらの集中度、理解度、及び、習得速度等の学習状態に関する情報のうちの例えば集中度及び理解度に関する情報が学習情報Cとして環境データ解析部82に与えられる。ただし、ユーザの集中度、理解度、及び、習得速度等の学習状態に関する情報のうちのいずれか1つ又は複数の情報が学習情報Cとして環境データ解析部82に与えられる場合であってもよい。 The learning information C includes information on the learning state such as the user's concentration level, comprehension level, and learning speed for learning. Among the information on the learning state such as the degree of concentration, the degree of understanding, and the learning speed, for example, the information on the degree of concentration and the degree of understanding is given to the environmental data analysis unit 82 as the learning information C. However, there may be a case where any one or more information on the learning state such as the degree of concentration, the degree of understanding, and the learning speed of the user is given to the environmental data analysis unit 82 as the learning information C. ..
 環境データ解析部82は、環境情報E、ユーザ状態G、及び、学習情報Cの各々を、過去の値と比較する。例えば、環境情報Eの変化量に対して、ユーザ状態Gや学習情報Cがどの程度変化したかを計算し、学習環境のどの要素がユーザ状態Gや学習情報Cの変化に影響したかというような解析を行う。 The environmental data analysis unit 82 compares each of the environmental information E, the user state G, and the learning information C with the past values. For example, it is calculated how much the user state G and the learning information C have changed with respect to the amount of change in the environment information E, and which element of the learning environment has influenced the change in the user state G and the learning information C. Perform an analysis.
 そして、環境データ解析部82は、解析の結果、ユーザの集中度や理解度等の学習状態を向上(増加)させるための学習環境に対する変更内容(学習環境の各要素に対する変更内容)を算出して解析結果Aeとして制御信号生成部91に供給する。 Then, the environment data analysis unit 82 calculates the change contents (change contents for each element of the learning environment) for the learning environment for improving (increasing) the learning state such as the concentration degree and the comprehension degree of the user as a result of the analysis. As an analysis result Ae, it is supplied to the control signal generation unit 91.
 また、学習環境を変更した結果、学習情報Cが示すユーザの学習状態が低下した場合には、環境データ解析部82は、学習状態への影響が大きい学習環境の要素を制御する環境制御機器に対して前回とは別の制御方針(制御内容)に変更してもよい。例えば、環境データ解析部82は、機器92Bに対して音楽を出力させた結果、ユーザの学習状態が低下した場合に、機器92Bに対して音楽の出力を停止させるのではなく、出力する音楽のジャンルを変更させてもよい。 Further, when the learning state of the user indicated by the learning information C deteriorates as a result of changing the learning environment, the environment data analysis unit 82 becomes an environmental control device that controls the elements of the learning environment that have a large influence on the learning state. On the other hand, the control policy (control content) different from the previous one may be changed. For example, when the learning state of the user deteriorates as a result of outputting music to the device 92B, the environmental data analysis unit 82 does not stop the output of the music to the device 92B, but outputs the music. You may change the genre.
 学習環境を変更した結果、学習情報Cが示すユーザの学習状態が向上した場合には、環境データ解析部82は、現在の学習環境の状態を維持するか、または、現在の学習環境の各要素を所定の範囲内で変更する制御方針を設定する。環境データ解析部82は、設定した制御方針を解析結果Aeとして制御信号生成部91に出力し、ユーザの学習状態をより向上させることを目指す。 When the learning state of the user indicated by the learning information C is improved as a result of changing the learning environment, the environment data analysis unit 82 maintains the state of the current learning environment or each element of the current learning environment. Set a control policy to change within a predetermined range. The environment data analysis unit 82 outputs the set control policy as the analysis result Ae to the control signal generation unit 91, aiming to further improve the learning state of the user.
 また、環境データ解析部82は、ユーザが学習しているときの環境情報E、ユーザ状態G、及び、学習情報Cに対する制御方針、すなわち、解析結果Aeを環境制御用データベース101に保存しておきデータベース化しておく。 Further, the environment data analysis unit 82 stores the control policy for the environment information E, the user state G, and the learning information C when the user is learning, that is, the analysis result Ae in the environment control database 101. Create a database.
 環境データ解析部82は、環境制御用データベース101に保存されている前回までの解析結果Aeを参照することにより、環境情報Eやユーザ状態Gがどのようなパターンの場合でも、速やかにユーザの学習状態を良好な状態に遷移させることができるようになる。 By referring to the analysis results Ae up to the previous time stored in the environment control database 101, the environment data analysis unit 82 promptly learns the user regardless of the pattern of the environment information E and the user state G. It becomes possible to transition the state to a good state.
 また、環境データ解析部82は、学習状態を表す集中度、理解度、及び、習得速度等のうちのいずれかの値をユーザの学習状態を表す評価値としてもよいし、学習状態を表す要素(集中度、理解度、及び、習得速度等)の重み付け平均を評価値としてもよい。 Further, the environmental data analysis unit 82 may use any one of the concentration level, the comprehension level, the learning speed, and the like representing the learning state as an evaluation value representing the learning state of the user, or an element representing the learning state. The weighted average of (concentration ratio, comprehension ratio, learning speed, etc.) may be used as the evaluation value.
 環境データ解析部82は、学習データ解析部63から供給される学習情報Cに基づいて評価値を算出することにより、評価値が大きい程、学習状態が良好であると判定することができる。 By calculating the evaluation value based on the learning information C supplied from the learning data analysis unit 63, the environmental data analysis unit 82 can determine that the larger the evaluation value, the better the learning state.
 尚、環境データ解析部82は、環境制御部83により変更可能な学習環境の要素ごとに状態を変更させて評価値が最大となる状態に設定するようにしてもよい。 The environment data analysis unit 82 may change the state for each element of the learning environment that can be changed by the environment control unit 83 to set the state in which the evaluation value is maximized.
 また、環境データ解析部82は、学習環境の変更内容(制御方針である解析結果Ae)の算出処理に関して、ディープラーニングやその他の機械学習を用いて学習させたディープニューラルネットワーク(DNN:Deep Neural Network)の学習モデルを使用してもよい。機械学習における学習方法は、教師あり学習であってもよいし、強化学習であってもよい。以下において、学習モデルを用いて解析結果Aeを算出する場合の処理について説明する。 In addition, the environment data analysis unit 82 uses deep learning or other machine learning to learn the calculation process of the change content (analysis result Ae, which is the control policy) of the learning environment. ) Learning model may be used. The learning method in machine learning may be supervised learning or reinforcement learning. In the following, the process for calculating the analysis result Ae using the learning model will be described.
(環境データ解析部82の第1の処理例)
 環境データ解析部82の第1の処理例として、機械学習(ディープラーニング)を使用して教師あり学習により学習(訓練)させた学習モデル(DNN)を使用する場合について説明する。
(First processing example of the environmental data analysis unit 82)
As a first processing example of the environmental data analysis unit 82, a case where a learning model (DNN) trained (trained) by supervised learning using machine learning (deep learning) is used will be described.
 環境データ解析部82は、学習環境センシング部81から供給された環境情報Eと、ユーザ状態センシング部62から供給されたユーザ状態Gと、学習データ解析部63から供給された学習情報Cとが入力データとして入力されると、学習環境に対する適切な制御方針(ユーザの学習状態が向上する学習環境の変更内容)を示す出力データを出力する学習モデルを用いて解析結果Aeを算出する。 The environment data analysis unit 82 inputs the environment information E supplied from the learning environment sensing unit 81, the user state G supplied from the user state sensing unit 62, and the learning information C supplied from the learning data analysis unit 63. When input as data, the analysis result Ae is calculated using a learning model that outputs output data indicating an appropriate control policy for the learning environment (changes in the learning environment that improve the learning state of the user).
 ここで、制御方針とは、変更(制御)する学習環境の全ての要素についての変更内容(制御内容)を表し、各要素に対して採り得る制御内容の全ての組み合わせの数の制御方針が存在する。 Here, the control policy represents the change content (control content) for all the elements of the learning environment to be changed (controlled), and there is a control policy for the number of all combinations of the control content that can be adopted for each element. To do.
 例えば、制御する学習環境の要素として、温度と湿度があり、温度と湿度の制御内容としてそれぞれ上昇、下降、維持の3通りが存在すると仮定する。この場合において、温度と湿度との制御内容の組み合わせとして、(温度を上げる、かつ、湿度を上げる)、(温度を上げる、かつ、湿度を下げる)、(温度を上げる、かつ、湿度を維持する)、・・・、(温度を維持する、かつ、湿度を維持する)のように9通りの組み合わせが存在する。温度と湿度との制御内容について、それらの9通りの組み合わせの各々が1つの制御方針であり、全部で9通りの制御方針が存在する。 For example, it is assumed that there are temperature and humidity as elements of the learning environment to be controlled, and there are three types of temperature and humidity control contents, rising, falling, and maintaining, respectively. In this case, as a combination of the control contents of the temperature and the humidity, (increase the temperature and increase the humidity), (increase the temperature and decrease the humidity), (increase the temperature and maintain the humidity). ), ..., (Maintaining temperature and maintaining humidity), there are nine combinations. Regarding the control contents of temperature and humidity, each of these nine combinations has one control policy, and there are a total of nine control policies.
 また、学習モデルは全ての制御方針の各々に対応する出力ノードを有し、各出力ノードからは例えば0から1までの範囲の値が出力される。そして、各制御方針に対応する出力ノードからの出力値は、各制御方針の適性度を表す。 In addition, the learning model has output nodes corresponding to each of all control policies, and each output node outputs, for example, a value in the range of 0 to 1. Then, the output value from the output node corresponding to each control policy represents the appropriateness of each control policy.
 環境データ解析部82は、学習モデルの出力ノードから出力された適性度が最大となった制御方針をユーザの学習状態(例えば、集中度)を向上させる制御方針として決定する。また、環境データ解析部82は、決定した制御方針を解析結果Aeとして環境制御部83の制御信号生成部91に供給する。 The environmental data analysis unit 82 determines the control policy that maximizes the aptitude output from the output node of the learning model as the control policy that improves the learning state (for example, the degree of concentration) of the user. Further, the environmental data analysis unit 82 supplies the determined control policy as the analysis result Ae to the control signal generation unit 91 of the environmental control unit 83.
 なお、学習モデルへの入力データは、環境情報E、ユーザ状態G、及び、学習情報Cのうちのいずれか1つ又は複数の情報であってもよい。 The input data to the learning model may be any one or more of the environment information E, the user state G, and the learning information C.
 環境データ解析部82は、ユーザに特化した学習モデルが機械学習(ディープラーニング)により生成されるまでは、環境制御用データベース101に予め保存された学習済みの学習モデル(学習データ収集用の学習モデル)を用いて学習データを収集する。 The environment data analysis unit 82 is a learned learning model (learning for learning data collection) stored in advance in the environment control database 101 until a user-specific learning model is generated by machine learning (deep learning). The training data is collected using the model).
 学習データ収集用の学習モデルは、ユーザとは無関係の学習モデル(例えば未学習の学習モデル)であってもよいし、ユーザの学習に対する傾向に対応した学習モデルであってもよい。 The learning model for collecting learning data may be a learning model unrelated to the user (for example, an unlearned learning model), or a learning model corresponding to the user's tendency toward learning.
 ユーザの学習に対する傾向に対応した学習モデルを使用する場合、環境データ解析部82は、例えば、制御する学習環境の各要素について学習状態(集中度等)が向上する条件をユーザに入力部24(図1)から入力させてユーザ情報として取得する。制御する環境の各要素について学習状態が向上する条件とは、例えば、ノイズはあった方が学習に集中できるか否か、音楽は流れていた方が学習に集中できるか否か、温度は何度位が学習に集中できるか等、学習環境の各要素について学習状態が向上するとユーザ自身が判断する条件である。 When using a learning model corresponding to the tendency of the user to learn, the environment data analysis unit 82 inputs to the user a condition for improving the learning state (concentration ratio, etc.) for each element of the learning environment to be controlled, for example. It is input from Fig. 1) and acquired as user information. The conditions for improving the learning state for each element of the controlled environment are, for example, whether or not the person with noise can concentrate on learning, whether or not the person who is playing music can concentrate on learning, and what is the temperature. It is a condition for the user to judge that the learning state is improved for each element of the learning environment, such as whether the degree can concentrate on learning.
 一方、環境制御用データベース101には、同一又は類似のユーザ情報ごとに略適切な制御方針を算出する学習モデルが保存される。例えば、環境制御用データベース101には、他のユーザに対して生成された学習モデルとそのユーザのユーザ情報とが対応付けられて複数保存されている。環境データ解析部82は、ユーザから取得したユーザ情報と同一又は類似のユーザ情報に対応した学習モデルを環境制御用データベース101から読み出して学習データ収集用の学習モデルとして使用する。 On the other hand, the environment control database 101 stores a learning model that calculates a substantially appropriate control policy for each of the same or similar user information. For example, in the environment control database 101, a plurality of learning models generated for other users and user information of the user are stored in association with each other. The environment data analysis unit 82 reads a learning model corresponding to the same or similar user information as the user information acquired from the user from the environment control database 101 and uses it as a learning model for collecting learning data.
 また、環境データ解析部82は、学習データ収集用の学習モデルを使用して学習データの収集を行う際に、ユーザの学習状態の良好度の評価を行う。 Further, the environmental data analysis unit 82 evaluates the goodness of the learning state of the user when collecting the learning data using the learning model for collecting the learning data.
 その評価を行うために、環境データ解析部82は、学習データ解析部63からの学習情報Cに基づいて評価値Zを算出する。 In order to perform the evaluation, the environmental data analysis unit 82 calculates the evaluation value Z based on the learning information C from the learning data analysis unit 63.
 環境データ解析部82は、例えば、学習情報Cが、集中度及び理解度を含む場合や、集中度及び理解度以外にも習得速度という情報を含む場合のように複数の情報(要素)を含む場合には、それらの複数の要素の重み付け平均を評価値Zとする。評価値Zは、学習情報Cの複数の要素のうち、1つの要素以外の重みを0とした場合の重み付け平均であってもよく、この場合には、評価値Zは学習情報Cのうちのいずれか1つの要素の値となる。 The environmental data analysis unit 82 includes a plurality of information (elements), for example, when the learning information C includes the concentration level and the comprehension level, or when the learning information C includes information such as the acquisition speed in addition to the concentration level and the comprehension level. In the case, the weighted average of those plurality of elements is set as the evaluation value Z. The evaluation value Z may be a weighted average when the weights other than one element are set to 0 among the plurality of elements of the learning information C. In this case, the evaluation value Z is the learning information C. It is the value of any one element.
 評価値Zは、ユーザの学習状態が良好な程(集中度や理解度が高い程)、高い値を示し、解析結果Aeに基づく学習環境の制御により評価値Zの増加量が大きい程、その解析結果Aeに基づく環境制御がユーザにとって学習状態が向上する適切な制御内容であったことを表す。 The evaluation value Z indicates a higher value as the user's learning state is better (the higher the concentration and understanding), and the larger the increase in the evaluation value Z due to the control of the learning environment based on the analysis result Ae, the higher the value. It shows that the environmental control based on the analysis result Ae was an appropriate control content for the user to improve the learning state.
 環境データ解析部82は、環境制御部83に解析結果Aeを供給して学習環境を変更する(変化させる)ごとに、学習環境センシング部81からの環境情報Eと、ユーザ状態センシング部62からのユーザ状態Gと、学習データ解析部63からの学習情報Cとを取得する。 Each time the environment data analysis unit 82 supplies the analysis result Ae to the environment control unit 83 and changes (changes) the learning environment, the environment information E from the learning environment sensing unit 81 and the user state sensing unit 62 The user state G and the learning information C from the learning data analysis unit 63 are acquired.
 また、環境データ解析部82は、学習環境を変更するごとに、新たに取得した環境情報E、ユーザ状態G、及び、学習情報Cの取得に基づいて学習モデルを用いて次の制御方針を算出して解析結果Aeとして環境制御部83に供給する。 Further, each time the learning environment is changed, the environment data analysis unit 82 calculates the next control policy using the learning model based on the acquisition of the newly acquired environment information E, the user state G, and the learning information C. Then, it is supplied to the environment control unit 83 as the analysis result Ae.
 さらに、環境データ解析部82は、学習環境を変更するごとに、新たに取得した学習情報Cに基づいて評価値Zを算出する。 Further, the environment data analysis unit 82 calculates the evaluation value Z based on the newly acquired learning information C every time the learning environment is changed.
 ここで、解析結果Aeにより環境制御部83がある時刻で学習環境を変更した後、次に学習環境を変更するまでの時間を1時間ステップとして時刻tを時間ステップ数で表すとする。1時間ステップは、環境制御部83が解析結果Aeに基づいて学習環境を変更したときから、その学習環境の変更がユーザの学習状態に対する効果として現れるまでの時間よりも長い時間であるとする。 Here, it is assumed that the time t is represented by the number of time steps, with the time from the change of the learning environment at a certain time by the analysis result Ae to the next change of the learning environment as one hour step. It is assumed that the one-hour step is a time longer than the time from when the environment control unit 83 changes the learning environment based on the analysis result Ae until the change in the learning environment appears as an effect on the learning state of the user.
 また、ある時刻tにおいて(時刻tに該当する1時間ステップの時間内において)、環境データ解析部82が取得した環境情報E、ユーザ状態G、及び、学習情報CをそれぞれE[t]、G[t]、C[t]で表し、学習情報C[t]に基づいて算出した評価値ZをZ[t]で表す。また、環境情報E[t]、ユーザ状態G[t]、及び、学習情報C[t]に基づいて算出された解析結果AeをAe[t]で表す。 Further, at a certain time t (within the time of the one-hour step corresponding to the time t), the environmental information E, the user state G, and the learning information C acquired by the environmental data analysis unit 82 are E [t] and G, respectively. It is represented by [t] and C [t], and the evaluation value Z calculated based on the learning information C [t] is represented by Z [t]. Further, the analysis result Ae calculated based on the environmental information E [t], the user state G [t], and the learning information C [t] is represented by Ae [t].
 このとき、環境データ解析部82は、時刻tのときの評価値Z[t]に対する時刻t+1のときの評価値Z[t+1]の増加量ΔZ[t+1]を求め、その増加量ΔZ[t+1]が所定の閾値ΔZs以上である場合には、時刻tにおける解析結果Ae[t]による学習環境の制御が、適切な制御内容であったと判定する。 At this time, the environmental data analysis unit 82 obtains an increase amount ΔZ [t + 1] of the evaluation value Z [t + 1] at the time t + 1 with respect to the evaluation value Z [t] at the time t, and the increase amount ΔZ [t + 1]. When is equal to or greater than a predetermined threshold value ΔZs, it is determined that the control of the learning environment by the analysis result Ae [t] at the time t is an appropriate control content.
 評価値Z[t]に対する評価値Z[t+1]の増加量ΔZ[t+1]が前記所定の閾値ΔZs未満である場合には、解析結果Ae[t]による学習環境の制御が、適切な制御内容でなかったと判定する。 When the amount of increase ΔZ [t + 1] of the evaluation value Z [t + 1] with respect to the evaluation value Z [t] is less than the predetermined threshold value ΔZs, the control of the learning environment by the analysis result Ae [t] is appropriate control content. It is determined that it was not.
 そして、環境データ解析部82は、解析結果Ae[t]による学習環境の制御が適切な制御内容であったと判定した場合の環境情報E[t]、ユーザ状態G[t]、及び、学習情報C[t]を学習データにおける入力データとして環境制御用データベース101に保存する。 Then, the environment data analysis unit 82 determines that the control of the learning environment by the analysis result Ae [t] has appropriate control contents, the environment information E [t], the user state G [t], and the learning information. C [t] is stored in the environment control database 101 as input data in the training data.
 また、環境データ解析部82は、学習データ収集用の学習モデルに環境情報E[t]、ユーザ状態G[t]、及び、学習情報C[t]を入力データとして入力した際の学習モデルからの出力データを学習データにおける教師データとして環境制御用データベース101に保存する。 Further, the environmental data analysis unit 82 uses the learning model when the environmental information E [t], the user state G [t], and the learning information C [t] are input to the learning model for collecting the learning data as input data. The output data of is stored in the environment control database 101 as teacher data in the training data.
 なお、教師データは、学習データ収集用の学習モデルに環境情報E[t]、ユーザ状態G[t]、及び、学習情報C[t]を入力データとして入力した際の学習モデルからの出力データではなく、その出力データに調整を施したデータであってもよい。たとえば、その学習モデルからの出力データに対して、解析結果Ae[t]とした制御方針に対応する学習モデルの出力ノードからの出力値を1に近づけた値に調整し、その他の出力ノードからの出力値を0に近づけた値に調整したデータを学習データとしてもよい。 The teacher data is output data from the learning model when the environment information E [t], the user state G [t], and the learning information C [t] are input to the learning model for collecting the learning data as input data. However, the output data may be adjusted. For example, for the output data from the learning model, the output value from the output node of the learning model corresponding to the control policy set as the analysis result Ae [t] is adjusted to a value close to 1, and from other output nodes. The data obtained by adjusting the output value of the above to a value close to 0 may be used as the training data.
 環境データ解析部82は、環境情報E、ユーザ状態G、及び、学習情報Cの取得と、環境情報E、ユーザ状態G、及び、学習情報Cに基づく制御方針(解析結果Ae)の算出と、評価値Zの算出とを繰り返し行うことで、学習データを収集して環境制御用データベース101に蓄積させる。 The environmental data analysis unit 82 acquires the environmental information E, the user state G, and the learning information C, calculates the control policy (analysis result Ae) based on the environmental information E, the user state G, and the learning information C, and By repeating the calculation of the evaluation value Z, the learning data is collected and stored in the environment control database 101.
 ユーザが学習を終了した際などにおいて、環境データ解析部82は、環境制御用データベース101に所定数以上の学習データが蓄積されている場合、環境制御用データベース101に蓄積された学習データを用いて学習用の学習モデルの学習を実施する。 When the user finishes learning, the environmental data analysis unit 82 uses the learning data stored in the environment control database 101 when a predetermined number or more of the learning data is stored in the environment control database 101. Learn the learning model for learning.
 学習用の学習モデルは、学習データ収集用の学習モデルであってもよいし、パラメータ(重みやバイアス)が初期値の学習モデルであってもよい。また、学習データを用いた学習用の学習モデルの学習方法としてバッチ学習、ミニバッチ学習、及び、オンライン学習のうちのいずれを採用してもよい。 The learning model for learning may be a learning model for collecting learning data, or may be a learning model with initial values of parameters (weights and biases). Further, any of batch learning, mini-batch learning, and online learning may be adopted as a learning method of a learning model for learning using learning data.
 そして、環境データ解析部82は、学習用の学習モデルを学習させて生成した学習済みの学習モデルをユーザに特化した学習モデルとして環境制御用データベース101に保存する。 Then, the environment data analysis unit 82 stores the learned learning model generated by training the learning model for learning in the environment control database 101 as a learning model specialized for the user.
 環境データ解析部82は、ユーザの次回の学習時からユーザに特化した学習モデルを環境制御用データベース101から読み出して、制御方針(解析結果Ae)の算出に用いる。 The environmental data analysis unit 82 reads the learning model specialized for the user from the environment control database 101 from the next learning of the user, and uses it for calculating the control policy (analysis result Ae).
 なお、環境データ解析部82は、ユーザに特化した学習モデルを使用している場合においても、学習データ収集用の学習モデルを使用している場合と同様に、学習データの収集を行って環境制御用データベース101の学習データの蓄積数を増やしていくようにしてもよい。 Even when the environment data analysis unit 82 uses the learning model specialized for the user, the environment data analysis unit 82 collects the learning data in the same manner as when the learning model for collecting the learning data is used. The number of accumulated learning data in the control database 101 may be increased.
 この場合、環境データ解析部82は、環境制御用データベース101に蓄積された全ての学習データを用いて学習モデルの学習をやりなしてもよいし、前回の学習モデルの学習時からの増加分の学習データを用いて前回までの学習データにより学習させた学習モデルを更に学習させてもよい。 In this case, the environmental data analysis unit 82 may perform learning of the learning model using all the learning data accumulated in the environment control database 101, or the increase from the previous learning of the learning model. The learning model trained by the training data up to the previous time may be further trained using the training data.
 図8は、環境データ解析部82が第1の処理例において行う学習モデルの生成処理を説明するフローチャートである。 FIG. 8 is a flowchart illustrating the learning model generation process performed by the environmental data analysis unit 82 in the first processing example.
 ステップS31では、ユーザが初めて情報処理装置11を使用して学習を行う場合(ユーザに特化した学習モデルが生成されていない場合)に、環境データ解析部82は、環境制御用データベース101から学習データ収集用の学習モデルを読み出す。処理はステップS31からステップS32に進む。 In step S31, when the user first learns using the information processing device 11 (when a learning model specialized for the user is not generated), the environment data analysis unit 82 learns from the environment control database 101. Read the learning model for data collection. The process proceeds from step S31 to step S32.
 ステップS32では、環境データ解析部82は、時刻tを0とする。処理はステップS32からステップS33に進む。 In step S32, the environmental data analysis unit 82 sets the time t to 0. The process proceeds from step S32 to step S33.
 ステップS33では、環境データ解析部82は、学習環境センシング部81から環境情報E[t]を取得し、ユーザ状態センシング部62からユーザ状態G[t]を取得し、学習データ解析部63から学習情報C[t]を取得する。処理はステップS33からステップS34に進む。 In step S33, the environment data analysis unit 82 acquires the environment information E [t] from the learning environment sensing unit 81, acquires the user state G [t] from the user state sensing unit 62, and learns from the learning data analysis unit 63. Acquire information C [t]. The process proceeds from step S33 to step S34.
 ステップS34では、環境データ解析部82は、ステップS33で取得した学習情報C[t]に基づいてユーザの学習状態の良好度を表す評価値Z[t]を算出する。処理はステップS34からステップS35に進む。 In step S34, the environment data analysis unit 82 calculates an evaluation value Z [t] representing the goodness of the learning state of the user based on the learning information C [t] acquired in step S33. The process proceeds from step S34 to step S35.
 ステップS35では、環境データ解析部82は、ステップS33で取得した環境情報E[t]、ユーザ状態G[t]、及び、学習情報C[t]を入力データとして、ステップS31で読み出した学習データ収集用の学習モデルを用いて、学習環境に対する制御方針である解析結果Ae[t]を算出する。処理はステップS35からステップS36に進む。 In step S35, the environment data analysis unit 82 uses the environment information E [t], the user state G [t], and the learning information C [t] acquired in step S33 as input data, and the learning data read in step S31. Using the learning model for collection, the analysis result Ae [t], which is the control policy for the learning environment, is calculated. The process proceeds from step S35 to step S36.
 ステップS36では、環境データ解析部82は、ステップS35で算出した解析結果Ae[t]を環境制御部83の制御信号生成部91に供給し、解析結果Ae[t]にしたがって学習環境を変更させる。処理はステップS36からステップS37に進む。 In step S36, the environment data analysis unit 82 supplies the analysis result Ae [t] calculated in step S35 to the control signal generation unit 91 of the environment control unit 83, and changes the learning environment according to the analysis result Ae [t]. .. The process proceeds from step S36 to step S37.
 ステップS37では、ステップS34で算出した評価値Z[t]の、1時間ステップ前のステップS34で算出した評価値Z[t-1]に対する増加量ΔZ[t](=Z[t]-Z[t-1])が所定の閾値ΔZs以上か否かを判定する。なお、環境データ解析部82は時刻tが0のときは、ステップS37及びステップS38をスキップしてステップS39に進む。 In step S37, the amount of increase ΔZ [t] (= Z [t] −Z) of the evaluation value Z [t] calculated in step S34 with respect to the evaluation value Z [t-1] calculated in step S34 one hour before. It is determined whether or not [t-1]) is equal to or greater than a predetermined threshold value ΔZs. When the time t is 0, the environmental data analysis unit 82 skips step S37 and step S38 and proceeds to step S39.
 ステップS37において、評価値の増加量ΔZ[t]が所定の閾値ΔZs以上でないと判定された場合、処理はステップS38をスキップしてステップS39に進む。 If it is determined in step S37 that the increase amount ΔZ [t] of the evaluation value is not equal to or greater than the predetermined threshold value ΔZs, the process skips step S38 and proceeds to step S39.
 ステップS37において、評価値の増加量ΔZ[t]が所定の閾値ΔZs以上であると判定された場合、処理はステップS38に進み、環境データ解析部82は、1時間ステップ前の環境情報E[t-1]、ユーザ状態G[t-1]、及び、学習情報C[t-1]を学習データとして環境制御用データベース101に保存する。また、環境情報E[t-1]、ユーザ状態G[t-1]、及び、学習情報C[t-1]を学習データ収集用の学習モデルに入力データとして入力した際の学習モデルの出力データ又はその出力データを調整したデータを学習データ(教師データ)として環境制御用データベース101に保存する。処理はステップS38からステップS39に進む。 If it is determined in step S37 that the increase amount ΔZ [t] of the evaluation value is equal to or greater than the predetermined threshold ΔZs, the process proceeds to step S38, and the environmental data analysis unit 82 performs the environmental information E [1 hour before the step. t-1], the user state G [t-1], and the learning information C [t-1] are stored in the environment control database 101 as learning data. Further, the output of the learning model when the environmental information E [t-1], the user state G [t-1], and the learning information C [t-1] are input as input data to the learning model for collecting training data. The data or the adjusted data of the output data is stored in the environment control database 101 as learning data (teacher data). The process proceeds from step S38 to step S39.
 ステップS39では、環境データ解析部82は、ユーザの学習が終了したか否かを判定する。 In step S39, the environment data analysis unit 82 determines whether or not the user's learning has been completed.
 ステップS39において、ユーザの学習が終了していないと判定された場合、処理はステップS39からステップS40に進む。 If it is determined in step S39 that the user's learning has not been completed, the process proceeds from step S39 to step S40.
 ステップS40では、環境データ解析部82は、ステップS33の処理を開始したときから1時間ステップ分の時間が経過するまで待機する。また、時間ステップ数で表された時刻tを1だけインクリメントする。処理はステップS40からステップS33に戻り、ステップS33乃至ステップS40を繰り返す。 In step S40, the environmental data analysis unit 82 waits until the time for one hour step elapses from the time when the process of step S33 is started. Also, the time t represented by the number of time steps is incremented by 1. The process returns from step S40 to step S33, and steps S33 to S40 are repeated.
 一方、ステップS39において、ユーザの学習が終了したと判定された場合、処理はステップS39からステップS41に進む。 On the other hand, if it is determined in step S39 that the user's learning has been completed, the process proceeds from step S39 to step S41.
 ステップS41では、環境データ解析部82は、環境制御用データベース101に保存された学習データを用いて学習用の学習モデルを学習させて、ユーザに特化した学習モデルを生成し、環境制御用データベース101に保存する。 In step S41, the environmental data analysis unit 82 trains the learning model for learning using the learning data stored in the environment control database 101, generates a learning model specialized for the user, and generates the learning model specialized for the user, and the environment control database. Store in 101.
 なお、環境制御用データベース101に蓄積された学習データの数が予め決められた所定数より少ない場合には、環境データ解析部82は、ユーザの次回の学習時においても学習データ収集用の学習モデルを用いて、ステップS31乃至ステップS40の処理を実施して学習データを収集する。そして、環境制御用データベース101に蓄積された学習データの数が予め決められた所定数以上となった場合に、環境データ解析部82は、ステップS41での処理を実施してユーザに特化した学習モデルを生成する。 If the number of learning data stored in the environment control database 101 is less than a predetermined number determined in advance, the environment data analysis unit 82 will use the learning model for learning data collection even at the next learning of the user. Is used to perform the processes of steps S31 to S40 to collect learning data. Then, when the number of learning data stored in the environment control database 101 exceeds a predetermined number determined in advance, the environment data analysis unit 82 performs the process in step S41 to specialize in the user. Generate a learning model.
 ステップS41の処理が終了すると、本フローチャートの処理が終了する。 When the process of step S41 is completed, the process of this flowchart is completed.
<第1の処理例においてユーザに特化した学習済みの学習モデルを用いて解析結果Aeを算出する際の処理>
 図9は、環境データ解析部82が第1の処理例においてユーザに特化した学習済みの学習モデルを用いて解析結果Aeを算出する際の処理を説明するフローチャートである。
<Processing when calculating the analysis result Ae using the trained learning model specialized for the user in the first processing example>
FIG. 9 is a flowchart illustrating a process when the environment data analysis unit 82 calculates the analysis result Ae using the learned learning model specialized for the user in the first processing example.
 ステップS51では、ユーザに特化した学習モデルが生成されている場合、環境データ解析部82は、環境制御用データベース101からユーザに特化した学習済みの学習モデルを読み出す。処理はステップS51からステップS52に進む。 In step S51, when a user-specific learning model is generated, the environment data analysis unit 82 reads the user-specific learned learning model from the environment control database 101. The process proceeds from step S51 to step S52.
 ステップS52では、環境データ解析部82は、時刻tを0とする。処理はステップS52からステップS53に進む。 In step S52, the environmental data analysis unit 82 sets the time t to 0. The process proceeds from step S52 to step S53.
 ステップS53では、環境データ解析部82は、学習環境センシング部81から環境情報E[t]を取得し、ユーザ状態センシング部62からユーザ状態G[t]を取得し、学習データ解析部63から学習情報C[t]を取得する。処理はステップS53からステップS54に進む。 In step S53, the environment data analysis unit 82 acquires the environment information E [t] from the learning environment sensing unit 81, acquires the user state G [t] from the user state sensing unit 62, and learns from the learning data analysis unit 63. Acquire information C [t]. The process proceeds from step S53 to step S54.
 ステップS54では、環境データ解析部82は、ステップS53で取得した環境情報E[t]、ユーザ状態G[t]、及び、学習情報C[t]を入力データとして、ステップS51で読み出した学習モデルを用いて、学習環境に対する制御方針である解析結果Ae[t]を算出する。処理はステップS54からステップS55に進む。 In step S54, the environment data analysis unit 82 takes the environment information E [t], the user state G [t], and the learning information C [t] acquired in step S53 as input data, and reads the learning model in step S51. Is used to calculate the analysis result Ae [t], which is the control policy for the learning environment. The process proceeds from step S54 to step S55.
 ステップS55では、環境データ解析部82は、ステップS54で算出した解析結果Ae[t]を環境制御部83の制御信号生成部91に供給し、解析結果Ae[t]にしたがって学習環境を変更させる。処理はステップS55からステップS56に進む。 In step S55, the environment data analysis unit 82 supplies the analysis result Ae [t] calculated in step S54 to the control signal generation unit 91 of the environment control unit 83, and changes the learning environment according to the analysis result Ae [t]. .. The process proceeds from step S55 to step S56.
 ステップS56では、環境データ解析部82は、ユーザの学習が終了したか否かを判定する。 In step S56, the environment data analysis unit 82 determines whether or not the user's learning has been completed.
 ステップS56において、ユーザの学習が終了していないと判定された場合、処理はステップS57に進む。 If it is determined in step S56 that the user's learning has not been completed, the process proceeds to step S57.
 ステップS57では、環境データ解析部82は、ステップS53の処理を開始したときから1時間ステップ分の時間が経過するまで待機する。また、時間ステップ数で表された時刻tを1だけインクリメントする。処理はステップS57からステップS53に戻り、ステップS53乃至ステップS57を繰り返す。 In step S57, the environmental data analysis unit 82 waits until the time for one hour step elapses from the time when the process of step S53 is started. Also, the time t represented by the number of time steps is incremented by 1. The process returns from step S57 to step S53, and steps S53 to S57 are repeated.
 一方、ステップS56において、ユーザの学習が終了したと判定された場合、処理はステップS57をスキップして、本フローチャートの処理が終了する。 On the other hand, if it is determined in step S56 that the user's learning has been completed, the process skips step S57 and the process of this flowchart ends.
 以上の環境データ解析部82の第1の処理例によれば、環境データ解析部82により、学習環境センシング部81から供給された環境情報Eと、ユーザ状態センシング部62から供給されたユーザ状態Gと、学習データ解析部63から供給された学習情報Cとに基づいて、ユーザの学習状態を向上させる制御方針が学習モデルを用いて算出され、ユーザの手間なく、また、ユーザの嗜好や性格等を考慮した学習環境の適切な制御が行われるようになる。 According to the first processing example of the environmental data analysis unit 82 described above, the environmental data analysis unit 82 supplies the environmental information E supplied from the learning environment sensing unit 81 and the user state G supplied from the user state sensing unit 62. And the learning information C supplied from the learning data analysis unit 63, a control policy for improving the learning state of the user is calculated using the learning model, without the user's trouble, and the user's preference and personality, etc. Appropriate control of the learning environment will be performed in consideration of.
(環境データ解析部82の第2の処理例)
 環境データ解析部82の第2の処理例として、機械学習(ディープラーニング)を使用して強化学習により学習させた学習モデル(DNN)を使用する場合について説明する。
(Second processing example of the environmental data analysis unit 82)
As a second processing example of the environmental data analysis unit 82, a case where a learning model (DNN) trained by reinforcement learning using machine learning (deep learning) is used will be described.
 環境データ解析部82は、学習環境センシング部81から供給された環境情報Eと、ユーザ状態センシング部62から供給されたユーザ状態Gと、学習データ解析部63から供給された学習情報Cとが入力データとして入力されると、全ての制御方針aの価値Qを出力データとして出力するDNNを学習モデルとして用いて解析結果Aeを算出する。なお、学習モデルへの入力データは、環境情報E、ユーザ状態G、及び、学習情報Cのうちのいずれか1つ又は複数の情報であってもよい。 The environment data analysis unit 82 inputs the environment information E supplied from the learning environment sensing unit 81, the user state G supplied from the user state sensing unit 62, and the learning information C supplied from the learning data analysis unit 63. When input as data, the analysis result Ae is calculated using the DNN that outputs the value Q of all the control policies a as output data as a learning model. The input data to the learning model may be any one or more of the environment information E, the user state G, and the learning information C.
 制御方針aは、環境データ解析部82の第1の処理例において説明した制御方針の意味と同じであるので、ここでは説明を省略する。 Since the control policy a has the same meaning as the control policy explained in the first processing example of the environmental data analysis unit 82, the description thereof is omitted here.
 また、学習モデルは強化学習により徐々にユーザに特化した学習モデルへと更新されるが、環境データ解析部82が最初に使用する初期の学習モデルは、ユーザとは無関係の学習モデル(例えば未学習の学習モデル)であってもよいし、ユーザの学習に対する傾向に対応した学習モデルであってもよい。初期の学習モデルについては、環境データ解析部82の第1の処理例の場合における学習データ収集用の学習モデルと同じであるため説明を省略する。 Further, the learning model is gradually updated to a learning model specialized for the user by reinforcement learning, but the initial learning model used first by the environmental data analysis unit 82 is a learning model unrelated to the user (for example, not yet). It may be a learning model of learning), or it may be a learning model corresponding to a user's tendency toward learning. Since the initial learning model is the same as the learning model for collecting learning data in the case of the first processing example of the environmental data analysis unit 82, the description thereof will be omitted.
 制御方針aの価値Qは、制御方針aの良さを表し、制御方針aにしたがって学習環境を変更させたときの報酬Vに基づいて算出される。 The value Q of the control policy a represents the goodness of the control policy a, and is calculated based on the reward V when the learning environment is changed according to the control policy a.
 報酬Vは、環境データ解析部82の第1の処理例の場合の評価値Zと同様にユーザの学習状態の良好度を表す。環境データ解析部82は、学習データ解析部63からの学習情報Cに基づいて報酬Vを算出する。 The reward V represents the degree of goodness of the learning state of the user as in the evaluation value Z in the case of the first processing example of the environmental data analysis unit 82. The environmental data analysis unit 82 calculates the reward V based on the learning information C from the learning data analysis unit 63.
 環境データ解析部82は、例えば、学習情報Cが、集中度及び理解度を含む場合や、集中度及び理解度以外にも習得速度という情報を含む場合のように複数の情報(要素)を含む場合には、それらの複数の要素の重み付け平均を報酬Vとする。報酬Vは、学習情報Cの複数の要素のうち、1つの要素以外の重みを0とした場合の重み付け平均であってもよく、この場合には、報酬Vは学習情報Cのうちのいずれか1つの要素の値となる。 The environmental data analysis unit 82 includes a plurality of information (elements), for example, when the learning information C includes the concentration level and the comprehension level, or when the learning information C includes information such as the acquisition speed in addition to the concentration level and the comprehension level. In the case, the weighted average of those plurality of elements is defined as the reward V. The reward V may be a weighted average when the weights other than one element are set to 0 among the plurality of elements of the learning information C. In this case, the reward V is any one of the learning information C. It is the value of one element.
 また、報酬Vは、ユーザの学習状態が良好な程(集中度や理解度が高い程)、高い値を示す。 In addition, the reward V shows a higher value as the user's learning state is better (the higher the degree of concentration and understanding).
 そして、制御方針aの価値Qは、その制御方針aにしたがって学習環境を変更した後の学習環境における報酬Vとその後に得られるであろう報酬Vとの割引報酬和である(いわゆるベルマン方程式)。 The value Q of the control policy a is the sum of the discount rewards V in the learning environment after the learning environment is changed according to the control policy a and the reward V that will be obtained thereafter (so-called Bellman equation). ..
 また、制御方針aはその後も最適な(価値Qが最大となる)制御方針aを採用するものとして、制御方針aの価値Qである割引報酬和は、次式(1)で表される(いわゆるベルマン最適方程式)。 Further, assuming that the control policy a still adopts the optimum control policy a (the value Q is maximized), the discount reward sum, which is the value Q of the control policy a, is expressed by the following equation (1) (1). The so-called Bellman optimal equation).
 Q[t]=V[t+1]+γQmax[t+1] ・・・(1) Q [t] = V [t + 1] + γQmax [t + 1] ... (1)
 時刻tは、環境データ解析部82の第1の処理例の場合と同様に、時間ステップ数で表された時刻である。価値Q[t]はある時刻tにおける学習環境に対して制御方針aに従って学習環境を変更した場合のその制御方針aの価値Qを表す。なお、時刻tにおける制御方針aを以下、a[t]で表す。報酬V[t+1]は、時刻tにおける学習環境に対して制御方針a[t]にしたがって学習環境を変更した後の学習環境における報酬Vを表す。Q[t+1]は、時刻t+1における学習環境を制御方針a[t+1]に従って変更した場合のその制御方針a[t+1]の価値Qを表し、学習モデルにより算出される。Qmax[t+1]は、時刻t+1において全ての制御方針a[t+1]の各々の価値Q[t+1]のうちの最大値を表す。γは割引率(γは0以上で1以下の値)であり、事前に決められた値である。 The time t is a time represented by the number of time steps, as in the case of the first processing example of the environmental data analysis unit 82. The value Q [t] represents the value Q of the control policy a when the learning environment is changed according to the control policy a with respect to the learning environment at a certain time t. The control policy a at time t is hereinafter represented by a [t]. The reward V [t + 1] represents the reward V in the learning environment after the learning environment is changed according to the control policy a [t] with respect to the learning environment at time t. Q [t + 1] represents the value Q of the control policy a [t + 1] when the learning environment at time t + 1 is changed according to the control policy a [t + 1], and is calculated by the learning model. Qmax [t + 1] represents the maximum value of each value Q [t + 1] of all control policies a [t + 1] at time t + 1. γ is a discount rate (γ is a value of 0 or more and 1 or less), which is a predetermined value.
 環境データ解析部82は、ある時刻tにおいて取得した学習環境センシング部81からの環境情報E[t]と、ユーザ状態センシング部62からのユーザ状態G[t]と、学習データ解析部63からの学習情報C[t]とを学習モデルに入力し、学習モデルの出力データとして全ての制御方針a[t]の価値Q[t]を算出する。 The environment data analysis unit 82 receives the environment information E [t] from the learning environment sensing unit 81 acquired at a certain time t, the user state G [t] from the user state sensing unit 62, and the learning data analysis unit 63. The learning information C [t] is input to the learning model, and the value Q [t] of all the control policies a [t] is calculated as the output data of the learning model.
 なお、学習モデルは、全ての制御方針aの各々に対応する出力ノードを有し、各出力ノードからは例えば0から1までの範囲の値が出力される。学習モデルの各出力ノードからの出力値は、各出力ノードに対応する制御方針aの価値Qを表す。 The learning model has output nodes corresponding to each of all control policies a, and each output node outputs, for example, a value in the range of 0 to 1. The output value from each output node of the learning model represents the value Q of the control policy a corresponding to each output node.
 そして、環境データ解析部82は、学習モデルから出力された全ての制御方針a[t]の価値Q[t]のうち、たとえば、価値Q[t]が最大となる制御方針a[t]を解析結果Ae[t]として環境制御部83に供給する。 Then, the environmental data analysis unit 82 determines, for example, the control policy a [t] that maximizes the value Q [t] among all the value Q [t] of the control policy a [t] output from the learning model. The analysis result Ae [t] is supplied to the environment control unit 83.
 学習モデルから出力される全ての制御方針a[t]の価値Q[t]のうち、最大の価値をQmax[t]で表し、解析結果Ae[t]とされた制御方針をac[t]で表す。なお、学習モデルから出力される全ての制御方針a[t]の価値Q[t]のうち最大の価値Qmax[t]ではない制御方針a[t]を低頻度で解析結果Ae[t]として採用してもよい。 Of all the value Q [t] of all control policies a [t] output from the learning model, the maximum value is represented by Qmax [t], and the control policy with the analysis result Ae [t] is ac [t]. It is represented by. Of all the control policy a [t] values Q [t] output from the learning model, the control policy a [t] that is not the maximum value Qmax [t] is frequently used as the analysis result Ae [t]. It may be adopted.
 環境制御部83が、解析結果Ae[t]に基づいて学習環境を変更すると、次に、環境データ解析部82は、時刻t+1において取得した学習環境センシング部81からの環境情報E[t+1]と、ユーザ状態センシング部62からのユーザ状態G[t+1]と、学習データ解析部63からの学習情報C[t+1]とを学習モデルに入力して、学習モデルの出力データである全ての制御方針a[t+1]の各々の価値Q[t+1]を算出する。そして、環境データ解析部82は、学習モデルから出力された価値Q[t+1]のうち、たとえば、価値Q[t+1]が最大となる制御方針ac[t+1]を時刻t+1における解析結果Aeとして環境制御部83に供給する。 When the environment control unit 83 changes the learning environment based on the analysis result Ae [t], the environment data analysis unit 82 then receives the environment information E [t + 1] from the learning environment sensing unit 81 acquired at time t + 1. , The user state G [t + 1] from the user state sensing unit 62 and the learning information C [t + 1] from the learning data analysis unit 63 are input to the learning model, and all the control policies a which are the output data of the learning model. Each value Q [t + 1] of [t + 1] is calculated. Then, the environmental data analysis unit 82 controls the environment by using, for example, the control policy ac [t + 1] at which the value Q [t + 1] is maximized as the analysis result Ae at the time t + 1 among the value Q [t + 1] output from the learning model. It is supplied to the unit 83.
 また、環境データ解析部82は、学習情報C[t+1]に基づいて報酬V[t+1]を算出する。これにより、環境データ解析部82は、時刻tにおいて解析結果Ae[t]とした制御方針ac[t]の価値Q[t]を上式(1)により算出し、算出した価値Q[t]を制御方針ac[t]の価値の正解値Q′[t]とする。 Further, the environmental data analysis unit 82 calculates the reward V [t + 1] based on the learning information C [t + 1]. As a result, the environmental data analysis unit 82 calculates the value Q [t] of the control policy ac [t], which is the analysis result Ae [t] at time t, by the above equation (1), and the calculated value Q [t]. Let be the correct answer value Q'[t] of the value of the control policy ac [t].
 そして、環境データ解析部82は、時刻tにおける環境情報E[t]、ユーザ状態G[t]、学習情報C[t]を学習モデルに入力した場合に、学習モデルから出力される制御方針ac[t]の価値Q[t]が、正解値Q′[t]となるように学習モデルを学習させる。 Then, when the environment data analysis unit 82 inputs the environment information E [t], the user state G [t], and the learning information C [t] at the time t into the learning model, the control policy ac output from the learning model. The learning model is trained so that the value Q [t] of [t] becomes the correct answer value Q'[t].
 環境データ解析部82は、環境情報E、ユーザ状態G、及び、学習情報Cの取得と、環境情報E、ユーザ状態G、及び、学習情報Cに基づく解析結果Aeの算出と、価値Qの算出とを繰り返し行いながら、学習モデルの強化学習を行う。これにより学習モデルはユーザに特化した学習モデルに徐々に更新される。 The environmental data analysis unit 82 acquires the environmental information E, the user state G, and the learning information C, calculates the analysis result Ae based on the environmental information E, the user state G, and the learning information C, and calculates the value Q. While repeating the above steps, the learning model is strengthened and learned. As a result, the learning model is gradually updated to a user-specific learning model.
 なお、環境データ解析部82は、環境情報E[t]、ユーザ状態G[t]、及び、学習情報C[t]の入力に対する学習モデルの出力データの正解値Q′[t]を取得するごとに学習モデルの学習を行わなくてもよい。 The environment data analysis unit 82 acquires the correct answer value Q'[t] of the output data of the learning model for the input of the environment information E [t], the user state G [t], and the learning information C [t]. It is not necessary to train the learning model for each.
 例えば、環境データ解析部82は、環境情報E[t]、ユーザ状態G[t]、及び、学習情報C[t]と、それらの入力に対する学習モデルの出力データの正解値Q′[t]とを学習データとして環境制御用データベース101に保存し蓄積させる。 For example, the environment data analysis unit 82 has the environment information E [t], the user state G [t], the learning information C [t], and the correct answer value Q'[t] of the output data of the learning model for those inputs. And are stored as learning data in the environment control database 101 and stored.
 環境データ解析部82は、学習データが所定数蓄積されるごとに蓄積された学習データを用いて学習モデルの学習を実施する。 The environmental data analysis unit 82 learns the learning model using the learning data accumulated every time a predetermined number of learning data are accumulated.
 図10は、環境データ解析部82が第2の処理例において学習モデルを用いて解析結果Aeを算出する際の処理を説明するフローチャートである。 FIG. 10 is a flowchart illustrating the processing when the environment data analysis unit 82 calculates the analysis result Ae using the learning model in the second processing example.
 ステップS71では、ユーザが初めて情報処理装置11を使用して学習を行う場合(学習モデルの強化学習が全く行われていない場合)には、環境データ解析部82は、環境制御用データベース101から初期の学習モデルを読み出す。既に学習モデルの強化学習が行われている場合には、環境データ解析部82は、環境制御用データベース101から強化学習が行われた学習モデルを読み出す。処理はステップS71からステップS72に進む。 In step S71, when the user first learns using the information processing device 11 (when the learning model reinforcement learning is not performed at all), the environment data analysis unit 82 initially starts from the environment control database 101. Read the learning model of. When the reinforcement learning of the learning model has already been performed, the environment data analysis unit 82 reads out the learning model in which the reinforcement learning has been performed from the environment control database 101. The process proceeds from step S71 to step S72.
 ステップS72では、環境データ解析部82は、時刻tを0とする。処理はステップS72からステップS73に進む。 In step S72, the environmental data analysis unit 82 sets the time t to 0. The process proceeds from step S72 to step S73.
 ステップS73では、環境データ解析部82は、学習環境センシング部81から環境情報E[t]を取得し、ユーザ状態センシング部62からユーザ状態G[t]を取得し、学習データ解析部63から学習情報C[t]を取得する。処理はステップS73からステップS74に進む。 In step S73, the environment data analysis unit 82 acquires the environment information E [t] from the learning environment sensing unit 81, acquires the user state G [t] from the user state sensing unit 62, and learns from the learning data analysis unit 63. Acquire information C [t]. The process proceeds from step S73 to step S74.
 ステップS74では、環境データ解析部82は、ステップS73で取得した学習情報C[t]に基づいてユーザの学習状態の良好度を表す報酬V[t]を算出する。処理はステップS74からステップS75に進む。なお、時刻tが0の場合、処理はステップS74をスキップしてステップS75に進む。 In step S74, the environmental data analysis unit 82 calculates a reward V [t] representing the goodness of the learning state of the user based on the learning information C [t] acquired in step S73. The process proceeds from step S74 to step S75. When the time t is 0, the process skips step S74 and proceeds to step S75.
 ステップS75では、環境データ解析部82は、ステップS73で取得した環境情報E[t]、ユーザ状態G[t]、及び、学習情報C[t]を入力データとして、学習モデルを用いて、学習環境に対する制御方針ac[t]である解析結果Ae[t]を算出する。処理はステップS75からステップS76に進む。 In step S75, the environment data analysis unit 82 learns using the learning model with the environment information E [t], the user state G [t], and the learning information C [t] acquired in step S73 as input data. The analysis result Ae [t], which is the control policy ac [t] for the environment, is calculated. The process proceeds from step S75 to step S76.
 ステップS76では、環境データ解析部82は、ステップS75で算出した解析結果Ae[t]を環境制御部83の制御信号生成部91に供給し、解析結果Ae[t]にしたがって学習環境を変化させる。処理はステップS76からステップS77に進む。 In step S76, the environment data analysis unit 82 supplies the analysis result Ae [t] calculated in step S75 to the control signal generation unit 91 of the environment control unit 83, and changes the learning environment according to the analysis result Ae [t]. .. The process proceeds from step S76 to step S77.
 ステップS77では、ステップS75で学習モデルから出力された各制御方針の価値Q[t]のうちの最大の価値Qmax[t]と、ステップS74で算出した報酬V[t]とを加算して、1時間ステップ前のステップS75において解析結果Ae[t-1]とした制御方針ac[t-1]の価値Q[t-1]を算出する。そして、算出した価値Q[t-1]を制御方針ac[t-1]に対する価値の正解値Q′[t-1]とする。処理はステップS77からステップS78に進む。 In step S77, the maximum value Qmax [t] of the value Q [t] of each control policy output from the learning model in step S75 and the reward V [t] calculated in step S74 are added. The value Q [t-1] of the control policy ac [t-1], which is the analysis result Ae [t-1] in step S75 one hour before the step, is calculated. Then, the calculated value Q [t-1] is set as the correct value Q'[t-1] of the value with respect to the control policy ac [t-1]. The process proceeds from step S77 to step S78.
 なお、ステップS77の処理により算出された価値Q[t-1]は、上式(1)において、時刻tを時刻t-1に置き換えた算出式により算出された値である。また、時刻tが0の場合、処理はステップS77及びステップS78をスキップしてステップS79に進む。 The value Q [t-1] calculated by the process of step S77 is a value calculated by the calculation formula in which the time t is replaced with the time t-1 in the above formula (1). When the time t is 0, the process skips step S77 and step S78 and proceeds to step S79.
 ステップS78では、環境データ解析部82は、1時間ステップ前の環境情報E[t-1]、ユーザ状態G[t-1]、及び、学習情報C[t-1]を入力データとして学習モデルに入力した場合に学習モデルから出力される制御方針ac[t-1]に対する価値Q[t-1]が正解値Q′[t-1]となるように学習モデルを学習させる。また、環境データ解析部82は、学習させた学習モデルを環境制御用データベース101に保存する。処理はステップS78からステップS79に進む。 In step S78, the environment data analysis unit 82 uses the environment information E [t-1], the user state G [t-1], and the learning information C [t-1] one hour before as input data as input data for the learning model. The learning model is trained so that the value Q [t-1] for the control policy ac [t-1] output from the learning model when inputting to is the correct answer value Q'[t-1]. Further, the environment data analysis unit 82 stores the trained learning model in the environment control database 101. The process proceeds from step S78 to step S79.
 ステップS79では、環境データ解析部82は、ユーザの学習が終了したか否かを判定する。 In step S79, the environment data analysis unit 82 determines whether or not the user's learning has been completed.
 ステップS79において、ユーザの学習が終了していないと判定された場合、処理はステップS80に進む。 If it is determined in step S79 that the user's learning has not been completed, the process proceeds to step S80.
 ステップS80では、環境データ解析部82は、ステップS73の処理を開始したときから1時間ステップ分の時間が経過するまで待機する。また、時間ステップ数で表された時刻tを1だけインクリメントする。処理はステップS80からステップS73に戻り、ステップS73乃至ステップS80を繰り返す。 In step S80, the environmental data analysis unit 82 waits until the time for one hour step elapses from the time when the process of step S73 is started. Also, the time t represented by the number of time steps is incremented by 1. The process returns from step S80 to step S73, and steps S73 to S80 are repeated.
 一方、ステップS79において、ユーザの学習が終了したと判定された場合、本フローチャートの処理が終了する。 On the other hand, if it is determined in step S79 that the user's learning has been completed, the processing of this flowchart ends.
 環境データ解析部82の第2の処理例によれば、環境データ解析部82により、学習環境センシング部81から供給された環境情報Eと、ユーザ状態センシング部62から供給されたユーザ状態Gと、学習データ解析部63から供給された学習情報Cとに基づいて、ユーザの学習状態を向上させる制御方針が学習モデルを用いて算出され、ユーザの手間なく、また、ユーザの嗜好や性格等を考慮した学習環境の適切な制御が行われるようになる。 According to the second processing example of the environment data analysis unit 82, the environment data analysis unit 82 supplies the environment information E supplied from the learning environment sensing unit 81, the user state G supplied from the user state sensing unit 62, and the user state G. Based on the learning information C supplied from the learning data analysis unit 63, a control policy for improving the learning state of the user is calculated using the learning model, and the user's taste and personality are taken into consideration without the user's trouble. Appropriate control of the learning environment will be performed.
 上記図1の情報処理装置11において、情報処理部12は、ユーザの学習空間に配備された情報端末(スマートフォンやパーソナルコンピュータ等)に対してインターネット等の通信回線により接続されたサーバ装置であってもよい。この場合に、情報端末が備える入力部及び出力部が図1の入力部24及び出力部25の代わりとして機能する。また、情報端末は、各種センサ13や各種環境制御機器14とサーバ装置との間の情報のやり取りを仲介する装置として機能する。 In the information processing device 11 of FIG. 1, the information processing unit 12 is a server device connected to an information terminal (smartphone, personal computer, etc.) deployed in the user's learning space by a communication line such as the Internet. May be good. In this case, the input unit and the output unit provided in the information terminal function as a substitute for the input unit 24 and the output unit 25 of FIG. Further, the information terminal functions as a device that mediates the exchange of information between the various sensors 13 and the various environmental control devices 14 and the server device.
 なお、本技術は、以下のような構成も取ることができる。
<1> ユーザの学習状態に基づいて、前記ユーザの学習環境に対する変更内容であって前記学習状態が向上する前記変更内容を算出する処理部
 を有する情報処理装置。
<2> 前記学習状態は、前記ユーザの学習に対する質の良さを表す
 <1>に記載の情報処理装置。
<3> 前記学習状態は、学習に対する前記ユーザの集中度、理解度、及び、習得速度のうちのいずれか1つ又は複数を含む
 <1>又は<2>に記載の情報処理装置。
<4> 前記処理部は、前記集中度、前記理解度、及び、前記習得速度のいずれか1つ又は複数の情報に基づいて算出した評価値に基づいて前記学習環境に対する前記変更内容を算出する
 <3>に記載の情報処理装置。
<5> 前記処理部は、前記学習環境の現在の状態に基づいて、前記学習環境に対する前記変更内容を算出する
 <1>乃至<4>のいずれかに記載の情報処理装置。
<6> 前記学習環境の状態は、音、映像、照度、温度、湿度、気圧、窓又は扉の開閉状態、部屋の散らかり具合、他者の有無、天候、及び、時間に関する状態のうちのいずれか1つ又は複数の状態である
 <5>に記載の情報処理装置。
<7> 前記処理部は、前記ユーザの状態に基づいて、前記学習環境に対する前記変更内容を算出する
 <1>乃至<6>のいずれかに記載の情報処理装置。
<8> 前記ユーザの状態は、前記ユーザの位置、行動、向き、脈拍、発汗、脳波、触覚、嗅覚、及び、味覚に関する状態のうちのいずれか1つ又は複数の状態である
 <7>に記載の情報処理装置。
<9> 前記処理部は、機械学習により学習させた学習モデルを用いて前記学習環境に対する前記変更内容を算出する
 <1>乃至<8>のいずれかに記載の情報処理装置。
<10> 前記処理部は、前記学習環境を変更して前記ユーザの学習状態が向上したときに収集した学習データに基づいて前記学習モデルの学習を行う
 <9>に記載の情報処理装置。
<11> 前記処理部は、前記学習モデルを用いて前記学習環境に対して採り得る全ての変更内容の各々の価値を算出し、前記価値に基づいて前記学習環境に対する前記変更内容を決定する
 <9>に記載の情報処理装置。
<12> 前記学習モデルはディープニューラルネットワークである
 <9>乃至<11>のいずれかに記載の情報処理装置。
<13> 前記ユーザの前記学習状態に応じた問題を前記ユーザに提示する問題生成部
 をさらに有する
 <1>乃至<12>のいずれかに記載の情報処理装置。
<14> 前記処理部により算出された前記学習環境に対する前記変更内容に基づいて前記学習環境を変更する環境制御部
 をさらに有する
 <1>乃至<13>のいずれかに記載の情報処理装置。
<15> 処理部
 を含む
 情報処理装置の
 前記処理部が、ユーザの学習状態に基づいて、前記ユーザの学習環境に対する変更内容であって前記学習状態が向上する前記変更内容を算出する
 情報処理方法。
<16> コンピュータを、
 ユーザの学習状態に基づいて、前記ユーザの学習環境に対する変更内容であって前記学習状態が向上する前記変更内容を算出する処理部
 として機能させるためのプログラム。
The present technology can also have the following configurations.
<1> An information processing device having a processing unit that calculates changes to the learning environment of the user based on the learning state of the user and improves the learning state.
<2> The information processing device according to <1>, wherein the learning state represents the quality of the user's learning.
<3> The information processing apparatus according to <1> or <2>, wherein the learning state includes any one or more of the user's concentration level, comprehension level, and learning speed with respect to learning.
<4> The processing unit calculates the change content for the learning environment based on the evaluation value calculated based on any one or more of the concentration level, the comprehension level, and the learning speed. The information processing device according to <3>.
<5> The information processing apparatus according to any one of <1> to <4>, wherein the processing unit calculates the change contents with respect to the learning environment based on the current state of the learning environment.
<6> The state of the learning environment is any of sound, image, illuminance, temperature, humidity, atmospheric pressure, open / closed state of windows or doors, clutter of the room, presence / absence of others, weather, and time. The information processing apparatus according to <5>, which is in one or more states.
<7> The information processing device according to any one of <1> to <6>, wherein the processing unit calculates the change contents with respect to the learning environment based on the state of the user.
<8> The state of the user is one or more of the states related to the position, behavior, orientation, pulse, sweating, brain wave, touch, smell, and taste of the user <7>. The information processing device described.
<9> The information processing apparatus according to any one of <1> to <8>, wherein the processing unit calculates the changed content with respect to the learning environment using a learning model trained by machine learning.
<10> The information processing apparatus according to <9>, wherein the processing unit learns the learning model based on the learning data collected when the learning environment of the user is changed and the learning state of the user is improved.
<11> The processing unit calculates the value of each of the changes that can be taken for the learning environment using the learning model, and determines the changes for the learning environment based on the values. The information processing apparatus according to 9>.
<12> The information processing apparatus according to any one of <9> to <11>, wherein the learning model is a deep neural network.
<13> The information processing apparatus according to any one of <1> to <12>, further comprising a problem generation unit that presents a problem corresponding to the learning state of the user to the user.
<14> The information processing apparatus according to any one of <1> to <13>, further comprising an environment control unit that changes the learning environment based on the change contents with respect to the learning environment calculated by the processing unit.
<15> An information processing method in which the processing unit of the information processing device including the processing unit calculates, based on the learning state of the user, the changed content for the learning environment of the user and the changed content for which the learning state is improved. ..
<16> Computer
A program for functioning as a processing unit for calculating the changed contents for the learning environment of the user and improving the learning state based on the learning state of the user.
 11 情報処理装置, 12 情報処理部, 13 各種センサ, 14 各種環境制御機器, 21 CPU, 24 入力部, 25 出力部, 41 学習コンテンツ制御部, 42 学習環境制御部, 61 ユーザインタフェース部, 62 ユーザ状態センシング部, 63 学習データ解析部, 64 問題生成部, 81 学習環境センシング部, 82 環境データ解析部, 83 環境制御部, 91 制御信号生成部, 101 環境制御用データベース 11 information processing device, 12 information processing unit, 13 various sensors, 14 various environment control devices, 21 CPU, 24 input unit, 25 output unit, 41 learning content control unit, 42 learning environment control unit, 61 user interface unit, 62 users State sensing unit, 63 learning data analysis unit, 64 problem generation unit, 81 learning environment sensing unit, 82 environment data analysis unit, 83 environment control unit, 91 control signal generation unit, 101 environment control database

Claims (16)

  1.  ユーザの学習状態に基づいて、前記ユーザの学習環境に対する変更内容であって前記学習状態が向上する前記変更内容を算出する処理部
     を有する情報処理装置。
    An information processing device having a processing unit that calculates changes to the learning environment of the user based on the learning state of the user and improves the learning state.
  2.  前記学習状態は、前記ユーザの学習に対する質の良さを表す
     請求項1に記載の情報処理装置。
    The information processing device according to claim 1, wherein the learning state represents the quality of the learning of the user.
  3.  前記学習状態は、学習に対する前記ユーザの集中度、理解度、及び、習得速度のうちのいずれか1つ又は複数を含む
     請求項1に記載の情報処理装置。
    The information processing apparatus according to claim 1, wherein the learning state includes any one or more of the user's concentration level, comprehension level, and learning speed with respect to learning.
  4.  前記処理部は、前記集中度、前記理解度、及び、前記習得速度のうちのいずれか1つ又は複数の情報に基づいて算出した評価値に基づいて前記学習環境に対する前記変更内容を算出する
     請求項3に記載の情報処理装置。
    The processing unit calculates the change contents for the learning environment based on the evaluation value calculated based on the information of any one or more of the concentration level, the comprehension level, and the learning speed. Item 3. The information processing apparatus according to item 3.
  5.  前記処理部は、前記学習環境の現在の状態に基づいて、前記学習環境の前記変更内容を算出する
     請求項1に記載の情報処理装置。
    The information processing device according to claim 1, wherein the processing unit calculates the changed content of the learning environment based on the current state of the learning environment.
  6.  前記学習環境の状態は、音、映像、照度、温度、湿度、気圧、窓又は扉の開閉状態、部屋の散らかり具合、他者の有無、天候、及び、時間に関する状態のうちのいずれか1つ又は複数の状態である
     請求項5に記載の情報処理装置。
    The state of the learning environment is any one of sound, image, illuminance, temperature, humidity, atmospheric pressure, window or door open / closed state, room clutter, presence / absence of others, weather, and time. Or the information processing apparatus according to claim 5, which is in a plurality of states.
  7.  前記処理部は、前記ユーザの状態に基づいて、前記学習環境に対する前記変更内容を算出する
     請求項1に記載の情報処理装置。
    The information processing device according to claim 1, wherein the processing unit calculates the change contents with respect to the learning environment based on the state of the user.
  8.  前記ユーザの状態は、前記ユーザの位置、行動、向き、脈拍、発汗、脳波、触覚、嗅覚、及び、味覚に関する状態のうちのいずれか1つ又は複数の状態である
     請求項7に記載の情報処理装置。
    The information according to claim 7, wherein the state of the user is any one or more of states related to the position, behavior, orientation, pulse, sweating, brain wave, touch, smell, and taste of the user. Processing equipment.
  9.  前記処理部は、機械学習により学習させた学習モデルを用いて前記学習環境に対する前記変更内容を算出する
     請求項1に記載の情報処理装置。
    The information processing device according to claim 1, wherein the processing unit calculates the change contents with respect to the learning environment by using a learning model trained by machine learning.
  10.  前記処理部は、前記学習環境を変更して前記ユーザの学習状態が向上したときに収集した学習データに基づいて前記学習モデルの学習を行う
     請求項9に記載の情報処理装置。
    The information processing device according to claim 9, wherein the processing unit learns the learning model based on the learning data collected when the learning environment of the user is changed and the learning state of the user is improved.
  11.  前記処理部は、前記学習モデルを用いて前記学習環境に対して採り得る全ての変更内容の各々の価値を算出し、前記価値に基づいて前記学習環境に対する前記変更内容を決定する
     請求項9に記載の情報処理装置。
    The processing unit calculates the value of each of the changes that can be made to the learning environment using the learning model, and determines the changes to the learning environment based on the values. The information processing device described.
  12.  前記学習モデルはディープニューラルネットワークである
     請求項9に記載の情報処理装置。
    The information processing device according to claim 9, wherein the learning model is a deep neural network.
  13.  前記ユーザの前記学習状態に応じた問題を前記ユーザに提示する問題生成部
     をさらに有する
     請求項1に記載の情報処理装置。
    The information processing device according to claim 1, further comprising a problem generation unit that presents a problem corresponding to the learning state of the user to the user.
  14.  前記処理部により算出された前記学習環境に対する前記変更内容に基づいて前記学習環境を変更する環境制御部
     をさらに有する
     請求項1に記載の情報処理装置。
    The information processing apparatus according to claim 1, further comprising an environment control unit that changes the learning environment based on the content of the change to the learning environment calculated by the processing unit.
  15.  処理部
     を含む
     情報処理装置の
     前記処理部が、ユーザの学習状態に基づいて、前記ユーザの学習環境に対する変更内容であって前記学習状態が向上する前記変更内容を算出する
     情報処理方法。
    An information processing method in which the processing unit of an information processing device including the processing unit calculates, based on the learning state of the user, the content of change to the learning environment of the user and the content of the change for which the learning state is improved.
  16.  コンピュータを、
     ユーザの学習状態に基づいて、前記ユーザの学習環境に対する変更内容であって前記学習状態が向上する前記変更内容を算出する処理部
     として機能させるためのプログラム。
    Computer,
    A program for functioning as a processing unit for calculating the changed contents for the learning environment of the user and improving the learning state based on the learning state of the user.
PCT/JP2020/040771 2019-11-15 2020-10-30 Information processing device, information processing method, and program WO2021095561A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2019-207032 2019-11-15
JP2019207032 2019-11-15

Publications (1)

Publication Number Publication Date
WO2021095561A1 true WO2021095561A1 (en) 2021-05-20

Family

ID=75912504

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2020/040771 WO2021095561A1 (en) 2019-11-15 2020-10-30 Information processing device, information processing method, and program

Country Status (1)

Country Link
WO (1) WO2021095561A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7436594B1 (en) 2022-09-27 2024-02-21 デジタルアーツ株式会社 Information processing device, information processing method, information processing program, and information processing system

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2014063128A (en) * 2012-08-31 2014-04-10 Panasonic Corp Concentration level measurement device and program
WO2016088463A1 (en) * 2014-12-03 2016-06-09 ソニー株式会社 Information processing device, information processing method, and computer program
JP2017116933A (en) * 2015-12-24 2017-06-29 トムソン ライセンシングThomson Licensing Method and device for providing adapted learning information to user
JP2017144225A (en) * 2016-02-17 2017-08-24 パナソニックIpマネジメント株式会社 Biological information detection device
US20180322798A1 (en) * 2017-05-03 2018-11-08 Florida Atlantic University Board Of Trustees Systems and methods for real time assessment of levels of learning and adaptive instruction delivery
JP2019082311A (en) * 2017-10-30 2019-05-30 ダイキン工業株式会社 Concentration degree estimation device

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2014063128A (en) * 2012-08-31 2014-04-10 Panasonic Corp Concentration level measurement device and program
WO2016088463A1 (en) * 2014-12-03 2016-06-09 ソニー株式会社 Information processing device, information processing method, and computer program
JP2017116933A (en) * 2015-12-24 2017-06-29 トムソン ライセンシングThomson Licensing Method and device for providing adapted learning information to user
JP2017144225A (en) * 2016-02-17 2017-08-24 パナソニックIpマネジメント株式会社 Biological information detection device
US20180322798A1 (en) * 2017-05-03 2018-11-08 Florida Atlantic University Board Of Trustees Systems and methods for real time assessment of levels of learning and adaptive instruction delivery
JP2019082311A (en) * 2017-10-30 2019-05-30 ダイキン工業株式会社 Concentration degree estimation device

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7436594B1 (en) 2022-09-27 2024-02-21 デジタルアーツ株式会社 Information processing device, information processing method, information processing program, and information processing system

Similar Documents

Publication Publication Date Title
KR102334942B1 (en) Data processing method and device for caring robot
JP6730843B2 (en) Communication support system
Kühnel et al. I'm home: Defining and evaluating a gesture set for smart-home control
JP5060978B2 (en) Information presentation system, program, information storage medium, and information presentation system control method
CN106235931A (en) Control the method and device of face cleaning instrument work
JP2010091228A (en) Air conditioner
JP7009342B2 (en) Devices, programs and methods that can evaluate meals based on the amount of chewing and smiling
WO2021095561A1 (en) Information processing device, information processing method, and program
CN112147911A (en) Equipment control method and device
CN113760100B (en) Man-machine interaction equipment with virtual image generation, display and control functions
Ramadan et al. The intelligent classroom: towards an educational ambient intelligence testbed
JP2012251731A (en) Air conditioning system
CN112113317A (en) Indoor thermal environment control system and method
CN110989370B (en) Intelligent household interaction control method and system
CN111260864A (en) Voice broadcast control method and device and air conditioner
JP7367530B2 (en) Information processing device, information processing method, and program
JP2020173787A (en) Information processing apparatus, information processing system, information processing method, and information processing program
CN213119453U (en) Indoor thermal environment control system
TW202022647A (en) Controlling system and controlling method for social network
CN113033336A (en) Home device control method, apparatus, device and computer readable storage medium
JP7085266B2 (en) Devices, programs and methods that can estimate satisfaction based on behavioral information
JP2021189573A (en) Information processing system and information processing method
Flessner et al. Perception based method for indoor air quality control
Dargie et al. A distributed architecture for reasoning about a higher-level context
JP5858228B2 (en) Identification apparatus, identification method, and identification program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20887166

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20887166

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP