WO2022209912A1 - Concentration value calculation system, concentration value calculation method, program, and concentration value calculation model generation system - Google Patents

Concentration value calculation system, concentration value calculation method, program, and concentration value calculation model generation system Download PDF

Info

Publication number
WO2022209912A1
WO2022209912A1 PCT/JP2022/012007 JP2022012007W WO2022209912A1 WO 2022209912 A1 WO2022209912 A1 WO 2022209912A1 JP 2022012007 W JP2022012007 W JP 2022012007W WO 2022209912 A1 WO2022209912 A1 WO 2022209912A1
Authority
WO
WIPO (PCT)
Prior art keywords
concentration
subject
concentration value
value calculation
target
Prior art date
Application number
PCT/JP2022/012007
Other languages
French (fr)
Japanese (ja)
Inventor
徹 臼倉
Original Assignee
パナソニックIpマネジメント株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by パナソニックIpマネジメント株式会社 filed Critical パナソニックIpマネジメント株式会社
Priority to JP2023510917A priority Critical patent/JP7531148B2/en
Publication of WO2022209912A1 publication Critical patent/WO2022209912A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules

Definitions

  • the present disclosure relates to a concentration value calculation system and a concentration value calculation method.
  • an information processing device that calculates a person's degree of concentration (also called a concentration value) is known.
  • the maximum value of the concentration value is set to 100, and the sum of the amount of change in facial expression and the amount of change in action multiplied by the facing rate is subtracted to obtain the concentration value. calculate.
  • the concentration value of a subject looking at one object is not necessarily high, and the concentration value of a subject looking at a plurality of objects is not necessarily low.
  • the conventional information processing apparatus calculates a high concentration value even though it cannot be said that the person is concentrating. be.
  • the conventional information processing apparatus calculates a low concentration value even though it can be said that the student is concentrating on the study.
  • the conventional information processing apparatus has a problem that the accuracy of the calculated concentration value is low.
  • the present disclosure provides a concentration value calculation system and a concentration value calculation method capable of calculating concentration values with higher accuracy.
  • a concentration value calculation system includes: an image acquisition unit that acquires an image stream in which a subject is imaged; A region acquisition unit that acquires a plurality of concentration target regions, each of which corresponds to each of a plurality of concentration targets to be watched by the target person, and a concentration value calculation unit that calculates the concentration value of the target person. and the concentration value calculation unit calculates the face orientation or line-of-sight orientation of the subject from the acquired image stream, and calculates the calculated face orientation or line-of-sight orientation of the subject; Determining whether or not the target person is gazing at any one of the plurality of focus target objects based on the obtained plurality of focus target regions, Calculate and output the concentration value.
  • an image stream in which a subject is imaged is acquired, and a plurality of concentration target areas, each of the plurality of concentration target areas, each of which is the subject's Obtaining a plurality of intensive target areas corresponding to each of a plurality of intensive target objects to be gazed at, and calculating the direction of the face or the direction of the line of sight of the target from the obtained image stream, and calculating the target person.
  • a concentration value of the subject is calculated and output based on the determination result.
  • one aspect of the present disclosure can be implemented as a program for causing a computer to execute the concentration value calculation method.
  • it can be realized as a computer-readable recording medium storing the program.
  • the concentration value can be calculated with higher accuracy.
  • FIG. 1 is a diagram for explaining a usage example of a concentration value calculation system according to an embodiment.
  • FIG. 2 is a block diagram showing the configuration of the concentration value calculation system according to the embodiment.
  • FIG. 3 is a diagram for explaining a concentration target area according to the embodiment.
  • FIG. 4 is a flow chart showing a concentration value calculation method according to the embodiment.
  • FIG. 5 is a diagram for explaining determination of a gaze state according to the embodiment.
  • FIG. 6 is a diagram for explaining transition state determination according to the embodiment.
  • FIG. 7 is a diagram for explaining concentration values calculated in the embodiment.
  • FIG. 8 is a diagram explaining another example of the focused object.
  • FIG. 9 is a block diagram showing a functional configuration of a concentration value calculation unit according to another example.
  • Patent Literature 1 As shown in Japanese Patent Laid-Open No. 2002-200012, an information processing apparatus that calculates a concentration value of a person is conventionally known.
  • the information processing device disclosed in Patent Literature 1 can calculate the concentration value only in a specific situation. Specifically, in the information processing apparatus, the more concentrated one object is viewed, the higher the calculated concentration value. Therefore, in order to use the information processing device for a target person who is working and calculate the concentration value of the target person for the work, it is necessary that the work performed by the target person is a work performed while gazing at one object. , and can only be applied to limited applications.
  • an office worker who performs office work using a computer equipped with a display unit and a sub-display naturally selects both the display unit provided in the computer and the sub-display (that is, alternatively to). Therefore, in a configuration in which a concentration value is calculated based only on gazing at a display unit provided in a computer, an inaccurate concentration value may be calculated in a work scene in which the user gazes at the sub-display even in a concentrated state.
  • a student who studies using a tablet terminal that reproduces a lecture movie, a text, and a notebook selectively gazes at all of the tablet terminal, the text, and the notebook. Therefore, in a configuration in which a concentration value is calculated based only on gazing at a tablet terminal, an inaccurate concentration value may be calculated in a work situation in which the user gazes at a text or a notebook even in a concentrated state.
  • the concentration value calculated using the information processing device cannot be said to have high accuracy.
  • the present disclosure has been made in view of the above circumstances, and provides a concentration value calculation system and the like that can be applied to a wide range of uses and that can calculate a subject's concentration value with high accuracy.
  • each figure is a schematic diagram and is not necessarily strictly illustrated. Therefore, scales and the like are not always the same in each drawing.
  • the same reference numerals are assigned to substantially the same configurations, and overlapping descriptions are omitted or simplified.
  • the first concentrated object and the second concentrated object are exemplified as the plurality of concentrated objects, but there may be three or more concentrated objects.
  • FIG. 1 is a diagram for explaining a concentration value calculation system according to an embodiment.
  • a concentration value calculation system 100 (see FIG. 2 to be described later) according to the present embodiment is built in, for example, a computer (an example of a first concentration object 97) used by a subject 99. Realized.
  • a camera, display, etc. mounted on the computer can be used as a part of the concentration value calculation system 100.
  • an externally connected camera attached to a sub-display (an example of the second focused object 98) used by the subject 99 is used as the imaging device 20.
  • the concentration value calculation system 100 can be incorporated into the computer or the like used by the subject 99, the input to the concentration value calculation system 100 can be obtained using a camera, and the Output can be presented using a display. Further, when the work performed by the subject 99 is work using a computer, the computer used for the work can be used to calculate the concentration value of the subject 99 in parallel with the work. It should be noted that the centralized value calculation system 100 may have a part of processing functions, information storage functions, etc. implemented by a cloud server or the like.
  • the concentration value calculation system 100 is a system that uses an image stream in which the subject 99 is captured and calculates the concentration value of the subject 99 on the image stream. Therefore, if the concentration value calculation system 100 can acquire an image stream in which the subject 99 is imaged, the concentration value calculation system 100 can concentrate even in a situation where the gaze target of the subject 99 shifts to each of two or more focused objects. A value can be calculated. In other words, the concentration value calculation system 100 can be applied to the subject 99 who gazes at each of a plurality of concentration objects in a time division manner.
  • FIG. 2 is a block diagram showing the configuration of the concentration value calculation system according to the embodiment.
  • the concentration value calculation system 100 includes an arithmetic device 10, an imaging device 20, a storage device 30, and an output device 40.
  • Each device constituting the centralized value calculation system 100 may be housed in one housing or the like and integrated, or may be connected to each other via a communication line to form a plurality of individual devices. may be implemented as a device of
  • the computing device 10 has an image acquisition unit 11 , an area acquisition unit 12 , and a concentration value calculation unit 13 .
  • Arithmetic device 10 is implemented by a processor, a memory, and a program executed using these.
  • the computing device 10 is installed, for example, as one of the functions in a computer, which is an example of the first concentration object 97 .
  • the image acquisition unit 11 is a functional unit that acquires an image stream in which the subject 99 is imaged.
  • the image acquisition unit 11 acquires an image stream in which the subject 99 is imaged by the imaging device 20 .
  • the image acquisition unit 11 may be integrated with the imaging device 20 .
  • the concentration value of the subject 99 is immediately calculated from the image stream acquired by the image acquisition section 11 .
  • “immediately” includes a delay of several milliseconds to several seconds considering the time required for calculation processing, data transfer, and the like.
  • the image acquisition unit 11 may be implemented in any way as long as it can acquire an image stream.
  • the image acquisition unit 11 may acquire an image stream stored in the storage device 30 in the past. In this way, the measurement of the concentration value by the concentration value calculation system 100 does not have to be instantaneous.
  • the image acquisition unit 11 transmits the acquired image stream to the concentration value calculation unit 13 .
  • the area acquisition unit 12 acquires a plurality of concentration target areas corresponding to a plurality of concentration objects to be watched by the subject 99, including the first concentration object 97 and the second concentration object 98.
  • the concentration target area is one area set for each one concentration target object, and when the target person 99 gazes at the concentration target object, the direction of the face of the target person 99 falls within the area. It is a set area. Such a concentration target area will be described.
  • FIG. 3 is a diagram for explaining a concentration target area according to the embodiment. In FIG. 3, the concentration target areas set for the first concentration target object 97 and the second concentration target object 98 when the target person 99 is viewed from above are indicated by dot hatching.
  • one focused target area is set for each of the first focused target object 97 and the second focused target object 98 .
  • a first concentration target area 97a is set for the first concentration target object 97
  • a second concentration target area 98a is set for the second concentration target object 98, respectively.
  • the focused target area is, for example, a virtual line connecting the center of the visual field 99a of the target person 99 (that is, the position of the target person 99, more specifically, the position of the eyes of the target person 99) to one end of the focused target, It is set between the virtual line connecting to the other end. That is, the focused target area is set within a predetermined angular range in the field of view 99 a of the subject 99 .
  • the concentration target area is an area dependent on the target person 99 in this way, it is preferable to set it for each target person 99 . Therefore, for example, an operation for setting a concentration target region for each subject is performed prior to calculation of the concentration value.
  • the concentration target area can be determined based on the orientation of the face of the subject 99. is determined experimentally. It should be noted that the calculation of the orientation of the face of the target person 99 is performed based on the region determination image acquired by imaging the target person 99 with the imaging device 20 when the above instruction is presented. Since this operation is the same as the calculation of the face direction of the subject 99 from the image stream performed by the arithmetic unit 10, the explanation here will be omitted by referring to the explanation of the calculation of the concentration value described later. do.
  • the second concentration target area 98a is determined.
  • Information about the determined concentration target area is stored in the storage device 30 in advance.
  • the area acquisition unit 12 acquires the necessary concentration target area by referring to the information stored in the storage device 30 .
  • acquiring the concentration target area means reading and acquiring information indicating the concentration target area.
  • the concentration target area determined by actual measurement may be obtained as it is and used for calculating the concentration value.
  • a first concentration target region 97a and a second concentration target region 98a are acquired as the concentration target regions.
  • an instruction to gaze at four corners of the focused target is given, but depending on the shape of the focused target, five or more corners may be gazed at, or in a horizontal or vertical direction. Only one end and the other end of the focused object in a predetermined direction may be gazed at.
  • the focused object is of a size that fits in the effective visual field of the target person 99 (that is, the visual field area in which information can be effectively obtained that spreads from the central visual field in one direction to the surroundings), simply gaze at the center of the focused object. Only can be used. In this case, an area of about plus or minus 10 degrees from the direction of the face of the subject 99 is automatically determined as the concentration target area.
  • the concentration target area may be determined by machine learning the direction in which the target person 99 is likely to gaze during work.
  • the focus target area can be automatically determined by detecting the position of the center of the field of view 99a of the subject 99 without actually depending on a physical focus target such as a computer or a sub-display. .
  • the concentration target area may be determined according to the space used by the target person 99. For example, a subject 99 using a desk on which a computer with two displays is installed is naturally expected to gaze at the two displays. Therefore, if the position of the center of the field of view 99a of the subject 99 is detected, the focused target area can be determined based on the positional relationship between the imaging device 20 and the two displays.
  • the concentration value calculation unit 13 has a function of calculating the concentration value of the subject 99 based on the image stream acquired by the image acquisition unit 11 and the concentration target area acquired by the area acquisition unit 12. Department. The calculation of the concentration value, which is the main function of the concentration value calculator 13, will be described later in detail.
  • the concentration value of the subject 99 calculated by the concentration value calculator 13 is output and presented on, for example, a computer screen.
  • the calculated concentration value of the target person 99 is output to and stored in a server device (not shown) or the like, and can be made available for confirmation by a manager or the like who is in a position to supervise the work of the target person 99. .
  • the imaging device 20 is a camera that captures images as described above.
  • the imaging device 20 continuously acquires images to generate and output an image stream.
  • the storage device 30 is a device for storing information such as a semiconductor memory.
  • the storage device 30 stores information such as the concentration target area, and receives reference to the information by the area acquisition unit 12 or the like.
  • the output device 40 is, for example, a display controller, converts the information into a presentation image in order to present the information of the concentration value calculated by the concentration value calculation unit 13 on the display, and outputs a signal for presenting the presentation image. output to
  • FIG. 4 is a flow chart showing a concentration value calculation method according to the embodiment.
  • the concentration value calculation system 100 may perform some operations not shown in FIG. 4, such as determination of a concentration target region.
  • the image acquisition unit 11 acquires an image stream (S101).
  • An image stream is a group of images that are captured in sequence. Therefore, the image acquisition unit 11 acquires an image stream by sequentially acquiring a plurality of continuously captured images.
  • the area acquisition unit 12 acquires a plurality of concentration target areas corresponding to each of the plurality of concentration objects by referring to the storage device 30 (S102). Acquisition of the concentration target area (S102) may be performed before acquisition of the image stream (S101). Thus, the order of some operations of the centralized value calculation system 100 may be changed.
  • the concentration value calculator 13 calculates the concentration value of the subject 99 based on the acquired image stream and the acquired concentration target area (S103). Specifically, the concentration value calculation unit 13 calculates the facial orientation and the like of the subject 99 from the acquired image stream. Based on the corresponding plurality of concentration target areas, it is determined whether or not the target person 99 is in a gaze state in which one of the plurality of concentration targets is being gazed at, and the concentration of the target person 99 is determined based on the determination result. Calculate and output the value.
  • concentration value calculator 13 [Calculation of Concentration Value by Concentration Value Calculation System] The operation of the concentration value calculator 13 described above will be described in more detail below. First, calculation of the orientation of the face of the subject 99 will be described. The subject's 99 face orientation is calculated based on the captured image stream. The concentration value calculation unit 13 inputs the obtained image stream to a machine-learned face orientation calculation model, thereby obtaining the face orientation of the subject 99 on the image stream as an output. More specifically, the face orientation calculation model outputs the face orientation of the subject 99 as a normal vector of the front side of the face.
  • the output normal vector of the front side of the face is The relative angle is treated as the orientation of the face of the subject 99, and more specifically, the orientation of said subject's face with respect to the imaging device capturing the image stream.
  • the face orientation calculation model is an example of the orientation calculation model, and outputs the face orientation for each of a plurality of images forming the image stream, so that the face orientation of the subject 99 can be changed according to the image stream. be able to.
  • the face orientation calculation model consists of a teacher image of the subject 99 (corresponding to the image stream) and the correct face orientation data corresponding to the teacher image (corresponding to the relative angle indicating the face orientation of the subject). ) is a trained model that has been trained in advance using a dataset that is a combination of
  • the calculation of the face orientation of the subject 99 is not limited to the example using the face orientation calculation model described above.
  • the feature points of the face of the subject 99 (the corners of the eyes, the tip of the nose, the corners of the mouth, the chin, etc.) of the subject 99 are used to fit the subject to a three-dimensional model.
  • the orientation of the person's face 99 may be calculated.
  • the concentration value calculator 13 may calculate the orientation of the face of the subject 99 from the image stream using any existing technique.
  • the orientation of the line of sight of the subject 99 can be used instead of the orientation of the face of the subject 99.
  • the direction of the line of sight of the subject 99 can be calculated by image analysis centering on the eyeball of the subject 99 .
  • the orientation of the line of sight of the target person 99 can be handled in substantially the same way as the orientation of the target person's 99 face. Therefore, an example using the above-mentioned direction of sight line will be explained by appropriately reading "direction of sight line" for "direction of face” in the description of the present disclosure.
  • FIG. 5 is a diagram for explaining determination of a gaze state according to the embodiment.
  • FIG. 6 is a diagram for explaining determination of a transitional state according to the embodiment. 5 and 6 show the subject 99 from the same viewpoint as in FIG. In FIG. 5, the subject 99 is in a fixation state. Also, in FIG. 6, the subject 99 is in a transitional state.
  • the face orientation of the subject 99 described above is indicated as direction 99b by the dashed arrow (that is, the normal vector) in FIGS.
  • the concentration value calculation unit 13 determines that the target person 99 is in the gaze state if the direction 99b is within the concentration target area. In other words, the direction 99b only needs to fall within the angular range of either the first focused target area 97a or the second focused target area 98a.
  • the face orientation may have a certain angular range. That is, the face direction may be in range 99c or the like.
  • the direction 99b is the direction that bisects the range 99c (that is, the center line).
  • a non-gazing state that is, a state of looking away can be determined.
  • the target person 99 at this time is a plurality of concentration target objects. It is determined that there is a transition state in which the line of sight transitions between two of them. In this way, in the calculation of the concentration value in the present embodiment, by classifying whether the target person 99 is in the gazing state, the transitional state, or the non-gazing state, the To calculate a concentration value with high accuracy.
  • the determination as to whether the subject is in the transitional state or the non-gazing state may be calculated based on the movement vector of the subject's 99 face direction over time.
  • the transition of the face orientation of subject 99 from a region that is neither of the plurality of focus target regions, nor to a region that is none of the plurality of focus target regions happens to be two focus target regions.
  • the duration of a state in which the target person 99 is stationary at a fixed position may be taken into consideration. In this example, even if the direction 99b is located between the first concentration target region 97a and the second concentration target region 98a, if this state continues for a certain period of time, it is considered a non-gazing state. It should be judged.
  • one region and another region is a line segment that connects an arbitrary point in one region and an arbitrary point in another region. It refers to the area to which it does not belong.
  • the concentration value calculation unit 13 further calculates a performance value, which is a unit concentration value of the subject 99, from the acquired image stream.
  • the performance value is a numerical value that is the base of the concentration value calculated from the body movement, posture, facial expression, etc. of the subject 99 on the image. Any existing technique may be used to calculate the performance value. For example, in the above body movement, if an image is acquired in which body movement is greater in number and degree than in the previous image, the performance value of the subject 99 is calculated to be low. Further, for example, in the above posture, a performance value is linked in advance for each posture of the subject 99, and the linked performance value is calculated by matching the posture on the acquired image.
  • the performance value is calculated by summing the numerical values of the features seen in the subject 99 on the image.
  • the state of the target person 99 may be taken into account.
  • the state of the subject 99 is further integrated to optimize the performance value and calculate the concentration value. For example, even if the performance value is a high value, the concentration value should be calculated to be low if the target person 99 is actually in a non-gazing state.
  • the concentration value calculation unit 13 multiplies the calculated performance value by the first coefficient when it is determined that the target person 99 is in the gaze state, and when it is determined that the target person 99 is in the transition state, multiplies the calculated performance value by a second coefficient less than or equal to the first coefficient, and if it is determined that the target person 99 is not in the gaze state and is not in the transition state, the calculated performance value is multiplied by the second coefficient
  • a concentration value of the subject 99 is calculated by multiplying by a third coefficient that is smaller than .
  • condition of the subject 99 and each coefficient for optimizing the performance value may contribute to the habits that the subject 99 can take when concentrating, so it is possible to conduct a preliminary test in advance.
  • Each coefficient may be set for each subject 99 .
  • the transitional state may be treated in the same way as the gaze state. That is, the first coefficient and the second coefficient may have the same value.
  • FIG. 7 is a diagram for explaining concentration values calculated in the embodiment.
  • FIG. 7 shows a graph of concentration values calculated for each image forming an image stream, that is, per time.
  • the target person 99 is in a transitional state (solid line graph), or in a gaze state and a non-gazing state.
  • An example is shown in which the calculated concentration value differs depending on which one (broken line graph).
  • the reliability of the performance value calculated when the target person 99 is in the transitional state is lower than in the case where the subject 99 is in the gaze state and the non-gazing state.
  • a concentration value is calculated to reduce the influence of the performance value on the concentration value.
  • the performance value at the first point in time and the performance value at the second point in time are used to calculate the concentration value at the second point in time.
  • the second point in time is a point in time that follows the first point in time, and includes the minimum unit period for calculating the concentration value in the concentration value calculation system 100 .
  • the minimum unit period for calculating the concentration value in the concentration value calculation system 100 is, for example, one second. Therefore, the concentration value for 1 second is calculated at the second time immediately after the concentration value for 1 second is calculated at the first time.
  • the first term on the right side of the above formula is the first value obtained by multiplying the performance value at the first point in time when it is determined that the subject 99 is not in the transitional state by the first weighting factor, or the first value when the subject 99 is in the transitional state.
  • the third value obtained by multiplying the performance value at the first point in time by the third weighting factor is shown.
  • the second term on the right side of the above equation is the second value obtained by multiplying the performance value at the second time point by the second weighting factor when it is determined that the subject 99 is not in the transition state, or A fourth value obtained by multiplying the performance value at the second time point by a fourth weighting factor when it is determined to be in the state is shown.
  • ⁇ in the formula is a weighting factor for determining which of the performance value at the first time point and the performance value at the second time point should be emphasized.
  • the performance values at the first point in time need not be considered as much. That is, since the performance value at the second point in time in this case has sufficient reliability, it is appropriate to increase the weight of the performance value at the second point in time. Therefore, ⁇ should be set to a relatively small value. ⁇ should be a value greater than 0 and less than 1 in order to satisfy the above equation. If ⁇ is set to 0, the concentration value at the second point in time can be calculated from only the performance value at the second point in time without considering the performance value at the first point in time.
  • should be set to a relatively large value.
  • the reliability of the performance value may be set as a numerical value (that is, a weighting factor) based on the state of the subject 99, and the previous performance value may be incorporated into the calculation of the concentration value for one minimum unit period.
  • the concentration value of the target person 99 can be calculated with higher accuracy while considering the state of the target person 99 when there are a plurality of focused objects.
  • the concentration value calculation system 100 includes the image acquisition unit 11 that acquires an image stream in which the subject 99 is imaged, and a plurality of concentration target regions, each of which is a plurality of concentration target regions. a region acquisition unit 12 for acquiring a plurality of concentration target regions, each corresponding to each of a plurality of concentration targets to be watched by the target person 99; and a concentration value calculation unit 13 for calculating the concentration value of the target person 99.
  • the concentration value calculation unit 13 calculates the face direction or the line-of-sight direction of the subject 99 from the acquired image stream, and the calculated face direction or line-of-sight direction of the subject 99 and the acquired It is determined whether or not the target person 99 is gazing at any one of the plurality of focused objects based on the determined concentration target areas, and the concentration value of the target person 99 is determined based on the determination result. Calculate and output.
  • Such a concentration value calculation system 100 can calculate the concentration value of the subject 99 based on whether the subject 99 is gazing at any one of the plurality of concentration objects.
  • the concentration value is relatively high, and when the target person 99 is not gazing at the focused object, the concentration value is relatively low. can be done. That is, even when there are two or more focused objects, a high concentration value is calculated by gazing at one of the plurality of focused objects.
  • the concentration value is calculated to be low.
  • the concentration value calculation system 100 can thus calculate the concentration value with higher accuracy even when there are a plurality of concentration objects.
  • each of the plurality of concentration target areas is determined as an area imaged when the subject 99 is presented with an instruction to gaze at the concentration target corresponding to the concentration target area. It may be determined based on the image and stored in advance in the storage unit, and the area acquiring unit 12 may acquire a plurality of concentration target areas by referring to the storage unit.
  • each of the plurality of concentration target areas is set to the target person 99 who gazes at one end in response to an instruction to gaze at one end and the other end of the concentration target object corresponding to the concentration target area. It may be determined as an area between the face direction or line of sight direction and the face direction or line of sight direction of the subject 99 who gazes at the other end, and stored in advance in the storage unit.
  • a focused target area determined as an area between the orientation of the face or the orientation of the line of sight of the target person 99 when gazing at one end and the other end.
  • the concentration target area is set in advance for each space used by the subject 99 and stored in the storage unit, and the area acquisition unit 12 refers to the storage unit to determine the plurality of concentration target areas. may be obtained.
  • the concentration target area is set in advance for each subject 99 and stored in the storage unit, and the area acquisition unit 12 may acquire a plurality of concentration target areas by referring to the storage unit. good.
  • the orientation of the subject's 99 face is calculated from the acquired image stream as a normal vector of the subject's 99 face, and each of the plurality of concentrated target regions is centered on the subject's 99 position. has a predetermined angle range, and the concentration value calculation unit 13 determines that the target person 99 corresponds to the concentration target region having the predetermined angle range when the calculated normal vector is within the predetermined angle range. It may be determined that the gaze state is gazing at the focused object.
  • the normal vector of the target person's 99 face can be calculated from the image stream, and it can be determined whether or not the target person's 99 is in a gaze state.
  • the concentration value calculation unit 13 determines that the target person 99 is not in the gaze state, it further determines whether the target person 99 is in a transition state in which the line of sight transitions between two of the plurality of concentration objects. may be determined, and the concentration value of the subject 99 may be calculated based on the determination result.
  • the target person 99 when the target person 99 is not in the gaze state, it is not simply determined as the non-gazing state, but it is determined whether the line of sight transitions between a certain concentration target area and another concentration target area. I can judge. A concentration value can be calculated based on this determination result.
  • the concentration value calculation unit 13 further calculates a performance value, which is a unit concentration value, of the target person 99 from the acquired image stream, and when it is determined that the target person 99 is in the gaze state, , the calculated performance value is multiplied by the first coefficient, and if it is determined that the target person 99 is in a transitional state, the calculated performance value is multiplied by a second coefficient that is equal to or less than the first coefficient, and the target person 99 is in the gaze state. If it is determined that it is not, and if it is determined that the state is not transitional, the concentration value of the subject 99 may be calculated by multiplying the calculated performance value by a third coefficient that is smaller than the second coefficient.
  • a performance value which is a unit concentration value
  • the concentration value is the highest value based on the performance value in the gaze state, the following value in the gaze state based on the performance value in the transition state, and the non-gazing state If there is, it is calculated based on the performance value so that it will be a smaller value than in the case of transitional state.
  • the concentration value calculation unit 13 further calculates a performance value, which is a unit concentration value, of the subject 99 from the acquired image stream at the first time point and at a second time point following the first time point,
  • a performance value which is a unit concentration value
  • the calculated performance value at the first time point is and a second value obtained by multiplying the calculated performance value at the second time point by a second weighting factor that is the difference between the first weighting factor and 1.
  • a third value obtained by multiplying the calculated performance value at the first time point by a third weighting factor different from the first weighting factor, and the calculated A concentration value of the subject 99 is calculated by adding a fourth value obtained by multiplying the performance value at the second time point by a fourth weighting factor which is a difference between the third weighting factor and 1, and
  • the 1 weighting factor and the 3rd weighting factor may be numerical values greater than 0 and less than 1.
  • the concentration value at the second point in time taking into consideration the performance values calculated at the first point in time and the second point in time.
  • the weight that is, the first A more accurate concentration value can be calculated by changing the degree to which the performance value at the time point affects the performance value at the second time point.
  • the concentration value calculation method obtains an image stream in which the subject 99 is captured, and obtains a plurality of concentration target areas, each of which is a concentration target area that the subject 99 is gazing at.
  • a plurality of focused target areas corresponding to each of a plurality of focused targets are acquired, the direction of the face or the direction of the line of sight of the target person 99 is calculated from the acquired image stream, and the calculated face direction of the target person 99 is calculated.
  • the concentration value of the subject 99 is calculated and output.
  • Such a concentration value calculation method can provide the same effects as the concentration value calculation system described above.
  • it may be a program for causing a computer to execute the concentration value calculation method described above.
  • the centralized value calculation system may be realized only by the arithmetic device by providing only the arithmetic device described above and connecting the arithmetic device to an external imaging device, an external storage device, and an external output device.
  • the imaging device, storage device, and output device are not essential components.
  • each of the plurality of focused objects may not be a physical object.
  • the first application window 97b displayed on the display 96 of the computer may be the first focus object
  • the second application window 98b may be the second focus object
  • the subject matter of the present disclosure may be applied.
  • FIG. 9 is a block diagram showing a functional configuration of a concentration value calculation unit according to another example.
  • the computing device 10 includes a concentration value calculator 13a in place of the concentration value calculator 13 in the embodiment.
  • the concentration value calculation unit 13a can directly output the concentration value of the subject 99 by inputting the obtained image stream and the obtained concentration target region to the concentration value calculation model 13b. Then, as the concentration value of the subject 99, the output result output from the concentration value calculation model 13b is output as it is.
  • the concentration value calculation model 13b is a learning model in which the correlation between the image stream, the concentration target region, and the concentration value is learned in advance by machine learning.
  • the concentration value calculation system 100 further includes a model generation unit 13c for generating (learning) the concentration value calculation model 13b.
  • the model generation unit 13c In order to generate the concentration value calculation model 13b, the model generation unit 13c generates input data corresponding to the two pieces of information of the image stream and the concentration target area, and correct (or correct and incorrect) output data for the input data. and are used as training data.
  • an image stream and a teacher image/teacher region D1 corresponding to the concentration target region are input for learning.
  • the teacher concentration value D2 of the subject 99 is input as the output data for learning.
  • the concentration value calculation model 13b is adjusted using a data set combining the teacher image/teacher region D1 and the teacher concentration value D2.
  • the weighting coefficient assigned to each neuron is adjusted by a method such as back propagation to obtain an appropriate value for the input data.
  • Machine learning is performed to obtain output data.
  • the concentration value calculation unit 13a inputs the acquired image stream and the acquired concentration target area to the learned concentration value calculation model 13b, so that an appropriate concentration value of the subject 99 is output. In this way, the calculation of the concentration value by the concentration value calculation unit 13a can also be realized using a machine-learned learning model.
  • the concentration value calculation system 100 including the model generating unit 13c has been described. After that, it is also possible to realize a configuration in which only the recorded concentration value calculation model 13b is used without going through the learning process. is also possible.
  • the present disclosure can be realized not only as a centralized value calculation system, but also as a program including, as steps, processes performed by each component of the centralized value calculation system, and a computer-readable recording medium recording the program.
  • the program may be pre-recorded on a recording medium, or may be supplied to the recording medium via a wide area network including the Internet.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Strategic Management (AREA)
  • Human Resources & Organizations (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Economics (AREA)
  • Educational Administration (AREA)
  • Development Economics (AREA)
  • Game Theory and Decision Science (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Marketing (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Tourism & Hospitality (AREA)
  • General Business, Economics & Management (AREA)
  • Image Analysis (AREA)

Abstract

A concentration value calculation system (100) comprises an image acquisition unit (11) that acquires a stream of images obtained by imaging a person (99), a region acquisition unit (12) that acquires a plurality of concentration target regions each corresponding to each of a plurality of concentration target objects to be gazed at by the person (99), and a concentration value calculation unit (13) that calculates a concentration value of the person (99), wherein the concentration value calculation unit (13) calculates the orientation of the face of or orientation of the line of sight of the person (99) from the acquired stream of images, determines if the person (99) is in a gaze state in which the person (99) is gazing at any of the plurality of concentration target objects on the basis of the calculated orientation of the face of or orientation of the line of sight of the person (99) and the plurality of acquired concentration target regions, and calculates and outputs a concentration value of the person (99) on the basis of the result of determination.

Description

集中値算出システム、集中値算出方法、プログラム、及び、集中値算出モデル生成システムConcentrated Value Calculation System, Concentrated Value Calculation Method, Program, and Concentrated Value Calculation Model Generation System
 本開示は、集中値算出システム及び集中値算出方法に関する。 The present disclosure relates to a concentration value calculation system and a concentration value calculation method.
 従来、人の集中の度合い(集中値ともいう)を算出する情報処理装置が知られている。例えば、特許文献1に記載された情報処理装置では、集中値の最大値を100として、表情の変化量と動作の変化量との和に顔向け率を乗じたものを減算することで集中値を算出する。 Conventionally, an information processing device that calculates a person's degree of concentration (also called a concentration value) is known. For example, in the information processing apparatus described in Patent Literature 1, the maximum value of the concentration value is set to 100, and the sum of the amount of change in facial expression and the amount of change in action multiplied by the facing rate is subtracted to obtain the concentration value. calculate.
特開2014-120137号公報JP 2014-120137 A
 上記従来の情報処理装置では、1つの対象物を集中して見ている程、算出される集中値が高い値になる。しかしながら、現実では、1つの対象物を見ている対象者の集中値が必ず高いという訳でもなく、複数の対象物を見ている対象者の集中値が必ず低いという訳でもない。例えば、学習または執務作業などの知的作業を行う対象者がテレビに見入っている場合、集中している状況とは言えないにも関わらず、上記従来の情報処理装置では高い集中値が算出される。また、学習において参考書とノートとを交互に見ている場合は、学習に集中している状況と言えるにも関わらず、上記従来の情報処理装置では低い集中値が算出される。このように、上記従来の情報処理装置では、算出される集中値の正確度が低いという問題がある。 In the conventional information processing apparatus described above, the more concentrated one is looking at one object, the higher the calculated concentration value becomes. However, in reality, the concentration value of a subject looking at one object is not necessarily high, and the concentration value of a subject looking at a plurality of objects is not necessarily low. For example, when a target person who is doing intellectual work such as study or office work is watching TV, the conventional information processing apparatus calculates a high concentration value even though it cannot be said that the person is concentrating. be. Further, when the user alternately looks at a reference book and a notebook during study, the conventional information processing apparatus calculates a low concentration value even though it can be said that the student is concentrating on the study. As described above, the conventional information processing apparatus has a problem that the accuracy of the calculated concentration value is low.
 そこで、本開示は、より高い正確度で集中値を算出することができる集中値算出システムおよび集中値算出方法を提供する。 Therefore, the present disclosure provides a concentration value calculation system and a concentration value calculation method capable of calculating concentration values with higher accuracy.
 上記課題を解決するために、本開示の一態様に係る集中値算出システムは、対象者が撮像された画像ストリームを取得する画像取得部と、複数の集中対象領域であって、前記複数の集中対象領域の各々が、前記対象者が注視すべき複数の集中対象物の各々に対応する、複数の集中対象領域を取得する領域取得部と、前記対象者の集中値を算出する集中値算出部と、を備え、前記集中値算出部は、取得された前記画像ストリームから、前記対象者の顔の向き又は視線の向きを算出し、算出した前記対象者の顔の向き又は視線の向きと、取得された前記複数の集中対象領域とに基づいて、前記対象者が前記複数の集中対象物のいずれかを注視している注視状態か否かを判定し、判定結果に基づいて前記対象者の集中値を算出して出力する。 In order to solve the above problems, a concentration value calculation system according to an aspect of the present disclosure includes: an image acquisition unit that acquires an image stream in which a subject is imaged; A region acquisition unit that acquires a plurality of concentration target regions, each of which corresponds to each of a plurality of concentration targets to be watched by the target person, and a concentration value calculation unit that calculates the concentration value of the target person. and the concentration value calculation unit calculates the face orientation or line-of-sight orientation of the subject from the acquired image stream, and calculates the calculated face orientation or line-of-sight orientation of the subject; Determining whether or not the target person is gazing at any one of the plurality of focus target objects based on the obtained plurality of focus target regions, Calculate and output the concentration value.
 また、本開示の一態様に係る集中値算出方法では、対象者が撮像された画像ストリームを取得し、複数の集中対象領域であって、前記複数の集中対象領域の各々が、前記対象者が注視すべき複数の集中対象物の各々に対応する複数の集中対象領域を取得し、取得された前記画像ストリームから、前記対象者の顔の向き又は視線の向きを算出し、算出した前記対象者の顔の向き又は視線の向きと、取得された前記複数の集中対象領域とに基づいて、前記対象者が前記複数の集中対象物のいずれかを注視している注視状態であるか否かを判定し、判定結果に基づいて前記対象者の集中値を算出して出力する。 Further, in a concentration value calculation method according to an aspect of the present disclosure, an image stream in which a subject is imaged is acquired, and a plurality of concentration target areas, each of the plurality of concentration target areas, each of which is the subject's Obtaining a plurality of intensive target areas corresponding to each of a plurality of intensive target objects to be gazed at, and calculating the direction of the face or the direction of the line of sight of the target from the obtained image stream, and calculating the target person. determining whether or not the target person is in a gaze state in which he or she is gazing at any one of the plurality of focused objects based on the orientation of the face or the orientation of the line of sight and the plurality of acquired focused target areas; A concentration value of the subject is calculated and output based on the determination result.
 また、本開示の一態様は、上記集中値算出方法をコンピュータに実行させるためのプログラムとして実現することができる。あるいは、当該プログラムを格納したコンピュータ読み取り可能な記録媒体として実現することもできる。 Also, one aspect of the present disclosure can be implemented as a program for causing a computer to execute the concentration value calculation method. Alternatively, it can be realized as a computer-readable recording medium storing the program.
 本開示によれば、より高い正確度で集中値を算出することができる。 According to the present disclosure, the concentration value can be calculated with higher accuracy.
図1は、実施の形態に係る集中値算出システムの使用例を説明する図である。FIG. 1 is a diagram for explaining a usage example of a concentration value calculation system according to an embodiment. 図2は、実施の形態に係る集中値算出システムの構成を示すブロック図である。FIG. 2 is a block diagram showing the configuration of the concentration value calculation system according to the embodiment. 図3は、実施の形態に係る集中対象領域を説明するための図である。FIG. 3 is a diagram for explaining a concentration target area according to the embodiment. 図4は、実施の形態に係る集中値算出方法を示すフローチャートである。FIG. 4 is a flow chart showing a concentration value calculation method according to the embodiment. 図5は、実施の形態に係る注視状態の判定について説明するための図である。FIG. 5 is a diagram for explaining determination of a gaze state according to the embodiment. 図6は、実施の形態に係る推移状態の判定について説明するための図である。FIG. 6 is a diagram for explaining transition state determination according to the embodiment. 図7は、実施の形態において算出される集中値について説明する図である。FIG. 7 is a diagram for explaining concentration values calculated in the embodiment. 図8は、集中対象物の別の例を説明する図である。FIG. 8 is a diagram explaining another example of the focused object. 図9は、別例に係る集中値算出部の機能構成を示すブロック図である。FIG. 9 is a block diagram showing a functional configuration of a concentration value calculation unit according to another example.
 (本開示を得るに至った知見)
 特許文献1に示すように、従来、人の集中値を算出する情報処理装置が知られている。特許文献1に開示された情報処理装置は、特定の状況においてのみ集中値を算出することができる。具体的には、当該情報処理装置では、1つの対象物を集中して見ている程、算出される集中値が高い値になる。したがって、作業中の対象者に対して当該情報処理装置を用い、作業に対する対象者の集中値を算出するためには、対象者の行う作業が1つの対象物に注視しながら行う作業である必要があり、限られた用途にしか適用できない。
(Knowledge leading to the present disclosure)
As shown in Japanese Patent Laid-Open No. 2002-200012, an information processing apparatus that calculates a concentration value of a person is conventionally known. The information processing device disclosed in Patent Literature 1 can calculate the concentration value only in a specific situation. Specifically, in the information processing apparatus, the more concentrated one object is viewed, the higher the calculated concentration value. Therefore, in order to use the information processing device for a target person who is working and calculate the concentration value of the target person for the work, it is necessary that the work performed by the target person is a work performed while gazing at one object. , and can only be applied to limited applications.
 例えば、ディスプレイユニットを備えるコンピュータと、サブディスプレイとを用いてオフィス業務などを行うオフィスワーカにおいては、当然コンピュータに備えられたディスプレイユニットと、サブディスプレイとの両方を選択的に(つまり、択一的に)注視することが想定される。このため、コンピュータに備えられたディスプレイユニットに対する注視のみに基づいて集中値を算出する構成では、集中した状態であってもサブディスプレイを注視する作業の場面で不正確な集中値が算出されうる。また、講義ムービーを再生するタブレット端末と、テキストと、ノートとを用いて学習を行う生徒においては、当然タブレット端末と、テキストと、ノートとのすべてを選択的に注視することが想定される。このため、タブレット端末に対する注視のみに基づいて集中値を算出する構成では、集中した状態であってもテキスト又はノートを注視する作業の場面で不正確な集中値が算出されうる。 For example, an office worker who performs office work using a computer equipped with a display unit and a sub-display naturally selects both the display unit provided in the computer and the sub-display (that is, alternatively to). Therefore, in a configuration in which a concentration value is calculated based only on gazing at a display unit provided in a computer, an inaccurate concentration value may be calculated in a work scene in which the user gazes at the sub-display even in a concentrated state. In addition, it is naturally assumed that a student who studies using a tablet terminal that reproduces a lecture movie, a text, and a notebook selectively gazes at all of the tablet terminal, the text, and the notebook. Therefore, in a configuration in which a concentration value is calculated based only on gazing at a tablet terminal, an inaccurate concentration value may be calculated in a work situation in which the user gazes at a text or a notebook even in a concentrated state.
 このように、当該情報処理装置を用いて算出された集中値は、正確度が高いとはいえない。 Thus, the concentration value calculated using the information processing device cannot be said to have high accuracy.
 本開示は、以上の状況に鑑みてなされ、幅広い用途に適用でき、かつ対象者の集中値を高い正確度で算出することができる集中値算出システム等が提供される。 The present disclosure has been made in view of the above circumstances, and provides a concentration value calculation system and the like that can be applied to a wide range of uses and that can calculate a subject's concentration value with high accuracy.
 以下、本開示の実施の形態について、図面を参照して説明する。なお、以下に説明する実施の形態は、いずれも本開示の包括的または具体的な例を示すものである。したがって、以下の実施の形態で示される、数値、構成要素、構成要素の配置位置および接続形態、並びに、ステップおよびステップの順序等は、一例であって本開示を限定する主旨ではない。よって、以下の実施の形態における構成要素のうち、本開示の独立請求項に記載されていない構成要素については、任意の構成要素として説明される。 Hereinafter, embodiments of the present disclosure will be described with reference to the drawings. It should be noted that all of the embodiments described below represent comprehensive or specific examples of the present disclosure. Therefore, numerical values, components, arrangement positions and connection forms of components, steps, order of steps, and the like shown in the following embodiments are examples and are not intended to limit the present disclosure. Therefore, among constituent elements in the following embodiments, constituent elements not described in independent claims of the present disclosure will be described as optional constituent elements.
 また、各図は、模式図であり、必ずしも厳密に図示されたものではない。したがって、各図において縮尺などは必ずしも一致していない。各図において、実質的に同一の構成に対しては同一の符号を付しており、重複する説明は省略または簡略化する。 In addition, each figure is a schematic diagram and is not necessarily strictly illustrated. Therefore, scales and the like are not always the same in each drawing. In each figure, the same reference numerals are assigned to substantially the same configurations, and overlapping descriptions are omitted or simplified.
 なお、以下説明される実施の形態では、複数の集中対象物として、第1集中対象物及び第2集中対象物の2つを例示するが、集中対象物は、3以上存在してもよい。 Note that in the embodiment described below, the first concentrated object and the second concentrated object are exemplified as the plurality of concentrated objects, but there may be three or more concentrated objects.
 (実施の形態)
 [集中値算出システムの構成]
 はじめに、図1を用いて、実施の形態に係る集中値算出システムについて説明する。図1は、実施の形態に係る集中値算出システムを説明する図である。
(Embodiment)
[Configuration of centralized value calculation system]
First, using FIG. 1, a concentration value calculation system according to an embodiment will be described. FIG. 1 is a diagram for explaining a concentration value calculation system according to an embodiment.
 図1に示すように、本実施の形態における集中値算出システム100(後述する図2参照)は、例えば、対象者99が利用するコンピュータ(第1集中対象物97の一例)等に内蔵されて実現される。集中値算出システム100を対象者99が利用するコンピュータ等に内蔵する形態で実現することにより、コンピュータに搭載されるカメラおよびディスプレイ等を集中値算出システム100の一部の構成として用いることができる。なお、本実施の形態では、対象者99が使用するサブディスプレイ(第2集中対象物98の一例)上に取り付けられた外部接続式のカメラが、撮像装置20として用いられる。 As shown in FIG. 1, a concentration value calculation system 100 (see FIG. 2 to be described later) according to the present embodiment is built in, for example, a computer (an example of a first concentration object 97) used by a subject 99. Realized. By implementing the concentration value calculation system 100 in a form incorporated in a computer or the like used by the subject 99, a camera, display, etc. mounted on the computer can be used as a part of the concentration value calculation system 100. In this embodiment, an externally connected camera attached to a sub-display (an example of the second focused object 98) used by the subject 99 is used as the imaging device 20. FIG.
 このように、集中値算出システム100を対象者99が利用するコンピュータ等に内蔵することにより、集中値算出システム100への入力を、カメラを用いて取得でき、かつ、集中値算出システム100からの出力を、ディスプレイを用いて提示できる。また、対象者99が行う作業が、コンピュータを用いる作業である場合、作業に用いるコンピュータを用いて、当該作業と並行して対象者99の集中値の算出を実施できる。なお、集中値算出システム100は、一部の処理機能や情報記憶機能などがクラウドサーバ等によって実現されてもよい。 In this way, by incorporating the concentration value calculation system 100 into the computer or the like used by the subject 99, the input to the concentration value calculation system 100 can be obtained using a camera, and the Output can be presented using a display. Further, when the work performed by the subject 99 is work using a computer, the computer used for the work can be used to calculate the concentration value of the subject 99 in parallel with the work. It should be noted that the centralized value calculation system 100 may have a part of processing functions, information storage functions, etc. implemented by a cloud server or the like.
 本開示における集中値算出システム100は、対象者99が撮像された画像ストリームを用いて、当該画像ストリーム上の対象者99の集中値を算出するシステムである。したがって、集中値算出システム100は、対象者99が撮像された画像ストリームを取得できれば、対象者99が2以上の集中対象物のそれぞれに注視対象が移行していくような状況であっても集中値の算出を行うことができる。つまり、集中値算出システム100は、複数の集中対象物のそれぞれを時分割に注視する対象者99に対して適用できる。 The concentration value calculation system 100 according to the present disclosure is a system that uses an image stream in which the subject 99 is captured and calculates the concentration value of the subject 99 on the image stream. Therefore, if the concentration value calculation system 100 can acquire an image stream in which the subject 99 is imaged, the concentration value calculation system 100 can concentrate even in a situation where the gaze target of the subject 99 shifts to each of two or more focused objects. A value can be calculated. In other words, the concentration value calculation system 100 can be applied to the subject 99 who gazes at each of a plurality of concentration objects in a time division manner.
 次に、図2を用いて、本実施の形態における集中値算出システム100の機能構成を詳しく説明する。図2は、実施の形態に係る集中値算出システムの構成を示すブロック図である。 Next, with reference to FIG. 2, the functional configuration of the concentration value calculation system 100 according to this embodiment will be described in detail. FIG. 2 is a block diagram showing the configuration of the concentration value calculation system according to the embodiment.
 図2に示すように、本実施の形態における集中値算出システム100は、演算装置10、撮像装置20、記憶装置30、及び、出力装置40を備える。これらの集中値算出システム100を構成する各装置は、1つの筐体等に収められて一体化されていてもよいし、通信回線を介して各装置同士が接続されることで、複数の個別の装置として実現されてもよい。 As shown in FIG. 2, the concentration value calculation system 100 according to the present embodiment includes an arithmetic device 10, an imaging device 20, a storage device 30, and an output device 40. Each device constituting the centralized value calculation system 100 may be housed in one housing or the like and integrated, or may be connected to each other via a communication line to form a plurality of individual devices. may be implemented as a device of
 演算装置10は、画像取得部11、領域取得部12、及び、集中値算出部13を有する。演算装置10は、プロセッサとメモリと、これらを用いて実行されるプログラムとによって実現される。演算装置10は、例えば、第1集中対象物97の一例であるコンピュータに機能の1つとして搭載される。 The computing device 10 has an image acquisition unit 11 , an area acquisition unit 12 , and a concentration value calculation unit 13 . Arithmetic device 10 is implemented by a processor, a memory, and a program executed using these. The computing device 10 is installed, for example, as one of the functions in a computer, which is an example of the first concentration object 97 .
 画像取得部11は、対象者99が撮像された画像ストリームを取得する機能部である。画像取得部11は、一例として、撮像装置20によって対象者99が撮像された画像ストリームを取得する。なお、画像取得部11は、撮像装置20と一体化されてもよい。本実施の形態の一例では、画像取得部11によって取得された画像ストリームから、即時的に対象者99の集中値が算出される。ここでの即時的とは、計算処理やデータ転送等の時間を考慮した数ミリ秒~数秒程度の遅延を含むものである。 The image acquisition unit 11 is a functional unit that acquires an image stream in which the subject 99 is imaged. For example, the image acquisition unit 11 acquires an image stream in which the subject 99 is imaged by the imaging device 20 . Note that the image acquisition unit 11 may be integrated with the imaging device 20 . In one example of the present embodiment, the concentration value of the subject 99 is immediately calculated from the image stream acquired by the image acquisition section 11 . Here, "immediately" includes a delay of several milliseconds to several seconds considering the time required for calculation processing, data transfer, and the like.
 なお、画像取得部11は、画像ストリームを取得できればどのように実現されてもよい。例えば、画像取得部11は、記憶装置30に過去に記憶された画像ストリームを取得してもよい。このように、集中値算出システム100による集中値の計測は即時的でなくてもよい。画像取得部11は、取得した画像ストリームを集中値算出部13へと送信する。 Note that the image acquisition unit 11 may be implemented in any way as long as it can acquire an image stream. For example, the image acquisition unit 11 may acquire an image stream stored in the storage device 30 in the past. In this way, the measurement of the concentration value by the concentration value calculation system 100 does not have to be instantaneous. The image acquisition unit 11 transmits the acquired image stream to the concentration value calculation unit 13 .
 領域取得部12は、第1集中対象物97及び第2集中対象物98を含む、対象者99が注視すべき複数の集中対象物に対応する複数の集中対象領域を取得する。集中対象領域とは、1つの集中対象物ごとに設定された1つの領域であり、対象者99が集中対象物を注視するときに対象者99の顔の向きが当該領域内に該当するように設定された領域である。このような集中対象領域について説明する。図3は、実施の形態に係る集中対象領域を説明するための図である。図3では、対象者99を上方から見た場合における第1集中対象物97及び第2集中対象物98のそれぞれに設定される集中対象領域を、ドットハッチングを付した領域として示している。 The area acquisition unit 12 acquires a plurality of concentration target areas corresponding to a plurality of concentration objects to be watched by the subject 99, including the first concentration object 97 and the second concentration object 98. The concentration target area is one area set for each one concentration target object, and when the target person 99 gazes at the concentration target object, the direction of the face of the target person 99 falls within the area. It is a set area. Such a concentration target area will be described. FIG. 3 is a diagram for explaining a concentration target area according to the embodiment. In FIG. 3, the concentration target areas set for the first concentration target object 97 and the second concentration target object 98 when the target person 99 is viewed from above are indicated by dot hatching.
 図3に示すように、集中対象領域は第1集中対象物97及び第2集中対象物98のそれぞれについて、1つずつ設定された領域である。具体的には、第1集中対象物97には、第1集中対象領域97aが、第2集中対象物98には、第2集中対象領域98aがそれぞれ設定されている。集中対象領域は、例えば対象者99の視野99aの中心(つまり、対象者99の位置、より具体的には対象者99の目の位置)から、集中対象物の一端までを結ぶ仮想線と、他端までを結ぶ仮想線との間に設定されている。すなわち、集中対象領域は、対象者99の視野99aにおける所定の角度範囲に設定されている。 As shown in FIG. 3, one focused target area is set for each of the first focused target object 97 and the second focused target object 98 . Specifically, a first concentration target area 97a is set for the first concentration target object 97, and a second concentration target area 98a is set for the second concentration target object 98, respectively. The focused target area is, for example, a virtual line connecting the center of the visual field 99a of the target person 99 (that is, the position of the target person 99, more specifically, the position of the eyes of the target person 99) to one end of the focused target, It is set between the virtual line connecting to the other end. That is, the focused target area is set within a predetermined angular range in the field of view 99 a of the subject 99 .
 このように集中対象領域は、対象者99に依存する領域であるため、対象者99ごとに設定されることが好ましい。このため、例えば、集中対象領域を対象者ごとに設定するための動作が、集中値の算出に先立って実施される。本実施の形態では、例えば、コンピュータの画面上に、「画面上の4つの角を順番に注視してください」などの指示を表示することで、対象者99の顔の向きなどから集中対象領域が実測的に決定される。なお、対象者99の顔の向き等の算出は、上記の指示が提示されたときに、対象者99を撮像装置20が撮像することで取得された領域決定画像に基づいて行われる。この動作については、演算装置10において行われる画像ストリームからの対象者99の顔の向き等の算出と同様のため、後述の集中値の算出の説明を参照することで、ここでの説明を省略する。 Since the concentration target area is an area dependent on the target person 99 in this way, it is preferable to set it for each target person 99 . Therefore, for example, an operation for setting a concentration target region for each subject is performed prior to calculation of the concentration value. In the present embodiment, for example, by displaying an instruction such as "Look at the four corners of the screen in turn" on the screen of the computer, the concentration target area can be determined based on the orientation of the face of the subject 99. is determined experimentally. It should be noted that the calculation of the orientation of the face of the target person 99 is performed based on the region determination image acquired by imaging the target person 99 with the imaging device 20 when the above instruction is presented. Since this operation is the same as the calculation of the face direction of the subject 99 from the image stream performed by the arithmetic unit 10, the explanation here will be omitted by referring to the explanation of the calculation of the concentration value described later. do.
 また、サブディスプレイなども同様の動作を実施することで、第2集中対象領域98aを決定することができる。決定された集中対象領域に関する情報は、あらかじめ記憶装置30に格納される。そして、領域取得部12は、記憶装置30に格納された情報を参照することで、必要な集中対象領域を取得する。つまり、集中対象領域を取得するとは、集中対象領域を示す情報を読み出して取得することを意味する。また、記憶装置30を参照することは必須ではなく、実測的に決定された集中対象領域をそのまま取得して集中値の算出に用いてもよい。ここでは、集中対象領域として第1集中対象領域97a及び第2集中対象領域98aが取得される。 In addition, by performing the same operation on the sub-display and the like, it is possible to determine the second concentration target area 98a. Information about the determined concentration target area is stored in the storage device 30 in advance. Then, the area acquisition unit 12 acquires the necessary concentration target area by referring to the information stored in the storage device 30 . In other words, acquiring the concentration target area means reading and acquiring information indicating the concentration target area. Further, it is not essential to refer to the storage device 30, and the concentration target area determined by actual measurement may be obtained as it is and used for calculating the concentration value. Here, a first concentration target region 97a and a second concentration target region 98a are acquired as the concentration target regions.
 なお、上記では、集中対象物の4つの角を注視する指示を与える構成としたが、集中対象物の形状によっては、5つ以上の角などを注視させてもよいし、水平方向又は鉛直方向などの所定方向における集中対象物の一端と他端とのみを注視させてもよい。また、集中対象物が対象者99の有効視野(つまり、一方向の中心視野から周囲に広がる、有効に情報を得られる視野領域)に収まるサイズであれば、単に集中対象物の中心を注視させるのみでもよい。この場合は、対象者99の顔の向き等の方向からプラスマイナス10度程度の領域が自動的に集中対象領域として決定される。 In the above description, an instruction to gaze at four corners of the focused target is given, but depending on the shape of the focused target, five or more corners may be gazed at, or in a horizontal or vertical direction. Only one end and the other end of the focused object in a predetermined direction may be gazed at. In addition, if the focused object is of a size that fits in the effective visual field of the target person 99 (that is, the visual field area in which information can be effectively obtained that spreads from the central visual field in one direction to the surroundings), simply gaze at the center of the focused object. Only can be used. In this case, an area of about plus or minus 10 degrees from the direction of the face of the subject 99 is automatically determined as the concentration target area.
 また、集中対象領域は、対象者99が作業中に注視しやすい方向を機械学習することで決定されてもよい。この場合は、実際にコンピュータやサブディスプレイ等の物理的な集中対象物に依存することなく、対象者99の視野99aの中心の位置を検知すれば自動的に集中対象領域を決定することができる。 Also, the concentration target area may be determined by machine learning the direction in which the target person 99 is likely to gaze during work. In this case, the focus target area can be automatically determined by detecting the position of the center of the field of view 99a of the subject 99 without actually depending on a physical focus target such as a computer or a sub-display. .
 さらに、集中対象領域は、対象者99が利用するスペースに応じて決定されていてもよい。例えば、2つのディスプレイを備えるコンピュータが設置されたデスクを利用する対象者99は、自ずと当該2つのディスプレイを注視することが予想される。したがって、対象者99の視野99aの中心の位置を検知すれば、撮像装置20と2つのディスプレイとの位置関係に基づいて集中対象領域を決定することができる。 Furthermore, the concentration target area may be determined according to the space used by the target person 99. For example, a subject 99 using a desk on which a computer with two displays is installed is naturally expected to gaze at the two displays. Therefore, if the position of the center of the field of view 99a of the subject 99 is detected, the focused target area can be determined based on the positional relationship between the imaging device 20 and the two displays.
 図2に戻り、集中値算出部13は、画像取得部11において取得された画像ストリーム、及び、領域取得部12において取得された集中対象領域に基づいて、対象者99の集中値を算出する機能部である。集中値算出部13の主たる機能である集中値の算出については、後で詳しく説明する。集中値算出部13において算出された対象者99の集中値は、例えば、コンピュータの画面などに出力されて提示される。また、算出された対象者99の集中値は、サーバ装置など(不図示)に出力されて格納され、対象者99の作業を監督する立場の管理者等が確認可能な状態とすることもできる。 Returning to FIG. 2, the concentration value calculation unit 13 has a function of calculating the concentration value of the subject 99 based on the image stream acquired by the image acquisition unit 11 and the concentration target area acquired by the area acquisition unit 12. Department. The calculation of the concentration value, which is the main function of the concentration value calculator 13, will be described later in detail. The concentration value of the subject 99 calculated by the concentration value calculator 13 is output and presented on, for example, a computer screen. In addition, the calculated concentration value of the target person 99 is output to and stored in a server device (not shown) or the like, and can be made available for confirmation by a manager or the like who is in a position to supervise the work of the target person 99. .
 撮像装置20は、上記したように画像を撮像するカメラである。撮像装置20は、画像を連続的に取得することで画像ストリームを生成して出力する。 The imaging device 20 is a camera that captures images as described above. The imaging device 20 continuously acquires images to generate and output an image stream.
 記憶装置30は半導体メモリ等の情報記憶用の装置である。記憶装置30は、集中対象領域などの情報を記憶し、領域取得部12などによる情報の参照を受け付ける。 The storage device 30 is a device for storing information such as a semiconductor memory. The storage device 30 stores information such as the concentration target area, and receives reference to the information by the area acquisition unit 12 or the like.
 出力装置40は、例えばディスプレイコントローラであり、集中値算出部13が算出した集中値の情報をディスプレイに提示させるために、情報を提示画像に変換し、当該提示画像を提示させるための信号をディスプレイへと出力する。 The output device 40 is, for example, a display controller, converts the information into a presentation image in order to present the information of the concentration value calculated by the concentration value calculation unit 13 on the display, and outputs a signal for presenting the presentation image. output to
 [集中値算出システムの動作]
 次に図4を参照して、集中値算出システム100の動作について説明する。図4は、実施の形態に係る集中値算出方法を示すフローチャートである。なお、集中値算出システム100は、図4に示されない、例えば、集中対象領域の決定などの一部の動作が実行される場合がある。図4に示すように、集中値算出システム100の動作が開始されると、画像取得部11は、画像ストリームを取得する(S101)。画像ストリームは、連続的に撮像された複数の画像からなる画像群である。したがって、画像取得部11は、連続的に撮像された複数の画像を順次取得することで、画像ストリームの取得を行う。
[Operation of centralized value calculation system]
Next, operation of the concentration value calculation system 100 will be described with reference to FIG. FIG. 4 is a flow chart showing a concentration value calculation method according to the embodiment. Note that the concentration value calculation system 100 may perform some operations not shown in FIG. 4, such as determination of a concentration target region. As shown in FIG. 4, when the concentration value calculation system 100 starts operating, the image acquisition unit 11 acquires an image stream (S101). An image stream is a group of images that are captured in sequence. Therefore, the image acquisition unit 11 acquires an image stream by sequentially acquiring a plurality of continuously captured images.
 また、領域取得部12は、記憶装置30を参照することで、複数の集中対象物のそれぞれに対応する複数の集中対象領域を取得する(S102)。なお、集中対象領域の取得(S102)は、画像ストリームの取得(S101)よりも前に行われてもよい。このように、集中値算出システム100の一部の動作は、順序が入れ替えられてもよい。 Also, the area acquisition unit 12 acquires a plurality of concentration target areas corresponding to each of the plurality of concentration objects by referring to the storage device 30 (S102). Acquisition of the concentration target area (S102) may be performed before acquisition of the image stream (S101). Thus, the order of some operations of the centralized value calculation system 100 may be changed.
 次いで、集中値算出部13は、取得された画像ストリーム及び取得された集中対象領域に基づいて、対象者99の集中値の算出を行う(S103)。具体的には、集中値算出部13は、取得した画像ストリームから、対象者99の顔の向き等を算出し、算出した対象者99の顔の向き等と、複数の集中対象物のそれぞれに対応する複数の集中対象領域とに基づいて、対象者99が複数の集中対象物のいずれかを注視している注視状態であるか否かを判定し、判定結果に基づいて対象者99の集中値を算出して出力する。 Next, the concentration value calculator 13 calculates the concentration value of the subject 99 based on the acquired image stream and the acquired concentration target area (S103). Specifically, the concentration value calculation unit 13 calculates the facial orientation and the like of the subject 99 from the acquired image stream. Based on the corresponding plurality of concentration target areas, it is determined whether or not the target person 99 is in a gaze state in which one of the plurality of concentration targets is being gazed at, and the concentration of the target person 99 is determined based on the determination result. Calculate and output the value.
 [集中値算出システムによる集中値の算出]
 以下、上記に説明した集中値算出部13の動作についてさらに詳しく説明する。まず、対象者99の顔の向き等の算出について説明する。対象者99の顔の向きは、取得された画像ストリームに基づいて算出される。集中値算出部13では、取得された画像ストリームを、機械学習済みの顔の向き算出モデルに入力することで、画像ストリーム上の対象者99の顔の向きを出力として得ることができる。より具体的には、顔の向き算出モデルは、対象者99の顔の向きを顔の正面側の法線ベクトルとして出力する。そして、集中値算出システム100においては、対象者99と、対象者99を撮像する撮像装置20とを直線で結ぶ方向を0度としたときの、出力された顔の正面側の法線ベクトルの相対的な角度を、対象者99の顔の向き、より詳しくは、画像ストリームを撮像する撮像装置に対する前記対象者の顔の向きとして扱う。
[Calculation of Concentration Value by Concentration Value Calculation System]
The operation of the concentration value calculator 13 described above will be described in more detail below. First, calculation of the orientation of the face of the subject 99 will be described. The subject's 99 face orientation is calculated based on the captured image stream. The concentration value calculation unit 13 inputs the obtained image stream to a machine-learned face orientation calculation model, thereby obtaining the face orientation of the subject 99 on the image stream as an output. More specifically, the face orientation calculation model outputs the face orientation of the subject 99 as a normal vector of the front side of the face. In the concentration value calculation system 100, when the direction connecting the target person 99 and the imaging device 20 imaging the target person 99 is assumed to be 0 degrees, the output normal vector of the front side of the face is The relative angle is treated as the orientation of the face of the subject 99, and more specifically, the orientation of said subject's face with respect to the imaging device capturing the image stream.
 顔の向き算出モデルは、向き算出モデルの一例であり、画像ストリームを構成する複数の画像のそれぞれについて顔の向きを出力するので、対象者99の顔の向きの画像ストリームに応じた変化を得ることができる。なお、顔の向き算出モデルは、対象者99の教師画像(画像ストリームに対応)と、当該教師画像に対応する顔の向きの正解データ(対象者の顔の向きを示す相対的な角度に対応)との組み合わせであるデータセットを用いてあらかじめ学習された学習済みモデルである。 The face orientation calculation model is an example of the orientation calculation model, and outputs the face orientation for each of a plurality of images forming the image stream, so that the face orientation of the subject 99 can be changed according to the image stream. be able to. The face orientation calculation model consists of a teacher image of the subject 99 (corresponding to the image stream) and the correct face orientation data corresponding to the teacher image (corresponding to the relative angle indicating the face orientation of the subject). ) is a trained model that has been trained in advance using a dataset that is a combination of
 なお、対象者99の顔の向きの算出は、上記の顔の向き算出モデルを用いる例に限られない。例えば、画像ストリームを構成する各画像のそれぞれにおいて、対象者99の顔の特徴点(対象者99の両目尻、鼻先、口角、顎等)を用いて3次元モデルにあてはめを行うことで、対象者99の顔の向きを算出してもよい。その他、集中値算出部13は、既存のあらゆる技術を用いて画像ストリームから対象者99の顔の向きを算出してもよい。 Note that the calculation of the face orientation of the subject 99 is not limited to the example using the face orientation calculation model described above. For example, in each of the images that make up the image stream, the feature points of the face of the subject 99 (the corners of the eyes, the tip of the nose, the corners of the mouth, the chin, etc.) of the subject 99 are used to fit the subject to a three-dimensional model. The orientation of the person's face 99 may be calculated. Alternatively, the concentration value calculator 13 may calculate the orientation of the face of the subject 99 from the image stream using any existing technique.
 なお、本実施の形態では、対象者99の顔の向きを用いる例を説明するが、対象者99の顔の向きに代えて対象者99の視線の向きを用いることもできる。対象者99の視線の向きは、対象者99の眼球を中心とした画像解析によって算出することができる。なお、対象者99の視線の向きは、対象者99の顔の向きと略同等に扱うことができる。したがって、本開示の説明における「顔の向き」を適宜「視線の向き」に読み替えることによって、上記の視線の向きを用いる例が説明される。 In the present embodiment, an example using the orientation of the face of the subject 99 will be described, but instead of the orientation of the face of the subject 99, the orientation of the line of sight of the subject 99 can be used. The direction of the line of sight of the subject 99 can be calculated by image analysis centering on the eyeball of the subject 99 . In addition, the orientation of the line of sight of the target person 99 can be handled in substantially the same way as the orientation of the target person's 99 face. Therefore, an example using the above-mentioned direction of sight line will be explained by appropriately reading "direction of sight line" for "direction of face" in the description of the present disclosure.
 ここで、図5及び図6を参照して、本実施の形態における対象者99の状態について説明する。図5は、実施の形態に係る注視状態の判定について説明するための図である。また、図6は、実施の形態に係る推移状態の判定について説明するための図である。図5及び図6では、図3と同様の視点における対象者99が示されている。図5では、対象者99は、注視状態である。また、図6では、対象者99は、推移状態である。 Here, the state of the subject 99 in this embodiment will be described with reference to FIGS. 5 and 6. FIG. FIG. 5 is a diagram for explaining determination of a gaze state according to the embodiment. Moreover, FIG. 6 is a diagram for explaining determination of a transitional state according to the embodiment. 5 and 6 show the subject 99 from the same viewpoint as in FIG. In FIG. 5, the subject 99 is in a fixation state. Also, in FIG. 6, the subject 99 is in a transitional state.
 上記に説明した対象者99の顔の向きは、図5及び図6において破線矢印(つまり法線ベクトル)によって方向99bとして示されている。図5に示すように、集中値算出部13は、方向99bが集中対象領域内にあれば対象者99が注視状態であると判定する。つまり、方向99bが、第1集中対象領域97a、又は、第2集中対象領域98aのいずれかの角度範囲に入っていればよい。 The face orientation of the subject 99 described above is indicated as direction 99b by the dashed arrow (that is, the normal vector) in FIGS. As shown in FIG. 5, the concentration value calculation unit 13 determines that the target person 99 is in the gaze state if the direction 99b is within the concentration target area. In other words, the direction 99b only needs to fall within the angular range of either the first focused target area 97a or the second focused target area 98a.
 なお、ヒトの視野には有効視野が存在するため、顔の向きはある程度の角度範囲を有していてもよい。すなわち、顔の向きは、範囲99cなどであってもよい。なお、方向99bは、範囲99cを2等分する方向(つまり中心線)である。範囲99cが用いられる場合、例えば、集中対象領域に範囲99cの一部が重なる状態であれば対象者99は注視状態であると判定されてもよい。 In addition, since the human visual field has an effective visual field, the face orientation may have a certain angular range. That is, the face direction may be in range 99c or the like. Note that the direction 99b is the direction that bisects the range 99c (that is, the center line). When the range 99c is used, for example, it may be determined that the target person 99 is in the gaze state if the range 99c partially overlaps the concentration target region.
 一方で、図6に示すように、方向99bが第1集中対象領域97aの角度範囲にも入っておらず、第2集中対象領域98aの角度範囲にも入っていない場合、従来であれば、非注視状態、すなわち、余所見をしている状態と判定されうる。一方で、本実施の形態では、方向99bが第1集中対象領域97aと第2集中対象領域98aとの間に位置していることから、このときの対象者99を、複数の集中対象物のうちの2つの間を視線が推移する推移状態であると判定する。このように、本実施の形態における集中値の算出では、対象者99が、注視状態であるか、推移状態であるか、又は、非注視状態であるかのそれぞれの場合に分けることで、より正確度の高い集中値の算出を行う。 On the other hand, as shown in FIG. 6, if the direction 99b does not fall within the angular range of the first focused target area 97a nor does it fall within the angular range of the second focused target area 98a, conventionally, A non-gazing state, that is, a state of looking away can be determined. On the other hand, in the present embodiment, since the direction 99b is located between the first concentration target region 97a and the second concentration target region 98a, the target person 99 at this time is a plurality of concentration target objects. It is determined that there is a transition state in which the line of sight transitions between two of them. In this way, in the calculation of the concentration value in the present embodiment, by classifying whether the target person 99 is in the gazing state, the transitional state, or the non-gazing state, the To calculate a concentration value with high accuracy.
 なお、推移状態であるか、非注視状態であるかの判定は、上記に加えて、対象者99の顔の向きの経時的な移動ベクトルによって算出されてもよい。この例では、複数の集中対象領域のいずれでもない領域からの、及び、複数の集中対象領域のいずれでもない領域への、対象者99の顔の向きの推移が、偶発的に2つの集中対象領域の間を通過した場合に、これを推移状態とは判定しないようにすればよい。また、対象者99の顔の向きが一定位置に静止している状態の継続時間などが考慮されてもよい。この例では、方向99bが第1集中対象領域97aと第2集中対象領域98aとの間に位置していたとしても、この状態が一定期間継続されるようであれば、非注視状態であると判定するようにすればよい。 In addition to the above, the determination as to whether the subject is in the transitional state or the non-gazing state may be calculated based on the movement vector of the subject's 99 face direction over time. In this example, the transition of the face orientation of subject 99 from a region that is neither of the plurality of focus target regions, nor to a region that is none of the plurality of focus target regions, happens to be two focus target regions. When passing between areas, it is sufficient not to judge this as a transitional state. Further, the duration of a state in which the target person 99 is stationary at a fixed position may be taken into consideration. In this example, even if the direction 99b is located between the first concentration target region 97a and the second concentration target region 98a, if this state continues for a certain period of time, it is considered a non-gazing state. It should be judged.
 また、一の領域と他の領域の間とは、一の領域内の任意の点と他の領域内の任意の点とを結ぶ線分のうち、一の領域及び他の領域のいずれにも属さない領域のことをいう。 In addition, between one region and another region is a line segment that connects an arbitrary point in one region and an arbitrary point in another region. It refers to the area to which it does not belong.
 集中値算出部13は、さらに、取得した画像ストリームから、対象者99の単位集中値であるパフォーマンス値の算出を行う。パフォーマンス値は、画像上の対象者99の体動、姿勢、及び、表情等から算出される集中値の基部となる数値である。パフォーマンス値の算出では、既存のいかなる技術が用いられてもよい。例えば、上記の体動では、連続する前の画像に比べて回数や程度の大きい体動が行われた画像が取得されれば、対象者99のパフォーマンス値を低く算出する。また、例えば、上記の姿勢では、対象者99の姿勢ごとにあらかじめパフォーマンス値が紐づけられ、取得された画像上の姿勢をマッチングすることで紐づけられたパフォーマンス値を算出する。また、例えば、上記の表情では、対象者の表情に現れるいくつかの特徴に、数値が割り当てられ、画像上の対象者99にみられる特徴の数値を合計することでパフォーマンス値を算出する。なお、パフォーマンス値の算出において、上記の対象者99の状態(注視状態であるか、推移状態であるか、又は、非注視状態であるか)を加味するとしてもよい。 The concentration value calculation unit 13 further calculates a performance value, which is a unit concentration value of the subject 99, from the acquired image stream. The performance value is a numerical value that is the base of the concentration value calculated from the body movement, posture, facial expression, etc. of the subject 99 on the image. Any existing technique may be used to calculate the performance value. For example, in the above body movement, if an image is acquired in which body movement is greater in number and degree than in the previous image, the performance value of the subject 99 is calculated to be low. Further, for example, in the above posture, a performance value is linked in advance for each posture of the subject 99, and the linked performance value is calculated by matching the posture on the acquired image. Also, for example, in the facial expression described above, numerical values are assigned to several features appearing in the subject's facial expression, and the performance value is calculated by summing the numerical values of the features seen in the subject 99 on the image. In calculating the performance value, the state of the target person 99 (whether it is a gaze state, a transitional state, or a non-gazing state) may be taken into account.
 このようにして算出されたパフォーマンス値に対して、本実施の形態では、さらに、対象者99の状態を統合することで、パフォーマンス値を適化して集中値として算出することを行う。例えば、パフォーマンス値が高い値であったとしても、実際には対象者99が非注視状態であれば、集中値が低く算出されるべきである。したがって、集中値算出部13は、対象者99が注視状態であると判定された場合には、算出したパフォーマンス値に第1係数を乗じ、対象者99が推移状態であると判定された場合には、算出したパフォーマンス値に第1係数以下の第2係数を乗じ、対象者99が注視状態でないと判定され、かつ、推移状態でないと判定された場合には、算出したパフォーマンス値に第2係数よりも小さい第3係数を乗じて対象者99の集中値を算出する。 For the performance value calculated in this way, in the present embodiment, the state of the subject 99 is further integrated to optimize the performance value and calculate the concentration value. For example, even if the performance value is a high value, the concentration value should be calculated to be low if the target person 99 is actually in a non-gazing state. Therefore, the concentration value calculation unit 13 multiplies the calculated performance value by the first coefficient when it is determined that the target person 99 is in the gaze state, and when it is determined that the target person 99 is in the transition state, multiplies the calculated performance value by a second coefficient less than or equal to the first coefficient, and if it is determined that the target person 99 is not in the gaze state and is not in the transition state, the calculated performance value is multiplied by the second coefficient A concentration value of the subject 99 is calculated by multiplying by a third coefficient that is smaller than .
 ただし、対象者99の状態と、パフォーマンス値の適化のための各係数は、対象者99の集中時に取りうる癖などが寄与する可能性があるので、事前に予備的な試験を行うことで対象者99ごとに各係数が設定されるようにしてもよい。また、推移状態は、注視状態と同等に扱ってもよい。すなわち、第1係数及び第2係数は同じ値であってもよい。 However, the condition of the subject 99 and each coefficient for optimizing the performance value may contribute to the habits that the subject 99 can take when concentrating, so it is possible to conduct a preliminary test in advance. Each coefficient may be set for each subject 99 . Also, the transitional state may be treated in the same way as the gaze state. That is, the first coefficient and the second coefficient may have the same value.
 また、以下では、図7を参照しつつ、集中値算出の別の例を説明する。図7は、実施の形態において算出される集中値について説明する図である。図7では、画像ストリームを構成する各画像あたり、すなわち、時間あたりに算出される集中値のグラフが示されている。本図では、特に時間軸上両矢印で示す期間において、パフォーマンス値が同じ値であった場合に、対象者99が推移状態であるか(実線のグラフ)、又は、注視状態及び非注視状態のいずれかであるか(破線のグラフ)によって、算出される集中値が異なる例を示している。 Another example of concentration value calculation will be described below with reference to FIG. FIG. 7 is a diagram for explaining concentration values calculated in the embodiment. FIG. 7 shows a graph of concentration values calculated for each image forming an image stream, that is, per time. In this figure, when the performance values are the same during the period indicated by the double-headed arrow on the time axis, the target person 99 is in a transitional state (solid line graph), or in a gaze state and a non-gazing state. An example is shown in which the calculated concentration value differs depending on which one (broken line graph).
 ここでは、対象者99の状態が注視状態及び非注視状態である場合に比べ、推移状態である場合に算出されているパフォーマンス値の信頼性が低いことから、推移状態である場合に算出されるパフォーマンス値による集中値への影響を低減するための集中値の算出を行う。 Here, the reliability of the performance value calculated when the target person 99 is in the transitional state is lower than in the case where the subject 99 is in the gaze state and the non-gazing state. A concentration value is calculated to reduce the influence of the performance value on the concentration value.
 本例では、{第2時点の集中値=(β×第1時点のパフォーマンス値)+((1-β)×第2時点のパフォーマンス値)}の式に則り、集中値の算出を行う。 In this example, the concentration value is calculated according to the formula {concentration value at second time point=(β×performance value at first time point)+((1−β)×performance value at second time point)}.
 ここでは、第2時点での集中値の算出のために第1時点のパフォーマンス値と第2時点のパフォーマンス値とを用いる。第2時点は、第1時点に連続する時点であり、いずれも集中値算出システム100における集中値算出のための最小単位の期間を含んでいる。なお、集中値算出システム100における集中値算出のための最小単位の期間は、例えば1秒である。したがって、第1時点で1秒間の集中値が算出された直後の第2時点で1秒間の集中値の算出が行われる。 Here, the performance value at the first point in time and the performance value at the second point in time are used to calculate the concentration value at the second point in time. The second point in time is a point in time that follows the first point in time, and includes the minimum unit period for calculating the concentration value in the concentration value calculation system 100 . Note that the minimum unit period for calculating the concentration value in the concentration value calculation system 100 is, for example, one second. Therefore, the concentration value for 1 second is calculated at the second time immediately after the concentration value for 1 second is calculated at the first time.
 上記式の右辺第1項は、対象者99が推移状態でないと判定された場合における第1時点でのパフォーマンス値に第1重み係数を乗じた第1値、又は、対象者99が推移状態であると判定された場合における第1時点でのパフォーマンス値に第3重み係数を乗じた第3値を示している。また、上記式の右辺第2項は、対象者99が推移状態でないと判定された場合における第2時点でのパフォーマンス値に第2重み係数を乗じた第2値、又は、対象者99が推移状態であると判定された場合における第2時点でのパフォーマンス値に第4重み係数を乗じた第4値を示している。 The first term on the right side of the above formula is the first value obtained by multiplying the performance value at the first point in time when it is determined that the subject 99 is not in the transitional state by the first weighting factor, or the first value when the subject 99 is in the transitional state. The third value obtained by multiplying the performance value at the first point in time by the third weighting factor is shown. In addition, the second term on the right side of the above equation is the second value obtained by multiplying the performance value at the second time point by the second weighting factor when it is determined that the subject 99 is not in the transition state, or A fourth value obtained by multiplying the performance value at the second time point by a fourth weighting factor when it is determined to be in the state is shown.
 ここで式中のβは、第1時点でのパフォーマンス値と第2時点でのパフォーマンス値とのいずれを重視するかを決定するための重み係数である。例えば、信頼性の高い注視状態又は非注視状態において、第1時点でのパフォーマンス値は、さほど考慮する必要がない。つまり、この場合の第2時点におけるパフォーマンス値は、十分な信頼性があるので、第2時点でのパフォーマンス値の重みを大きくすることが適切である。したがって、βを比較的小さな値にすればよい。上記式を成立させるためには、βは、0より大きく1より小さい値となる。なお、βを0とすれば、第1時点でのパフォーマンス値は全く考慮せずに、第2時点のパフォーマンス値のみから第2時点の集中値を算出することができる。  Here, β in the formula is a weighting factor for determining which of the performance value at the first time point and the performance value at the second time point should be emphasized. For example, in a reliable gaze state or non-gaze state, the performance values at the first point in time need not be considered as much. That is, since the performance value at the second point in time in this case has sufficient reliability, it is appropriate to increase the weight of the performance value at the second point in time. Therefore, β should be set to a relatively small value. β should be a value greater than 0 and less than 1 in order to satisfy the above equation. If β is set to 0, the concentration value at the second point in time can be calculated from only the performance value at the second point in time without considering the performance value at the first point in time.
 一方で、信頼性の低い推移状態において、第1時点でのパフォーマンス値を重視することで、低信頼の第2時点におけるパフォーマンス値が集中値に与える影響を小さくすることができる。したがって、βを比較的大きな値にすればよい。このように、対象者99の状態からパフォーマンス値の信頼性を数値(つまり重み係数)として設定して、1最小単位の期間だけ以前のパフォーマンス値を集中値の算出に取り入れてもよい。 On the other hand, by emphasizing the performance value at the first time point in the low-reliability transitional state, it is possible to reduce the influence of the low-reliability performance value at the second time point on the concentration value. Therefore, β should be set to a relatively large value. In this way, the reliability of the performance value may be set as a numerical value (that is, a weighting factor) based on the state of the subject 99, and the previous performance value may be incorporated into the calculation of the concentration value for one minimum unit period.
 この結果、図7に示すように、注視状態又は非注視状態の場合に比べて、推移状態での集中値の時間領域における変化が緩やかになる(第2時点のパフォーマンス値の影響を受けにくくなる)。 As a result, as shown in FIG. 7, changes in the concentration value in the time domain in the transitional state become more gradual than in the case of the gaze state or the non-gazing state (the effect of the performance value at the second point in time becomes less ).
 以上のようにして、複数の集中対象物が存在する場合に、対象者99の状態を考慮しながら、より高い正確度で対象者99の集中値を算出することができる。 As described above, the concentration value of the target person 99 can be calculated with higher accuracy while considering the state of the target person 99 when there are a plurality of focused objects.
 [効果等]
 以上説明したように、本実施の形態における集中値算出システム100は、対象者99が撮像された画像ストリームを取得する画像取得部11と、複数の集中対象領域であって、複数の集中対象領域の各々が、対象者99が注視すべき複数の集中対象物の各々に対応する、複数の集中対象領域を取得する領域取得部12と、対象者99の集中値を算出する集中値算出部13と、を備え、集中値算出部13は、取得された画像ストリームから、対象者99の顔の向き又は視線の向きを算出し、算出した対象者99の顔の向き又は視線の向きと、取得された複数の集中対象領域とに基づいて、対象者99が複数の集中対象物のいずれかを注視している注視状態か否かを判定し、判定結果に基づいて対象者99の集中値を算出して出力する。
[Effects, etc.]
As described above, the concentration value calculation system 100 according to the present embodiment includes the image acquisition unit 11 that acquires an image stream in which the subject 99 is imaged, and a plurality of concentration target regions, each of which is a plurality of concentration target regions. a region acquisition unit 12 for acquiring a plurality of concentration target regions, each corresponding to each of a plurality of concentration targets to be watched by the target person 99; and a concentration value calculation unit 13 for calculating the concentration value of the target person 99. , the concentration value calculation unit 13 calculates the face direction or the line-of-sight direction of the subject 99 from the acquired image stream, and the calculated face direction or line-of-sight direction of the subject 99 and the acquired It is determined whether or not the target person 99 is gazing at any one of the plurality of focused objects based on the determined concentration target areas, and the concentration value of the target person 99 is determined based on the determination result. Calculate and output.
 このような集中値算出システム100は、対象者99が複数の集中対象物のいずれかを注視しているか否かに基づいて、対象者99の集中値を算出できる。対象者99が集中対象物を注視している場合は、集中値が比較的高くなるように、また、対象者99が集中対象物を注視していない場合は、集中値が比較的低くなるようにできる。つまり、2以上のような複数の集中対象物があるような場合においても、複数の集中対象物のそれぞれのうちいずれかを注視していることで高い集中値が算出される。一方で、複数の集中対象物のいずれも注視していないのであれば、集中値が低く算出される。集中値算出システム100では、このようにして複数の集中対象物がある場合にもより高い正確度で集中値を算出することが可能となる。 Such a concentration value calculation system 100 can calculate the concentration value of the subject 99 based on whether the subject 99 is gazing at any one of the plurality of concentration objects. When the target person 99 is gazing at the focused object, the concentration value is relatively high, and when the target person 99 is not gazing at the focused object, the concentration value is relatively low. can be done. That is, even when there are two or more focused objects, a high concentration value is calculated by gazing at one of the plurality of focused objects. On the other hand, if none of the plurality of concentration targets is being watched, the concentration value is calculated to be low. The concentration value calculation system 100 can thus calculate the concentration value with higher accuracy even when there are a plurality of concentration objects.
 また、例えば、複数の集中対象領域の各々は、対象者99に対して当該集中対象領域に対応する集中対象物を注視させる指示が対象者99に対して提示されたときに撮像された領域決定画像に基づいて決定されて、あらかじめ記憶部に記憶されており、領域取得部12は、記憶部を参照して、複数の集中対象領域を取得してもよい。 Further, for example, each of the plurality of concentration target areas is determined as an area imaged when the subject 99 is presented with an instruction to gaze at the concentration target corresponding to the concentration target area. It may be determined based on the image and stored in advance in the storage unit, and the area acquiring unit 12 may acquire a plurality of concentration target areas by referring to the storage unit.
 これによれば、集中対象物を注視させる指示に対して、対象者99が注視するときの顔の向き又は視線の向きによって決定された集中対象領域を取得できる。 According to this, in response to an instruction to gaze at the focused object, it is possible to acquire the focused target area determined by the direction of the face or the direction of the line of sight when the target person 99 gazes.
 また、例えば、複数の集中対象領域の各々は、対象者99に対して当該集中対象領域に対応する集中対象物の一端及び他端を注視させる指示に対して、一端を注視する対象者99の顔の向き又は視線の向きと、他端を注視する対象者99の顔の向き又は視線の向きとの間の領域として決定されて、あらかじめ記憶部に記憶されていてもよい。 In addition, for example, each of the plurality of concentration target areas is set to the target person 99 who gazes at one end in response to an instruction to gaze at one end and the other end of the concentration target object corresponding to the concentration target area. It may be determined as an area between the face direction or line of sight direction and the face direction or line of sight direction of the subject 99 who gazes at the other end, and stored in advance in the storage unit.
 これによれば、一端及び他端のそれぞれを注視するときの対象者99の顔の向き又は視線の向きの間の領域として決定された集中対象領域を取得できる。 According to this, it is possible to obtain a focused target area determined as an area between the orientation of the face or the orientation of the line of sight of the target person 99 when gazing at one end and the other end.
 また、例えば、集中対象領域は、対象者99が使用するスペースごとにあらかじめ設定されて、記憶部に記憶されており、領域取得部12は、記憶部を参照して、複数の集中対象領域を取得してもよい。 Further, for example, the concentration target area is set in advance for each space used by the subject 99 and stored in the storage unit, and the area acquisition unit 12 refers to the storage unit to determine the plurality of concentration target areas. may be obtained.
 これによれば、スペースごとにあらかじめ設定された集中対象領域を取得できる。 According to this, it is possible to acquire a concentration target area preset for each space.
 また、例えば、集中対象領域は、対象者99ごとにあらかじめ設定されて、記憶部に記憶されており、領域取得部12は、記憶部を参照して、複数の集中対象領域を取得してもよい。 Further, for example, the concentration target area is set in advance for each subject 99 and stored in the storage unit, and the area acquisition unit 12 may acquire a plurality of concentration target areas by referring to the storage unit. good.
 これによれば、対象者99ごとにあらかじめ設定された集中対象領域を取得できる。 According to this, it is possible to obtain a concentration target area preset for each target person 99 .
 また、例えば、対象者99の顔の向きは、取得された画像ストリームから、対象者99の顔の法線ベクトルとして算出され、複数の集中対象領域のそれぞれは、対象者99の位置を中心とした所定の角度範囲を有し、集中値算出部13は、算出された法線ベクトルが、所定の角度範囲内にある場合に、対象者99が当該所定の角度範囲を有する集中対象領域に対応する集中対象物を注視している注視状態と判定してもよい。 Also, for example, the orientation of the subject's 99 face is calculated from the acquired image stream as a normal vector of the subject's 99 face, and each of the plurality of concentrated target regions is centered on the subject's 99 position. has a predetermined angle range, and the concentration value calculation unit 13 determines that the target person 99 corresponds to the concentration target region having the predetermined angle range when the calculated normal vector is within the predetermined angle range. It may be determined that the gaze state is gazing at the focused object.
 これによれば、画像ストリームから、対象者99の顔の法線ベクトルを算出して、注視状態であるか否かを判定できる。 According to this, the normal vector of the target person's 99 face can be calculated from the image stream, and it can be determined whether or not the target person's 99 is in a gaze state.
 また、例えば、集中値算出部13は、注視状態でないと判定した場合に、さらに、対象者99が、複数の集中対象物のうちの2つの間を視線が推移する推移状態であるか否かを判定し、判定結果に基づいて対象者99の集中値を算出してもよい。 Further, for example, when the concentration value calculation unit 13 determines that the target person 99 is not in the gaze state, it further determines whether the target person 99 is in a transition state in which the line of sight transitions between two of the plurality of concentration objects. may be determined, and the concentration value of the subject 99 may be calculated based on the determination result.
 これによれば、対象者99が注視状態でない場合に、単に非注視状態と判定するのではなく、ある集中対象領域と別の集中対象領域との間を視線が推移する推移状態であるかを判定できる。この判定結果によって集中値を算出することができる。 According to this, when the target person 99 is not in the gaze state, it is not simply determined as the non-gazing state, but it is determined whether the line of sight transitions between a certain concentration target area and another concentration target area. I can judge. A concentration value can be calculated based on this determination result.
 また、例えば、集中値算出部13は、さらに、取得された画像ストリームから、対象者99の単位集中値であるパフォーマンス値を算出し、対象者99が注視状態であると判定された場合には、算出したパフォーマンス値に第1係数を乗じ、対象者99が推移状態であると判定された場合には、算出したパフォーマンス値に第1係数以下の第2係数を乗じ、対象者99が注視状態でないと判定され、かつ、推移状態でないと判定された場合には、算出したパフォーマンス値に第2係数よりも小さい第3係数を乗じて対象者99の集中値を算出してもよい。 Further, for example, the concentration value calculation unit 13 further calculates a performance value, which is a unit concentration value, of the target person 99 from the acquired image stream, and when it is determined that the target person 99 is in the gaze state, , the calculated performance value is multiplied by the first coefficient, and if it is determined that the target person 99 is in a transitional state, the calculated performance value is multiplied by a second coefficient that is equal to or less than the first coefficient, and the target person 99 is in the gaze state. If it is determined that it is not, and if it is determined that the state is not transitional, the concentration value of the subject 99 may be calculated by multiplying the calculated performance value by a third coefficient that is smaller than the second coefficient.
 これによれば、集中値は、注視状態であれば、パフォーマンス値に基づいて最も高い数値となり、推移状態であれば、パフォーマンス値に基づいて注視状態である場合以下の数値となり、非注視状態であれば、パフォーマンス値に基づいて推移状態である場合よりも小さい数値となるように算出される。 According to this, the concentration value is the highest value based on the performance value in the gaze state, the following value in the gaze state based on the performance value in the transition state, and the non-gazing state If there is, it is calculated based on the performance value so that it will be a smaller value than in the case of transitional state.
 また、例えば、集中値算出部13は、さらに、取得された画像ストリームから、対象者99の単位集中値であるパフォーマンス値を、第1時点及び第1時点に連続する第2時点で算出し、対象者99が注視状態であると判定された場合、及び、対象者99が注視状態でないと判定され、かつ、推移状態でないと判定された場合には、算出した第1時点におけるパフォーマンス値に第1重み係数を乗じた第1値、及び、算出した第2時点におけるパフォーマンス値に第2重み係数であって第1重み係数と1との差分である第2重み係数を乗じた第2値を加算し、対象者99が推移状態であると判定された場合には、算出した第1時点におけるパフォーマンス値に第1重み係数とは異なる第3重み係数を乗じた第3値、及び、算出した第2時点におけるパフォーマンス値に第4重み係数であって第3重み係数と1との差分である第4重み係数を乗じた第4値を加算して対象者99の集中値を算出し、第1重み係数、及び、第3重み係数は、0より大きく1より小さい数値であってもよい。 Further, for example, the concentration value calculation unit 13 further calculates a performance value, which is a unit concentration value, of the subject 99 from the acquired image stream at the first time point and at a second time point following the first time point, When it is determined that the target person 99 is in the gaze state, and when it is determined that the target person 99 is not in the gaze state and is not in the transition state, the calculated performance value at the first time point is and a second value obtained by multiplying the calculated performance value at the second time point by a second weighting factor that is the difference between the first weighting factor and 1. If it is determined that the subject 99 is in a transitional state, a third value obtained by multiplying the calculated performance value at the first time point by a third weighting factor different from the first weighting factor, and the calculated A concentration value of the subject 99 is calculated by adding a fourth value obtained by multiplying the performance value at the second time point by a fourth weighting factor which is a difference between the third weighting factor and 1, and The 1 weighting factor and the 3rd weighting factor may be numerical values greater than 0 and less than 1.
 これによれば、第1時点と第2時点とのそれぞれにおいて算出されたパフォーマンス値を考慮した、第2時点における集中値を算出することができる。この時、対象者99が注視状態又は非注視状態であるか、又は、推移状態であるかに基づいて第1時点のパフォーマンス値と第2時点のパフォーマンス値とのそれぞれに対する重み(つまり、第1時点でのパフォーマンス値をどれほど第2時点のパフォーマンス値に影響させるか)を変更することで、より正確な集中値を算出することが可能となる。 According to this, it is possible to calculate the concentration value at the second point in time, taking into consideration the performance values calculated at the first point in time and the second point in time. At this time, the weight (that is, the first A more accurate concentration value can be calculated by changing the degree to which the performance value at the time point affects the performance value at the second time point.
 また、本実施の形態における集中値算出方法は、対象者99が撮像された画像ストリームを取得し、複数の集中対象領域であって、複数の集中対象領域の各々が、対象者99が注視すべき複数の集中対象物の各々に対応する複数の集中対象領域を取得し、取得された画像ストリームから、対象者99の顔の向き又は視線の向きを算出し、算出した対象者99の顔の向き又は視線の向きと、取得された複数の集中対象領域とに基づいて、対象者99が複数の集中対象物のいずれかを注視している注視状態であるか否かを判定し、判定結果に基づいて対象者99の集中値を算出して出力する。 In addition, the concentration value calculation method according to the present embodiment obtains an image stream in which the subject 99 is captured, and obtains a plurality of concentration target areas, each of which is a concentration target area that the subject 99 is gazing at. A plurality of focused target areas corresponding to each of a plurality of focused targets are acquired, the direction of the face or the direction of the line of sight of the target person 99 is calculated from the acquired image stream, and the calculated face direction of the target person 99 is calculated. Based on the orientation or the direction of the line of sight and the acquired plurality of focused target areas, it is determined whether or not the target person 99 is in a gaze state in which one of the plurality of focused targets is being gazed at, and a determination result is obtained. , the concentration value of the subject 99 is calculated and output.
 このような集中値算出方法では、上記に記載の集中値算出システムと同様の効果を奏することができる。 Such a concentration value calculation method can provide the same effects as the concentration value calculation system described above.
 また、例えば、上記に記載の集中値算出方法をコンピュータに実行させるためのプログラムとしてもよい。 Also, for example, it may be a program for causing a computer to execute the concentration value calculation method described above.
 これによれば、上記に記載の集中値算出方法をコンピュータに実行させることができる。 According to this, it is possible to cause the computer to execute the concentration value calculation method described above.
 (その他の実施の形態)
 以上、本開示に係る集中値算出システム、集中値算出方法、およびプログラムにについて、上記実施の形態等に基づいて説明したが、本開示は、上記の実施の形態に限定されるものではない。例えば、各実施の形態等に対して当業者が思いつく各種変形を施して得られる形態や、本開示の趣旨を逸脱しない範囲で各実施の形態における構成要素および機能を任意に組み合わせることで実現される形態も本開示に含まれる。
(Other embodiments)
Although the concentration value calculation system, the concentration value calculation method, and the program according to the present disclosure have been described above based on the above embodiments and the like, the present disclosure is not limited to the above embodiments. For example, a form obtained by applying various modifications that a person skilled in the art can think of for each embodiment, etc., or a form obtained by arbitrarily combining the constituent elements and functions of each embodiment within the scope of the present disclosure. Also included in the present disclosure is the form of
 例えば、上記した演算装置のみを備え、演算装置を外部の撮像装置、外部の記憶装置、及び外部の出力装置に接続する構成により、集中値算出システムを演算装置のみによって実現してもよい。このように、撮像装置、記憶装置、及び出力装置は必須の構成ではない。 For example, the centralized value calculation system may be realized only by the arithmetic device by providing only the arithmetic device described above and connecting the arithmetic device to an external imaging device, an external storage device, and an external output device. Thus, the imaging device, storage device, and output device are not essential components.
 また、例えば、図8に示すように、複数の集中対象物のそれぞれは、物理的な物体でなくてもよい。図では、コンピュータのディスプレイ96に表示された第1アプリケーションウインドウ97bが第1集中対象物として、第2アプリケーションウインドウ98bが第2集中対象物として本開示の内容が適用されてもよい。昨今のディスプレイ装置の大型化に伴い、従来の技術では対象者の集中値を高い正確度で算出することができなかったが、本開示の内容を適用することでこれを可能とすることができる。 Also, for example, as shown in FIG. 8, each of the plurality of focused objects may not be a physical object. In the figure, the first application window 97b displayed on the display 96 of the computer may be the first focus object, and the second application window 98b may be the second focus object, and the subject matter of the present disclosure may be applied. With the recent increase in size of display devices, it was not possible to calculate the concentration value of a subject with high accuracy using conventional techniques, but this can be made possible by applying the contents of the present disclosure. .
 また、図9に示すように、上記の実施の形態における集中値算出部に機械学習によって学習された学習モデルを利用することもできる。図9は、別例に係る集中値算出部の機能構成を示すブロック図である。この例では、演算装置10には、実施の形態における集中値算出部13に代えて集中値算出部13aが備えられる。集中値算出部13aは、取得した画像ストリーム及び取得した集中対象領域を集中値算出モデル13bに入力することで、対象者99の集中値を直接的に出力させることができる。そして、対象者99の集中値として、集中値算出モデル13bから出力された出力結果をそのまま出力する。 Also, as shown in FIG. 9, a learning model learned by machine learning can be used in the concentration value calculation unit in the above embodiment. FIG. 9 is a block diagram showing a functional configuration of a concentration value calculation unit according to another example. In this example, the computing device 10 includes a concentration value calculator 13a in place of the concentration value calculator 13 in the embodiment. The concentration value calculation unit 13a can directly output the concentration value of the subject 99 by inputting the obtained image stream and the obtained concentration target region to the concentration value calculation model 13b. Then, as the concentration value of the subject 99, the output result output from the concentration value calculation model 13b is output as it is.
 集中値算出モデル13bは、あらかじめ、機械学習によって、画像ストリーム及び集中対象領域と、集中値との相関関係が学習された学習モデルである。なお、本例では、集中値算出システム100は、集中値算出モデル13bを生成する(学習させる)ために、モデル生成部13cをさらに備える。モデル生成部13cは、集中値算出モデル13bを生成するために、画像ストリーム及び集中対象領域の2つの情報に対応する入力データと、入力データに対する正解の(又は正解及び不正解を含む)出力データとを教師データとして用いる。 The concentration value calculation model 13b is a learning model in which the correlation between the image stream, the concentration target region, and the concentration value is learned in advance by machine learning. In this example, the concentration value calculation system 100 further includes a model generation unit 13c for generating (learning) the concentration value calculation model 13b. In order to generate the concentration value calculation model 13b, the model generation unit 13c generates input data corresponding to the two pieces of information of the image stream and the concentration target area, and correct (or correct and incorrect) output data for the input data. and are used as training data.
 この例では、入力データとして、画像ストリーム及び集中対象領域に対応する教師画像/教師領域D1が学習用に入力されている。また、この例では、出力データとして、対象者99の教師集中値D2が学習用に入力されている。そして、教師画像/教師領域D1及び教師集中値D2を組み合わせたデータセットにより、集中値算出モデル13bを調整する。例えば、複数層の階層からなるニューラルネットワークで構成された集中値算出モデル13bである場合、各ニューロンに付される重み係数をバックプロパゲーションなどの手法によって調整して入力データに対して、適切な出力データが得られるように機械学習が行われる。 In this example, as input data, an image stream and a teacher image/teacher region D1 corresponding to the concentration target region are input for learning. In this example, the teacher concentration value D2 of the subject 99 is input as the output data for learning. Then, the concentration value calculation model 13b is adjusted using a data set combining the teacher image/teacher region D1 and the teacher concentration value D2. For example, in the case of the centralized value calculation model 13b composed of a neural network consisting of a plurality of layers, the weighting coefficient assigned to each neuron is adjusted by a method such as back propagation to obtain an appropriate value for the input data. Machine learning is performed to obtain output data.
 そして、集中値算出部13aが学習済みの集中値算出モデル13bに対して、取得した画像ストリーム及び取得した集中対象領域を入力することで、適切な対象者99の集中値が出力される。このように、集中値算出部13aによる集中値の算出は、機械学習された学習モデルを利用して実現することもできる。 Then, the concentration value calculation unit 13a inputs the acquired image stream and the acquired concentration target area to the learned concentration value calculation model 13b, so that an appropriate concentration value of the subject 99 is output. In this way, the calculation of the concentration value by the concentration value calculation unit 13a can also be realized using a machine-learned learning model.
 なお、この例において、モデル生成部13cを備える集中値算出システム100について説明したが、集中値算出システム100を製造するに際して、学習済みの集中値算出モデル13bを演算装置10内に記録しておき、その後は、学習の過程を経ることなく、記録された集中値算出モデル13bを利用するのみの構成も実現可能である、この場合においては、モデル生成部13cを備えることなく集中値算出システム100を実現することも可能である。 In this example, the concentration value calculation system 100 including the model generating unit 13c has been described. After that, it is also possible to realize a configuration in which only the recorded concentration value calculation model 13b is used without going through the learning process. is also possible.
 また、例えば、本開示は、集中値算出システムとして実現できるだけでなく、集中値算出システムの各構成要素が行う処理をステップとして含むプログラム、および、そのプログラムを記録したコンピュータ読み取り可能な記録媒体として実現することもできる。プログラムは、記録媒体に予め記録されていてもよく、あるいは、インターネットなどを含む広域通信網を介して記録媒体に供給されてもよい。 In addition, for example, the present disclosure can be realized not only as a centralized value calculation system, but also as a program including, as steps, processes performed by each component of the centralized value calculation system, and a computer-readable recording medium recording the program. You can also The program may be pre-recorded on a recording medium, or may be supplied to the recording medium via a wide area network including the Internet.
 つまり、上述した包括的または具体的な態様は、システム、装置、集積回路、コンピュータプログラムまたはコンピュータ読み取り可能な記録媒体で実現されてもよく、システム、装置、集積回路、コンピュータプログラムおよび記録媒体の任意な組み合わせで実現されてもよい。 That is, the general or specific aspects described above may be embodied in a system, device, integrated circuit, computer program or computer readable recording medium, and any of the system, device, integrated circuit, computer program and recording medium may be implemented. may be implemented in any combination.
  11 画像取得部
  12 領域取得部
  13 集中値算出部
  99 対象者
 100 集中値算出システム
REFERENCE SIGNS LIST 11 image acquisition unit 12 area acquisition unit 13 concentration value calculation unit 99 subject 100 concentration value calculation system

Claims (14)

  1.  対象者が撮像された画像ストリームを取得する画像取得部と、
     複数の集中対象領域であって、前記複数の集中対象領域の各々が、前記対象者が注視すべき複数の集中対象物の各々に対応する、複数の集中対象領域を取得する領域取得部と、
     前記対象者の集中値を算出する集中値算出部と、を備え、
     前記集中値算出部は、
      取得された前記画像ストリームから、前記対象者の顔の向き又は視線の向きを算出し、
      算出した前記対象者の顔の向き又は視線の向きと、取得された前記複数の集中対象領域とに基づいて、前記対象者が前記複数の集中対象物のいずれかを注視している注視状態か否かを判定し、
      判定結果に基づいて前記対象者の集中値を算出して出力する
     集中値算出システム。
    an image acquisition unit that acquires an image stream in which a subject is imaged;
    a region acquisition unit configured to acquire a plurality of concentration target regions, each of which corresponds to each of a plurality of concentration targets to be gazed at by the subject;
    a concentration value calculation unit that calculates the concentration value of the subject,
    The concentration value calculation unit
    calculating the orientation of the face or the orientation of the line of sight of the subject from the acquired image stream;
    Based on the calculated face direction or line-of-sight direction of the subject and the plurality of acquired focus target areas, whether the subject is in a gaze state gazing at any one of the plurality of focused objects determine whether or not
    A concentration value calculation system for calculating and outputting a concentration value of the subject based on a determination result.
  2.  前記複数の集中対象領域の各々は、前記対象者に対して当該集中対象領域に対応する集中対象物を注視させる指示が前記対象者に対して提示されたときに撮像された領域決定画像に基づいて決定されて、あらかじめ記憶部に記憶されており、
     前記領域取得部は、前記記憶部を参照して、前記複数の集中対象領域を取得する
     請求項1に記載の集中値算出システム。
    Each of the plurality of concentration target areas is based on an area determination image captured when the subject is presented with an instruction to gaze at a concentration target corresponding to the concentration target area. and is stored in advance in the storage unit,
    The concentration value calculation system according to claim 1, wherein the area acquisition unit acquires the plurality of concentration target areas by referring to the storage unit.
  3.  前記複数の集中対象領域の各々は、前記対象者に対して当該集中対象領域に対応する集中対象物の一端及び他端を注視させる指示に対して、前記一端を注視する前記対象者の顔の向き又は視線の向きと、前記他端を注視する前記対象者の顔の向き又は視線の向きとの間の領域として決定されて、あらかじめ記憶部に記憶されている
     請求項2に記載の集中値算出システム。
    Each of the plurality of concentration target areas is configured such that, in response to an instruction to cause the subject to gaze at one end and the other end of the concentration target corresponding to the concentration target area, the face of the subject gazes at the one end. 3. The concentration value according to claim 2, which is determined as an area between the orientation or the direction of the line of sight and the orientation of the face or the direction of the line of sight of the subject gazing at the other end, and stored in advance in a storage unit. calculation system.
  4.  前記集中対象領域は、前記対象者が使用するスペースごとにあらかじめ設定されて、記憶部に記憶されており、
     前記領域取得部は、前記記憶部を参照して、前記複数の集中対象領域を取得する
     請求項1に記載の集中値算出システム。
    The concentration target area is set in advance for each space used by the target person and stored in a storage unit,
    The concentration value calculation system according to claim 1, wherein the area acquisition unit acquires the plurality of concentration target areas by referring to the storage unit.
  5.  前記集中対象領域は、前記対象者ごとにあらかじめ設定されて、記憶部に記憶されており、
     前記領域取得部は、前記記憶部を参照して、前記複数の集中対象領域を取得する
     請求項1に記載の集中値算出システム。
    The concentration target area is set in advance for each target person and stored in a storage unit,
    The concentration value calculation system according to claim 1, wherein the area acquisition unit acquires the plurality of concentration target areas by referring to the storage unit.
  6.  前記対象者の顔の向きは、取得された前記画像ストリームから、前記対象者の顔の法線ベクトルとして算出され、
     前記複数の集中対象領域のそれぞれは、前記対象者の位置を中心とした所定の角度範囲を有し、
     前記集中値算出部は、算出された前記法線ベクトルが、前記所定の角度範囲内にある場合に、前記対象者が当該所定の角度範囲を有する集中対象領域に対応する集中対象物を注視している前記注視状態と判定する
     請求項1~5のいずれか1項に記載の集中値算出システム。
    the orientation of the subject's face is calculated from the acquired image stream as a normal vector of the subject's face;
    each of the plurality of focused target areas has a predetermined angular range centered on the subject's position;
    The concentration value calculation unit, when the calculated normal vector is within the predetermined angle range, causes the subject to gaze at the concentration target object corresponding to the concentration target region having the predetermined angle range. 6. The concentration value calculation system according to any one of claims 1 to 5, wherein the gaze state is determined to be one of the
  7.  前記集中値算出部は、前記注視状態でないと判定した場合に、さらに、前記対象者が、前記複数の集中対象物のうちの2つの間を視線が推移する推移状態であるか否かを判定し、判定結果に基づいて前記対象者の集中値を算出する
     請求項1~6のいずれか1項に記載の集中値算出システム。
    The concentration value calculation unit further determines whether or not the target person is in a transition state in which the line of sight transitions between two of the plurality of concentration objects when determining that the gaze state is not the state. and calculating the concentration value of the subject based on the determination result.
  8.  前記集中値算出部は、
      さらに、取得された前記画像ストリームから、前記対象者の単位集中値であるパフォーマンス値を算出し、
      前記対象者が前記注視状態であると判定された場合には、算出した前記パフォーマンス値に第1係数を乗じ、前記対象者が前記推移状態であると判定された場合には、算出した前記パフォーマンス値に前記第1係数以下の第2係数を乗じ、前記対象者が前記注視状態でないと判定され、かつ、前記推移状態でないと判定された場合には、算出した前記パフォーマンス値に前記第2係数よりも小さい第3係数を乗じて前記対象者の集中値を算出する
     請求項7に記載の集中値算出システム。
    The concentration value calculation unit
    Further, calculating a performance value, which is a unit concentration value of the subject, from the acquired image stream,
    When the target person is determined to be in the gaze state, the calculated performance value is multiplied by a first coefficient, and when the target person is determined to be in the transition state, the calculated performance value is multiplied by a second coefficient equal to or less than the first coefficient, and when it is determined that the target person is not in the gaze state and is not in the transition state, the calculated performance value is added to the second coefficient 8. The concentration value calculation system according to claim 7, wherein the concentration value of the subject is calculated by multiplying by a third coefficient smaller than .
  9.  前記集中値算出部は、
      さらに、取得された前記画像ストリームから、前記対象者の単位集中値であるパフォーマンス値を、第1時点及び前記第1時点に連続する第2時点で算出し、
      前記対象者が前記注視状態であると判定された場合、及び、前記対象者が前記注視状態でないと判定され、かつ、前記推移状態でないと判定された場合には、算出した前記第1時点におけるパフォーマンス値に第1重み係数を乗じた第1値、及び、算出した前記第2時点におけるパフォーマンス値に第2重み係数であって前記第1重み係数と1との差分である第2重み係数を乗じた第2値を加算し、
      前記対象者が前記推移状態であると判定された場合には、算出した前記第1時点におけるパフォーマンス値に前記第1重み係数とは異なる第3重み係数を乗じた第3値、及び、算出した前記第2時点におけるパフォーマンス値に第4重み係数であって前記第3重み係数と1との差分である第4重み係数を乗じた第4値を加算して前記対象者の集中値を算出し、
     前記第1重み係数、及び、前記第3重み係数は、0より大きく1より小さい数値である
     請求項7に記載の集中値算出システム。
    The concentration value calculation unit
    Furthermore, from the acquired image stream, a performance value, which is a unit concentration value of the subject, is calculated at a first time point and a second time point following the first time point,
    When it is determined that the target person is in the gaze state, and when it is determined that the target person is not in the gaze state and is not in the transitional state, A first value obtained by multiplying the performance value by a first weighting factor, and a second weighting factor, which is a difference between the first weighting factor and 1, to the calculated performance value at the second point in time. Add the multiplied second value,
    When the subject is determined to be in the transition state, a third value obtained by multiplying the calculated performance value at the first time point by a third weighting factor different from the first weighting factor, and the calculated adding a fourth value obtained by multiplying the performance value at the second time point by a fourth weighting factor which is a difference between the third weighting factor and 1 to calculate the concentration value of the subject; ,
    8. The concentration value calculation system according to claim 7, wherein the first weighting factor and the third weighting factor are numerical values greater than 0 and less than 1.
  10.  前記対象者の顔の向き又は視線の向きは、前記画像ストリームと、前記画像ストリームを撮像する撮像装置に対する前記対象者の顔の向き又は視線の向きとの相関関係が機械学習によって学習された向き算出モデルに対して前記画像ストリームを入力することで、出力結果として算出される
     請求項1~9のいずれか1項に記載の集中値算出システム。
    The direction of the subject's face or the direction of the line of sight is the direction in which the correlation between the image stream and the direction of the subject's face or the direction of the line of sight with respect to an imaging device that captures the image stream is learned by machine learning. The concentration value calculation system according to any one of claims 1 to 9, which is calculated as an output result by inputting the image stream to a calculation model.
  11.  対象者が撮像された画像ストリームを取得し、
     複数の集中対象領域であって、前記複数の集中対象領域の各々が、前記対象者が注視すべき複数の集中対象物の各々に対応する複数の集中対象領域を取得し、
     取得された前記画像ストリームから、前記対象者の顔の向き又は視線の向きを算出し、
     算出した前記対象者の顔の向き又は視線の向きと、取得された前記複数の集中対象領域とに基づいて、前記対象者が前記複数の集中対象物のいずれかを注視している注視状態であるか否かを判定し、
     判定結果に基づいて前記対象者の集中値を算出して出力する
     集中値算出方法。
    obtaining an image stream in which the subject is imaged;
    obtaining a plurality of focused regions of interest, each of the plurality of focused regions of interest corresponding to each of a plurality of focused objects to be gazed at by the subject;
    calculating the orientation of the face or the orientation of the line of sight of the subject from the acquired image stream;
    a gaze state in which the subject is gazing at any one of the plurality of focused objects based on the calculated face orientation or line-of-sight orientation of the subject and the plurality of acquired focused target areas; determine whether there is
    A concentration value calculation method for calculating and outputting a concentration value of the subject based on a determination result.
  12.  請求項11に記載の集中値算出方法をコンピュータに実行させるための
     プログラム。
    A program for causing a computer to execute the concentration value calculation method according to claim 11 .
  13.  対象者が撮像された画像ストリームを取得する画像取得部と、
     複数の集中対象領域であって、前記複数の集中対象領域の各々が、前記対象者が注視すべき複数の集中対象物の各々に対応する、複数の集中対象領域を取得する領域取得部と、
     前記対象者の集中値を算出する集中値算出部と、を備え、
     前記集中値算出部は、前記画像ストリーム及び前記複数の集中対象領域と、前記対象者の集中値との相関関係が機械学習によって学習された集中値算出モデルに対して前記画像ストリーム及び前記複数の集中対象領域を入力することで、前記対象者の集中値を算出する
     集中値算出システム。
    an image acquisition unit that acquires an image stream in which a subject is imaged;
    a region acquisition unit configured to acquire a plurality of concentration target regions, each of which corresponds to each of a plurality of concentration targets to be gazed at by the subject;
    a concentration value calculation unit that calculates the concentration value of the subject,
    The concentration value calculation unit is configured to apply the image stream and the plurality of concentration target regions to a concentration value calculation model in which a correlation between the image stream and the plurality of concentration target regions and the concentration value of the subject has been learned by machine learning. A concentration value calculation system for calculating a concentration value of the subject by inputting a concentration target region.
  14.  請求項13に記載の集中値算出モデルを生成するモデル生成部を備える、
     集中値算出モデル生成システム。
    A model generation unit that generates the concentration value calculation model according to claim 13,
    Concentrated value calculation model generation system.
PCT/JP2022/012007 2021-03-30 2022-03-16 Concentration value calculation system, concentration value calculation method, program, and concentration value calculation model generation system WO2022209912A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2023510917A JP7531148B2 (en) 2021-03-30 2022-03-16 Concentration value calculation system, concentration value calculation method, and program

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2021057345 2021-03-30
JP2021-057345 2021-03-30

Publications (1)

Publication Number Publication Date
WO2022209912A1 true WO2022209912A1 (en) 2022-10-06

Family

ID=83459109

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2022/012007 WO2022209912A1 (en) 2021-03-30 2022-03-16 Concentration value calculation system, concentration value calculation method, program, and concentration value calculation model generation system

Country Status (2)

Country Link
JP (1) JP7531148B2 (en)
WO (1) WO2022209912A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000259834A (en) * 1999-03-11 2000-09-22 Toshiba Corp Registering device and method for person recognizer
JP2016111612A (en) * 2014-12-09 2016-06-20 三星電子株式会社Samsung Electronics Co.,Ltd. Content display device
JP2017140107A (en) * 2016-02-08 2017-08-17 Kddi株式会社 Concentration degree estimation device
WO2020116181A1 (en) * 2018-12-03 2020-06-11 パナソニックIpマネジメント株式会社 Concentration degree measurement device and concentration degree measurement method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000259834A (en) * 1999-03-11 2000-09-22 Toshiba Corp Registering device and method for person recognizer
JP2016111612A (en) * 2014-12-09 2016-06-20 三星電子株式会社Samsung Electronics Co.,Ltd. Content display device
JP2017140107A (en) * 2016-02-08 2017-08-17 Kddi株式会社 Concentration degree estimation device
WO2020116181A1 (en) * 2018-12-03 2020-06-11 パナソニックIpマネジメント株式会社 Concentration degree measurement device and concentration degree measurement method

Also Published As

Publication number Publication date
JP7531148B2 (en) 2024-08-09
JPWO2022209912A1 (en) 2022-10-06

Similar Documents

Publication Publication Date Title
Arabadzhiyska et al. Saccade landing position prediction for gaze-contingent rendering
Lemaignan et al. From real-time attention assessment to “with-me-ness” in human-robot interaction
US10665022B2 (en) Augmented reality display system for overlaying apparel and fitness information
JP6165846B2 (en) Selective enhancement of parts of the display based on eye tracking
US10607063B2 (en) Information processing system, information processing method, and recording medium for evaluating a target based on observers
US10832483B2 (en) Apparatus and method of monitoring VR sickness prediction model for virtual reality content
Cidota et al. Workspace awareness in collaborative AR using HMDS: a user study comparing audio and visual notifications
US11442685B2 (en) Remote interaction via bi-directional mixed-reality telepresence
JP4868360B2 (en) Interest trend information output device, interest trend information output method, and program
KR20210043174A (en) Method for providing exercise coaching function and electronic device performing thereof
WO2013069344A1 (en) Gaze position estimation system, control method for gaze position estimation system, gaze position estimation device, control method for gaze position estimation device, program, and information storage medium
KR20190048144A (en) Augmented reality system for presentation and interview training
WO2022209912A1 (en) Concentration value calculation system, concentration value calculation method, program, and concentration value calculation model generation system
JP2016111612A (en) Content display device
CN115762772B (en) Method, device, equipment and storage medium for determining emotional characteristics of target object
Lee et al. A study on virtual reality sickness and visual attention
Lin et al. An eye-tracking and head-control system using movement increment-coordinate method
JP7233631B1 (en) posture improvement system
JP2008046802A (en) Interaction information output device, interaction information output method and program
CN111651043B (en) Augmented reality system supporting customized multi-channel interaction
WO2022070747A1 (en) Assist system, assist method, and assist program
Kao et al. Eye gaze tracking based on pattern voting scheme for mobile device
TWI674518B (en) Calibration method of eye-tracking and device thereof
Boczon State of the art: eye tracking technology and applications
US20140062997A1 (en) Proportional visual response to a relative motion of a cephalic member of a human subject

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22780132

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2023510917

Country of ref document: JP

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: 11202305456W

Country of ref document: SG

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 22780132

Country of ref document: EP

Kind code of ref document: A1