WO2016092933A1 - Information processing device, information processing method, and program - Google Patents

Information processing device, information processing method, and program Download PDF

Info

Publication number
WO2016092933A1
WO2016092933A1 PCT/JP2015/076633 JP2015076633W WO2016092933A1 WO 2016092933 A1 WO2016092933 A1 WO 2016092933A1 JP 2015076633 W JP2015076633 W JP 2015076633W WO 2016092933 A1 WO2016092933 A1 WO 2016092933A1
Authority
WO
WIPO (PCT)
Prior art keywords
information
action
user
additional information
image
Prior art date
Application number
PCT/JP2015/076633
Other languages
French (fr)
Japanese (ja)
Inventor
卓 青木
亮 向山
田村 錬志
賢一 山浦
Original Assignee
ソニー株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニー株式会社 filed Critical ソニー株式会社
Publication of WO2016092933A1 publication Critical patent/WO2016092933A1/en

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B69/00Training appliances or apparatus for special sports
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B71/00Games or sports accessories not covered in groups A63B1/00 - A63B69/00
    • A63B71/06Indicating or scoring devices for games or players, or for other sports activities

Definitions

  • This disclosure relates to an information processing apparatus, an information processing method, and a program.
  • Patent Document 1 in order to more accurately grasp the state of sports play by a user, the occurrence of a play event in sports is detected based on sensor information, and a play image is captured according to the occurrence of the play event. Techniques for controlling the are described.
  • Patent Document 1 the technique as described in Patent Document 1 is intended to more accurately grasp the state of sports play, and does not necessarily provide a better experience for users who exclusively watch play. It is not intended.
  • a new and improved information processing apparatus and information processing capable of providing more appropriate information reflecting an evaluation of an action based on an image to a user who views the image including the action Propose methods and programs.
  • an evaluation information acquisition unit that acquires action evaluation information obtained by evaluating the action based on an action image including an action of an actor, and user situation information indicating a situation of a user who views the action image are acquired.
  • An information processing apparatus comprising: a situation information acquisition unit; and an additional information generation unit that generates additional information that is presented to the user together with the action image and expresses at least part of the action evaluation information according to the user situation information Is provided.
  • acquiring action evaluation information obtained by evaluating the action based on an action image including an actor's action and acquiring user situation information indicating a situation of a user viewing the action image
  • the function which acquires the action evaluation information which evaluated the said action based on the action image containing an action of an actor and the function which acquires the user condition information which shows the condition of the user who views the said action image
  • Providing information that evaluates an action based on an image together with the action image helps the user understand the action because the provided information is linked to the content of the image. However, since what kind of information is appropriate may vary depending on the user's situation, more appropriate information is provided to the user by generating additional information based on action evaluation based on the user's situation. can do.
  • FIG. 14 is a flowchart illustrating an example of processing for controlling the amount of additional information according to an embodiment of the present disclosure.
  • FIG. 5 is a diagram illustrating an example of additional information for a user who views an action image alone in an embodiment of the present disclosure.
  • FIG. 4 is a diagram illustrating an example of additional information for a user who views a plurality of action images in an embodiment of the present disclosure. It is a figure showing the 1st example of other additional information in one embodiment of this indication. It is a figure showing the 2nd example of other additional information in one embodiment of this indication. It is a figure showing the 3rd example of other additional information in one embodiment of this indication.
  • FIG. 3 is a block diagram illustrating a hardware configuration example of an information processing apparatus according to an embodiment of the present disclosure.
  • FIG. 1 is a diagram illustrating a configuration example of a system according to an embodiment of the present disclosure.
  • a system 10 includes a camera 100, a server 200, and an output device 300 (a television 300a and a recorder 300b).
  • the camera 100 captures an action image including an actor's action.
  • an actor is a person or thing that performs some action.
  • Actions can include all kinds of actions that can be evaluated, such as sports play, performances such as dance, and professional skills. For example, if the action is sport play, the actor is also called a player.
  • actors such as performers, dancers, and testers are called in various ways depending on the type of action.
  • the camera 100 can be, for example, a digital video camera that can acquire a still image or a moving image.
  • the camera 100 has at least one image sensor and records an image including an actor as a subject as electronic data.
  • the camera 100 may include an image processing circuit that performs processing for converting electronic image data into a format suitable for the server 200 and the output device 300.
  • the camera 100 may include an information processing circuit that realizes part or all of the functions of the information processing apparatus described later.
  • the information processing circuit can be configured by applying a hardware configuration of an information processing apparatus to be described later, for example.
  • the server 200 receives an action image from the camera 100 (S11), and generates or receives action evaluation information that evaluates an action based on the action image. Further, in an example, the server 200 receives user status information indicating the status of the user viewing the action image from the output device 300 (S13a), and adds the additional information generated according to the user status information together with the action image. It transmits to the output device 300 (S12). Here, the additional information includes at least a part of the action evaluation information. Alternatively, the server 200 transmits the action image together with the action evaluation information to the output device 300 as it is (S12). In this case, the generation of additional information using the action evaluation information is performed by the output device 300 (S13b).
  • the server 200 can be realized by one or a plurality of server devices on a network. Each server device can be configured by applying a hardware configuration of an information processing device to be described later, for example.
  • the output device 300 receives an action image from the server 200 (S12), outputs the action image together with additional information, and presents it to the user.
  • the additional information may be generated by the server 200 based on the user situation information (S13a) transmitted by the output device 300, and transmitted from the server 200 to the output device 300 together with the action image (S12).
  • the output device 300 receives the action evaluation information together with the action image from the server 200 (S12), generates additional information from the action evaluation information based on the user situation information (S13b), and then displays the action image and additional information. It may be output.
  • the output device 300 is illustrated as a television 300a and a recorder 300b.
  • an action image transmitted from the server 200 by a broadcast wave is received, recorded, and saved by the recorder 300b, and later played and output via the television 300a.
  • the action image may be received and output directly by the television 300a without using the recorder 300b.
  • the television 300a or the recorder 300b may include an information processing circuit that performs information processing related to generation of additional information and the like.
  • the information processing circuit can be configured by applying a hardware configuration of an information processing apparatus to be described later, for example.
  • the example of the output device 300 is not limited to the combination of the television 300a and the recorder 300b in the illustrated example.
  • the television 300a may be used alone as the output device 300.
  • any device having a function of outputting a received image can be used as the output device 300.
  • a smartphone, a tablet, a personal computer, or the like may be used as the output device 300.
  • FIG. 2 is a diagram illustrating a functional configuration example of the information processing apparatus according to an embodiment of the present disclosure.
  • the information processing apparatus includes an image acquisition unit 201, an evaluation information generation unit 203, an evaluation information acquisition unit 205, a situation information acquisition unit 207, an additional information generation unit 209, and a screen generation unit 211.
  • the functional configuration of the information processing apparatus is realized by the camera 100, the server 200, or the output device 300 in the system 10 as described above. These functional configurations may be realized by being integrated in a single device, or may be realized by being distributed to a plurality of devices.
  • the functional configuration described below is realized by, for example, a processor such as a CPU of one or a plurality of devices operating according to a program stored in a memory or storage.
  • the image acquisition unit 201 acquires an action image taken by the camera 100.
  • the action image is an image including an action by an actor.
  • the image acquisition unit 201 may be realized by, for example, a communication device that receives image data from the camera 100 connected via a network or the like by wired or wireless communication.
  • the image acquisition unit 201 may be a software interface such as a driver program that acquires data from such a communication device. For example, when a screen generation unit 211 described later is realized in the camera 100, the image acquisition unit 201 may acquire an action image via an internal communication path such as a bus.
  • the evaluation information generation unit 203 generates action evaluation information that evaluates an action based on the action image.
  • the evaluation information generation unit 203 acquires an action image from the camera 100. Similar to the image acquisition unit 201 described above, the action image may be received by communication between apparatuses, or may be acquired via an internal communication path.
  • the evaluation information generation unit 203 performs image analysis on the action image, and estimates the displacement, rotation, posture, and the like of the actor. Furthermore, the evaluation information generation unit 203 evaluates the estimated information. At this time, the evaluation information generation unit 203 may evaluate the action by applying an action determination or scoring standard. Judgment or scoring standards are determined for each type of action (for example, sports, performance, professional skills, etc.).
  • action evaluation information or additional information based thereon may be provided to a referee or a judge.
  • the evaluation information generation unit 203 may generate action evaluation information based on information other than the action image. For example, the evaluation information generation unit 203 may generate action evaluation information based on a detection value obtained from a sensor such as an inertial sensor attached to the actor in addition to the action image.
  • the evaluation information acquisition unit 205 acquires the action evaluation information generated by the evaluation information generation unit 203.
  • the evaluation information generation unit 203 is realized by the camera 100, or the system
  • the evaluation information acquisition unit 205 may be realized by a communication device that receives the action evaluation information through communication between the devices. It may be a software interface for acquiring data from such a communication device.
  • the evaluation information acquisition unit 205 uses an internal communication path such as a bus. It can be realized as a software interface for acquiring action evaluation information through the interface.
  • the status information acquisition unit 207 acquires user status information indicating the status of the user who views the action image via the output device 300.
  • the user status information indicates, for example, the time for the user to refer to the additional information presented together with the action image, the user's familiarity regarding the action included in the action image, or the number of users. Can contain information.
  • the situation information acquisition unit 207 also receives user situation information through communication between devices when the additional information generation unit 209 is realized in a device different from the output device 300. It is realized by a device or a software interface for acquiring data from such a communication device.
  • the status information acquisition unit 207 is a software interface that acquires user status information via an internal communication path such as a bus. Can be realized.
  • the number of users who view an action image is detected by analyzing an image including a user in front of the display acquired by a camera attached to the display in the output device 300. Further, for example, a user who views an action image recognizes the user's face from an image acquired by a camera attached to a display in the output device 300 or when the output device 300 accepts a user login. Is specified by information of the logged-in user. For each identified user, by referring to the profile and schedule registered in the service, content usage history, etc., the depth of the user's knowledge about the action, the time to watch the action image, etc. Can be estimated.
  • the additional information generation unit 209 generates additional information that represents at least a part of the action evaluation information acquired by the evaluation information acquisition unit 205 based on the user situation information acquired by the situation information acquisition unit 207. As described above, the additional information is presented to the user together with the action image in the output device 300. Since the additional information expresses action evaluation information, it does not necessarily include the information included in the action evaluation information as it is.
  • the additional information generation unit 209 may generate additional information by adding some processing to the action evaluation information.
  • the additional information includes visual information displayed on the screen together with the action image. More specifically, for example, the additional information includes text or graphics. The text or graphic may be superimposed on the action image or displayed in the vicinity of the action image. A specific example of additional information will be described later.
  • the additional information generation unit 209 changes the amount of information and the type of information expressed by the additional information based on the user situation information.
  • the additional information generation unit 209 selects information included in the additional information from the action evaluation information generated by the evaluation information generation unit 203.
  • the embodiment of the present disclosure is such an example. Is not limited.
  • the evaluation information generation unit 203 generates action evaluation information according to the amount of information or the type of information that the additional information generation unit 209 has decided to generate according to the user situation information. Good. That is, in the embodiment of the present disclosure, the additional information may be generated by extracting part or all of the action evaluation information generated regardless of the user's situation, or generated according to the situation by the user. It may be generated based on the action evaluation information.
  • the screen generation unit 211 generates a screen including the action image acquired by the image acquisition unit 201 and the additional information generated by the additional information generation unit 209.
  • the additional information generation unit 209 includes text or graphics generated based on the action evaluation information.
  • the screen generation unit 211 arranges these texts or graphics so as to be superimposed on the action image or displayed in the vicinity of the action image.
  • the screen generation unit 211 may draw a screen including an action image and additional information, and provide the image data of the drawn screen for display by the output device 300.
  • the screen generation unit 211 generates image data for drawing an action image on the screen and image data for drawing additional information, respectively, and the output device 300 generates an action image and an image based on these image data. Additional information may be displayed on the screen.
  • the screen generation unit 211 may be realized, for example, in the server 200 or in the output device 300.
  • the functional configuration of the information processing apparatus illustrated in FIG. 2 is described corresponding to an example in which the information processing apparatus includes a processor that mainly implements the additional information generation unit 209.
  • the information processing apparatus includes, in addition to the additional information generation unit 209, an evaluation information acquisition unit 205 that is an interface for acquiring action evaluation information, and a situation information acquisition unit 207 that is an interface for acquiring user situation information.
  • the processor of the information processing apparatus may further realize the evaluation information generation unit 203.
  • the evaluation information acquisition unit 205 can be an internal program interface.
  • the information processing apparatus may include an image acquisition unit 201, and the processor of the information processing apparatus may realize the screen generation unit 211.
  • the information processing apparatus is included in the camera 100 or the output apparatus 300, and the functional configuration of the information processing apparatus illustrated in FIG. 2 may be implemented internally in the camera 100 or the output apparatus 300. .
  • FIG. 3 is a diagram illustrating a first example of additional information according to an embodiment of the present disclosure.
  • FIG. 3 shows an action image 1000 and a graphic display of a rotation axis 1010, a rotation axis blur 1020, a rotation speed 1030, and a rotation axis shift 1040 transparently superimposed on the action image 1000.
  • the spin of figure skating is captured as an action.
  • the evaluation information generation unit 203 evaluates the spin based on the attitude of the skater estimated from the action image 1000.
  • the evaluation information generation unit 203 may estimate the posture by extracting a skater image from the action image 1000 that is an RGB image by using an image recognition technique.
  • the evaluation information generation unit 203 may extract a skater image from the action image 1000 in which the depth information is added to the RGB image, using image recognition and space recognition techniques. Instead of or together with the depth information, RGB images taken simultaneously from a plurality of positions may be used. Furthermore, the evaluation information generation unit 203 may use a detection value of an inertial sensor attached to the skater in order to estimate the attitude of the skater.
  • the evaluation information generating unit 203 can evaluate the spin executed by the skater.
  • the evaluation information generation unit 203 evaluates the time series change of the spin rotation axis (represented by a straight line in space) and the spin rotation speed.
  • the additional information generation unit 209 Based on the action evaluation information including such evaluation, the additional information generation unit 209 generates additional information to be presented together with the action image.
  • the additional information generation unit 209 expresses, on the action image 1000, the rotation of the axis relative to the foot of the skater (the ground contact point of the skate) along with the rotation axis 1010 of the spin as the rotation axis blur 1020.
  • the additional information generation unit 209 determines the position of the skater in the action image 1000 and the posture of the skater estimated based on the action image 1000 in order to identify the position of the skater's foot or the position of the rotation axis. Use the information. Further, the additional information generation unit 209 displays the rotation speed 1030 according to the position of the specified rotation axis. Further, in the illustrated example, the additional information generation unit 209 represents the transition of the position of the skater's foot on the action image 1000 as a rotation axis shift 1040. In addition to this, in skating, for example, graphics indicating the jump height and distance, edge depth, and the like may be superimposed on the action image 1000 and displayed.
  • FIG. 4 is a diagram illustrating a second example of additional information according to an embodiment of the present disclosure.
  • FIG. 4 shows an action image 1000 and a rotation speed graph 1050 displayed together with the action image 1000.
  • the additional information generation unit 209 draws the graph 1050 based on the rotation speed included in the action evaluation information generated by the evaluation information generation unit 203.
  • the additional information generation unit 209 reads, for example, an ideal change in the rotation speed recorded in the internal storage of the information processing apparatus and compares it with the rotation speed of the skater's spin shown in the action image 1000.
  • the graph 1050 may be drawn. Note that information on the ideal rotation speed may be read from a storage external to the information processing apparatus.
  • the graph 1050 is an example of additional information that does not necessarily need to be superimposed on the action image 1000.
  • the rotating shaft 1010, the rotating shaft blur 1020, the rotating speed 1030, and the rotating shaft shift 1040 described above with reference to FIG. 3 are basically transparently superimposed on the action image 1000 and reflected in the action image 1000. While the additional information is visible along with the skater image, the graph 1050 may be transparently superimposed on the action image 1000, but may be opaquely superimposed as illustrated. Further, it may be displayed separately in the vicinity thereof without being superimposed on the action image 1000.
  • the present embodiment may include a configuration in which the additional information generation unit 209 controls the information amount of the additional information according to the user situation when generating the additional information based on the user situation information.
  • the additional information generation unit 209 controls the information amount of the additional information according to the user situation when generating the additional information based on the user situation information.
  • FIG. 5 is a diagram illustrating an example of additional information with a small amount of information according to an embodiment of the present disclosure.
  • FIG. 5 shows an action image 1100 and a jump locus 1110 that is transparently superimposed on the action image 1100.
  • a figure skating jump is captured as an action.
  • the evaluation information generation unit 203 evaluates the jump based on the skater posture estimated from the action image 1100.
  • the estimation of the skater posture by the evaluation information generation unit 203 can be performed, for example, in the same manner as the spin example described above with reference to FIG.
  • the evaluation information generating unit 203 can evaluate the jump executed by the skater.
  • the evaluation information generation unit 203 performs skating in the action image 1100 at the time of jump crossing (the moment when the skates are separated from the ice surface) and at the time of landing (the moment when the skates are touching the ice surface). Identify shoe coordinates.
  • the additional information generation unit 209 draws the trajectory of the skate shoes in the action image 1100 between the jump crossing and landing as the jump trajectory 1110.
  • the jump trajectory 1110 shown in FIG. 5 includes the skater's jump trajectory 1111 currently shown in the action image 1100 and the last year's jump trajectory 1112 by the same skater.
  • the additional information generation unit 209 draws the last year's jump trajectory 1112 based on, for example, information read from the storage.
  • Information for drawing last year's jump trajectory 1112 is stored in, for example, a storage inside or outside the information processing apparatus as reference information.
  • texts “TODAY” and “LAST YEAR” and texts indicating the jump distances of the respective jumps are displayed.
  • the jump-related information captured in the action image 1100 is simply expressed by the graphic and text of the jump trajectory 1110.
  • FIG. 6 is a diagram illustrating an example of additional information having a medium information amount according to an embodiment of the present disclosure.
  • FIG. 6 shows an action image 1100, a jump trajectory 1110, and jump information 1120.
  • the jump trajectory 1110 includes the current jump trajectory 1111 of the skater (assumed to be Jane Smith) shown in the action image 1100, the jump trajectory 1112 last year of Jane, and the current winner ( Unfortunately not Jane) jump trajectory 1113.
  • the jump information 1120 is displayed in association with each jump trajectory 1110 and includes text indicating the jump flight distance, height, and rotation speed during the jump.
  • more complete information regarding the jump captured in the action image 1100 is expressed by the graphic of the jump trajectory 1110 and the text of the jump information 1120. Yes.
  • FIG. 7 is a diagram illustrating an example of additional information with a large amount of information according to an embodiment of the present disclosure.
  • FIG. 7 shows an action image 1100, a jump locus 1110, jump information 1120, and a graph 1130.
  • jump trajectory 1110 and jump information 1120 are displayed so that they can be selected from Jane's current jump, Jane's last year's jump, and current winner's jump.
  • the user can display any set of jump trajectory 1110 and jump information 1120 by selecting arbitrary jump information 1120 using a pointing device such as a touch panel or a mouse on the output device 300.
  • a graph 1130 corresponds to the displayed set of jump trajectory 1110 and jump information 1120, and displays a change in rotational speed in the jump as a graph.
  • the graph 1130 is folded (wound up) at the start of display, and the non-display portion 1131 is expanded when the user selects the icon of the graph 1130 with a pointing device, for example.
  • the additional information displayed together with the action image 1100 includes more information than the examples shown in FIGS.
  • additional information that allows the user viewing the action image 1100 to select the content is generated.
  • the amount of information to be provided is large, visibility can be improved, for example, by displaying the same type of information according to the user's selection, rather than displaying all information at once. Accordingly, it may be useful to generate selectable additional information even if it is not necessarily the case of additional information having the largest amount of information (for example, in the case of the example shown in FIG. 6).
  • selectable additional information is generated when the user views the action image 1100.
  • the additional information generation unit 209 displays selectable additional information, for example, when it is estimated that the user has a long time for referring to the action image to some extent.
  • FIG. 8 is a flowchart illustrating an example of processing for controlling the amount of additional information according to an embodiment of the present disclosure.
  • the evaluation information acquisition unit 205 and the situation information acquisition unit 207 acquire action evaluation information and user situation information (S101).
  • the additional information generation unit 209 estimates the time for the user to refer to the additional information provided together with the action image from the user situation information.
  • the holding time may be estimated based on, for example, whether the user is viewing the action image live or viewing the recorded action image. For example, when the difference between the time when the actor performed the action included in the action image and the time when the user views the action image is below the threshold, that is, when the user is viewing the action image in a state close to live broadcast
  • the time for the user to refer to the additional information can be estimated to be short. Conversely, if the difference between the time when the actor performs the action and the time when the user views the action image exceeds the threshold, that is, if the user is viewing the recorded action image afterwards, the user will receive additional information. It can be estimated that the holding time for referring to is long.
  • the time for the user to refer to the additional information may be estimated based on the user's schedule, the playback time set for viewing the action image, or the like. Good.
  • information indicating the time for the user to refer to the additional information or the amount of additional information that the user considers appropriate for viewing the action image may be input by the user.
  • the additional information generating unit 209 that has estimated the holding time as described above determines whether or not the estimated holding time is less than the first threshold th1 (S103).
  • the additional information generation unit 209 when the possession time is less than the threshold th1 (YES), the additional information generation unit 209 generates simple additional information with a small amount of information, such as the example illustrated in FIG. 5 (S105).
  • the additional information generation unit 209 may determine not to generate additional information.
  • the additional information generation unit 209 determines whether or not the estimated time remaining exceeds the second threshold th2. Is determined (S107).
  • the additional information generation unit 209 when the holding time exceeds the threshold th2 (YES), the additional information generation unit 209 generates selectable additional information with a larger amount of information, such as the example illustrated in FIG. 7 (S109). .
  • the additional information generation unit 209 In other cases (NO in S107), the additional information generation unit 209 generates additional information with a medium amount of information, such as the example illustrated in FIG. 6 (S111).
  • the additional information in this case is normal additional information, but the normal additional information may not be defined.
  • an appropriate amount of information can be presented according to, for example, the time for the user to refer to the additional information.
  • the user may view the action image live, or may view the recorded action image.
  • the appropriate amount of additional information may vary depending on the user's situation such as a schedule. In the embodiment, it is possible to flexibly cope with such a change and provide an appropriate amount of information to the user.
  • the present embodiment may include a configuration in which the additional information generation unit 209 controls the content of the additional information in accordance with the user situation when generating the additional information based on the user situation information.
  • the additional information generation unit 209 controls the content of the additional information in accordance with the user situation when generating the additional information based on the user situation information.
  • FIG. 9 is a diagram illustrating an example of additional information for a user who views an action image alone in an embodiment of the present disclosure.
  • FIG. 9 shows an action image 1100, a jump trajectory 1110 that is transparently superimposed on the action image 1100, and jump information 1120.
  • the jump trajectory 1110 and the jump information 1120 are assumed to be a jump of a skater (Jane Smith) shown in the action image 1100 and a user who is watching the action image 1100 jumps temporarily. Jumps (labeled “YOU”), assumed jumps of the user's same age (30's), and assumed jumps of the user's youth (YOU IN YOUTH).
  • Such jump trajectory 1110 and jump information 1120 are drawn based on, for example, action evaluation information related to Jane Smith's jump and average basic physical strength data for each user's basic physical strength data, age, and age.
  • FIG. 10 is a diagram illustrating an example of additional information for a user who views a plurality of action images according to an embodiment of the present disclosure.
  • the additional information includes a jump locus 1110 that is transparently superimposed on the action image 1100, jump information 1120, and a ranking 1140 related to the jump.
  • the jump trajectory 1110 and the jump information 1120 are a skater (Jane Smith) jump shown in the action image 1100 and a plurality of users (JOHNNY and DAD) who are watching the action image 1100. It is displayed about the assumed jump when each jumps temporarily.
  • Such jump trajectory 1110 and jump information 1120 are drawn based on, for example, action evaluation information regarding Jane Smith's jump and basic physical strength data of each user.
  • the ranking 1140 displays the assumed flying distance, rotational speed, and overall ranking for the assumed jumps when JOHNNY and DAD each jump.
  • the additional information generation unit 209 controls the content of the additional information according to the number of users viewing the action image. More specifically, when one user who is viewing an action image is detected by the output device 300, the additional information generation unit 209, in addition to the skater (Jane Smith), Additional information including virtual jump information of the user himself, the same age, and young age is generated. On the other hand, when a plurality of users viewing an action image is detected by the output device 300, the additional information generation unit 209 compares virtual jumps of a plurality of users in addition to Jane Smith, Generate additional information to rank. As a result, the user who is watching the action image alone can feel the jump level of Jane Smith by providing a lot of familiar information about himself. In addition, users who are viewing action images in a plurality are provided with virtual jump information about each user, and further provided with information such as ranking that allows them to compete, for example, between each other. You can get a topic at.
  • the additional information such as the example shown in FIGS. 9 and 10 can be additional information for a user who is not deeply familiar with the action (figure skating in the illustrated example), for example.
  • the additional information generation unit 209 is, for example, the example described above with reference to FIG.
  • additional information that enables comparison with other skaters or past records may be provided.
  • users who are familiar with figure skating are interested in comparison with various skaters and past records as competitions.
  • users who are not very proficient are assumed to be interested in realizing a high level of skater based on comparison with themselves and nearby users. Note that such an assumption is merely an example, and additional information can be generated based on various assumptions.
  • FIG. 11 is a diagram illustrating a first example of other additional information according to an embodiment of the present disclosure.
  • FIG. 11 shows an action image 1200 and a composite image 1210 displayed on the action image 1200.
  • the performance of synchronized swimming is captured as an action.
  • the evaluation information generation unit 203 evaluates the performance of synchronized swimming based on the posture of the swimmer estimated from the action image 1200. More specifically, the evaluation information generation unit 203 evaluates the performance according to how much the movements of the respective swimmers are synchronized. For example, the evaluation information generation unit 203 compares the movements of the respective swimmers in time series, and compares the movement start timing, the magnitude of movement, the speed of movement, and the like.
  • the additional information generation unit 209 transparently superimposes the swimmer images extracted from the action image 1200 to generate a composite image 1210. More specifically, the additional information generation unit 209 combines the positions of the images 1211 to 1214 of the respective swimmers based on the postures of the swimmers estimated by the evaluation information generation unit 203 and transparently superimposes them. An image 1210 is generated. Thus, the user who views the action image 1200 can easily grasp how much the movements of the respective swimmers are synchronized.
  • FIG. 12 is a diagram illustrating a second example of other additional information according to an embodiment of the present disclosure.
  • FIG. 12 shows an action image 1300 and a punch icon 1310 displayed on the action image 1300.
  • a punch in boxing is captured as an action.
  • the evaluation information generation unit 203 estimates the strength of the punch based on the detection value of the impact sensor built in the glove.
  • the evaluation information generation unit 203 estimates the contact position between the glove and the boxer body at the time of punching, that is, the position where the punch hits, based on the posture of the boxer estimated from the action image 1300.
  • the evaluation information generation unit 203 determines whether or not the punch is an effective hit based on the position where the punch hits and the strength of the punch.
  • the evaluation information generation unit 203 may evaluate the aggressiveness of the boxer from the number of punches including those that are not effective hits and the tendency of the hit positions.
  • the additional information generation unit 209 displays a punch icon 1310 on the action image 1300 based on the punch hit position estimated by the evaluation information generation unit 203 and the punch strength. More specifically, the additional information generation unit 209 determines the display position of the punch icon 1310 according to the estimated position where the punch hits, and determines the size of the punch icon 1310 based on the estimated punch strength. Determine the color. Further, the additional information generation unit 209 may change the color of the punch icon 1310 depending on which punch of the fighting boxers is punched. As a result, the user who views the action image 1300 can intuitively understand where the punches emitted by the respective boxers hit and how strong they were.
  • FIG. 13 is a diagram illustrating a third example of other additional information according to an embodiment of the present disclosure.
  • FIG. 13 shows an action image 1400 and a ground area 1410 displayed on the action image 1400.
  • a throwing technique in judo is captured as an action.
  • the evaluation information generation unit 203 evaluates the effectiveness of the throwing technique based on the relationship between the player posture estimated from the action image 1400 and the floor surface recognized in the action image 1400. For example, the effectiveness of throwing techniques varies depending on whether the thrown athlete fell from the shoulder to the floor or from the back.
  • the evaluation information generation unit 203 also refers to the athlete's skeleton model to determine whether the contact position between the thrown athlete's body and the floor is the shoulder or the back. Thereby, the effectiveness of the throwing technique and the points given based on it can be determined.
  • the additional information generation unit 209 causes the action image 1400 to display the ground contact area 1410 based on the contact position between the body of the thrown player and the floor estimated by the evaluation information generation unit 203. More specifically, the additional information generation unit 209 displays a graphic of a predetermined color in an area corresponding to the estimated contact position. Further, the additional information generation unit 209 may change the color or the like of the ground contact area 1410 depending on, for example, whether the contact position is the shoulder or the back. Accordingly, the user who views the action image 1400 can intuitively understand why the points are given or not given to the throwing technique.
  • the additional information as described above with reference to FIGS. 11 to 13 and the action evaluation information based on the additional information include information related to action determination or scoring (synchronization of swimmer movements in synchronized swimming). , The hit position and strength of the punch in boxing, and the grounded position of the thrown athlete in judo). Therefore, in the above example, the action evaluation information includes information on determining or scoring the action, and the additional information can be said to indicate the progress or result of the determination or scoring. In the above example, the additional information indicates the progress of scoring, but the result (eg, given points) may be indicated in the same manner. Such information is not limited to users who watch action images as entertainment, for example, but is provided to assist users who actually evaluate actions, such as referees or judges. Also good.
  • the additional information generation unit 209 determines whether or not to generate the additional information as described above based on the action evaluation information. The determination may be made according to the depth of the user's knowledge about the competition captured in the action image. For example, if the user has a deep knowledge of the game, the additional information as in the above example may rather interfere with watching the game. On the other hand, if the user is not familiar with the game, the additional information as in the above example is useful for the user to understand why the points given by the judgment of the referee were given, for example. sell.
  • the additional information generation unit 209 generates simple additional information that does not generate additional information when the knowledge is deep, or displays only the determination result, according to the depth of knowledge about the user's competition, Otherwise, easy-to-understand additional information with graphics as in the above example may be generated.
  • the additional information generation unit 209 uses the additional information as described above with reference to FIGS. 5 to 7 and the additional information as described with reference to FIGS. It may be selectively generated according to the depth of the slag. For example, if the user has a deep knowledge of figure skating, additional information regarding the content of the competition, as described with reference to FIGS. 5 to 7, may be appropriate. On the other hand, if the user is not familiar with figure skating, additional information for comparing a familiar example such as the user himself with the skater performance shown in the action image as described with reference to FIGS. 9 and 10 is appropriate. It can be. As described above, the additional information generation unit 209 may change the content of the additional information according to the depth of the user's knowledge about the action.
  • FIG. 14 is a block diagram illustrating a hardware configuration example of the information processing apparatus according to the embodiment of the present disclosure.
  • the illustrated information processing apparatus 900 can realize, for example, the server apparatus, camera, or output apparatus in the above-described embodiment.
  • the information processing apparatus 900 includes a CPU (Central Processing unit) 901, a ROM (Read Only Memory) 903, and a RAM (Random Access Memory) 905.
  • the information processing apparatus 900 may include a host bus 907, a bridge 909, an external bus 911, an interface 913, an input device 915, an output device 917, a storage device 919, a drive 921, a connection port 923, and a communication device 925.
  • the information processing apparatus 900 may include an imaging device 933 and a sensor 935 as necessary.
  • the information processing apparatus 900 may include a processing circuit such as a DSP (Digital Signal Processor), an ASIC (Application Specific Integrated Circuit), or an FPGA (Field-Programmable Gate Array) instead of or in addition to the CPU 901.
  • DSP Digital Signal Processor
  • ASIC Application Specific Integrated Circuit
  • FPGA Field-Programmable Gate Array
  • the CPU 901 functions as an arithmetic processing device and a control device, and controls all or a part of the operation in the information processing device 900 according to various programs recorded in the ROM 903, the RAM 905, the storage device 919, or the removable recording medium 927.
  • the ROM 903 stores programs and calculation parameters used by the CPU 901.
  • the RAM 905 primarily stores programs used in the execution of the CPU 901, parameters that change as appropriate during the execution, and the like.
  • the CPU 901, the ROM 903, and the RAM 905 are connected to each other by a host bus 907 configured by an internal bus such as a CPU bus. Further, the host bus 907 is connected to an external bus 911 such as a PCI (Peripheral Component Interconnect / Interface) bus via a bridge 909.
  • PCI Peripheral Component Interconnect / Interface
  • the input device 915 is a device operated by the user, such as a mouse, a keyboard, a touch panel, a button, a switch, and a lever.
  • the input device 915 may be, for example, a remote control device that uses infrared rays or other radio waves, or may be an external connection device 929 such as a mobile phone that supports the operation of the information processing device 900.
  • the input device 915 includes an input control circuit that generates an input signal based on information input by the user and outputs the input signal to the CPU 901. The user operates the input device 915 to input various data and instruct processing operations to the information processing device 900.
  • the output device 917 is configured by a device capable of notifying the acquired information to the user using a sense such as vision, hearing, or touch.
  • the output device 917 can be, for example, a display device such as an LCD (Liquid Crystal Display) or an organic EL (Electro-Luminescence) display, an audio output device such as a speaker or headphones, or a vibrator.
  • the output device 917 outputs the result obtained by the processing of the information processing device 900 as video such as text or image, sound such as sound or sound, or vibration.
  • the storage device 919 is a data storage device configured as an example of a storage unit of the information processing device 900.
  • the storage device 919 includes, for example, a magnetic storage device such as an HDD (Hard Disk Drive), a semiconductor storage device, an optical storage device, or a magneto-optical storage device.
  • the storage device 919 stores, for example, programs executed by the CPU 901 and various data, and various data acquired from the outside.
  • the drive 921 is a reader / writer for a removable recording medium 927 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory, and is built in or externally attached to the information processing apparatus 900.
  • the drive 921 reads information recorded on the attached removable recording medium 927 and outputs the information to the RAM 905.
  • the drive 921 writes a record in the attached removable recording medium 927.
  • the connection port 923 is a port for connecting a device to the information processing apparatus 900.
  • the connection port 923 can be, for example, a USB (Universal Serial Bus) port, an IEEE 1394 port, a SCSI (Small Computer System Interface) port, or the like.
  • the connection port 923 may be an RS-232C port, an optical audio terminal, an HDMI (registered trademark) (High-Definition Multimedia Interface) port, or the like.
  • the communication device 925 is a communication interface configured with, for example, a communication device for connecting to the communication network 931.
  • the communication device 925 can be, for example, a communication card for LAN (Local Area Network), Bluetooth (registered trademark), Wi-Fi, or WUSB (Wireless USB).
  • the communication device 925 may be a router for optical communication, a router for ADSL (Asymmetric Digital Subscriber Line), or a modem for various communication.
  • the communication device 925 transmits and receives signals and the like using a predetermined protocol such as TCP / IP with the Internet and other communication devices, for example.
  • the communication network 931 connected to the communication device 925 is a network connected by wire or wireless, and may include, for example, the Internet, a home LAN, infrared communication, radio wave communication, satellite communication, or the like.
  • the imaging device 933 uses various members such as an imaging element such as a CMOS (Complementary Metal Oxide Semiconductor) or a CCD (Charge Coupled Device), and a lens for controlling the formation of a subject image on the imaging element. It is an apparatus that images a real space and generates a captured image.
  • the imaging device 933 may capture a still image or may capture a moving image.
  • the sensor 935 is various sensors such as an acceleration sensor, an angular velocity sensor, a geomagnetic sensor, an illuminance sensor, a temperature sensor, an atmospheric pressure sensor, or a sound sensor (microphone).
  • the sensor 935 acquires information about the state of the information processing apparatus 900 itself, such as the posture of the information processing apparatus 900, and information about the surrounding environment of the information processing apparatus 900, such as brightness and noise around the information processing apparatus 900, for example. To do.
  • the sensor 935 may include a GPS receiver that receives a GPS (Global Positioning System) signal and measures the latitude, longitude, and altitude of the device.
  • GPS Global Positioning System
  • Each component described above may be configured using a general-purpose member, or may be configured by hardware specialized for the function of each component. Such a configuration can be appropriately changed according to the technical level at the time of implementation.
  • an information processing apparatus for example, an information processing apparatus, a system, an information processing method executed by the information processing apparatus or system, a program for causing the information processing apparatus to function, and a program are recorded. It may include tangible media that is not temporary.
  • An evaluation information acquisition unit that acquires action evaluation information obtained by evaluating the action based on an action image including an action of an actor;
  • a situation information acquisition unit that acquires user situation information indicating a situation of a user who views the action image;
  • An information processing apparatus comprising: an additional information generation unit that generates additional information that is presented to the user together with the action image and expresses at least part of the action evaluation information according to the user situation information.
  • the additional information includes visual information.
  • the additional information includes text or graphics that are superimposed on the action image or displayed in the vicinity of the action image.
  • the action evaluation information includes information indicating coordinates of a body part of the actor in the action image, The information processing apparatus according to (3), wherein the additional information includes a graphic representing a locus of the portion.
  • the action evaluation information includes information for determining or scoring the action, The information processing apparatus according to (3) or (4), wherein the additional information includes text or graphics indicating a progress or result of the determination or scoring.
  • the user status information includes information for estimating the length of time the user has for referring to the additional information, 6.
  • the information processing apparatus according to (6), wherein the additional information generation unit increases the information amount of the additional information as the holding time increases.
  • the information processing apparatus according to (6) or (7), wherein the additional information generation unit generates the additional information from which the user can select contents when the holding time is long.
  • the user status information includes information indicating a relationship between a first time when the actor performs the action and a second time when the user views the action image, The additional information generation unit according to any one of (6) to (8), wherein the time is determined to be short when a difference between the first time and the second time is less than a threshold value.
  • Information processing device (10)
  • the user situation information includes information for estimating a depth of knowledge about the action of the user, The information processing apparatus according to any one of (1) to (9), wherein the additional information generation unit generates the additional information according to a depth of the sculpting.
  • the user status information includes information indicating the number of users, The information processing apparatus according to any one of (1) to (10), wherein the additional information generation unit generates the additional information according to the number of users.
  • the action evaluation information includes information obtained by evaluating a contact position between the actor and another object estimated based on the action image.

Landscapes

  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Physical Education & Sports Medicine (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

Provided is an information processing device provided with: an evaluation information acquisition unit (205) for acquiring action evaluation information that evaluates an action of an actor on the basis of an action image including the action; a situation information acquisition unit (207) for acquiring user situation information indicating the situation of a user viewing the action image; and an additional information generation unit (209) for generating, according to the user situation information, additional information that is presented to the user with the action image and expresses at least part of the action evaluation information.

Description

情報処理装置、情報処理方法およびプログラムInformation processing apparatus, information processing method, and program
 本開示は、情報処理装置、情報処理方法およびプログラムに関する。 This disclosure relates to an information processing apparatus, an information processing method, and a program.
 近年、例えば高性能化および低廉化した加速度センサなどのセンサを利用して、ユーザ自身がプレーするスポーツをより多様に楽しむためのサービスが提案されている。例えば、特許文献1には、ユーザによるスポーツのプレーの状態をより的確に把握するために、センサ情報に基づいてスポーツにおけるプレーイベントの発生を検出し、プレーイベントの発生に応じてプレー画像の撮影を制御する技術が記載されている。 In recent years, services for enjoying sports played by users themselves in a variety of ways have been proposed by using sensors such as high-performance and low-cost acceleration sensors. For example, in Patent Document 1, in order to more accurately grasp the state of sports play by a user, the occurrence of a play event in sports is detected based on sensor information, and a play image is captured according to the occurrence of the play event. Techniques for controlling the are described.
特開2013-188426号公報JP 2013-188426 A
 しかしながら、スポーツの楽しみ方は、必ずしもユーザ自身がプレーすることには限られない。テレビなどを介して、プロフェッショナルなスポーツや、スポーツの競技会を観戦することも、ごく一般的なスポーツの楽しみ方である。しかしながら、特許文献1に記載されたような技術は、スポーツのプレーの状態をより的確に把握することを意図としており、必ずしも、専らプレーを観戦するユーザに対してより良い体験を提供することを目的としたものではない。 However, the way of enjoying sports is not necessarily limited to being played by the users themselves. Watching professional sports and sports competitions via television etc. is also a very common way to enjoy sports. However, the technique as described in Patent Document 1 is intended to more accurately grasp the state of sports play, and does not necessarily provide a better experience for users who exclusively watch play. It is not intended.
 そこで、本開示では、アクションを含む画像を視聴するユーザに対して、画像に基づくアクションの評価を反映したより適切な情報を提供することが可能な、新規かつ改良された情報処理装置、情報処理方法、およびプログラムを提案する。 Therefore, in the present disclosure, a new and improved information processing apparatus and information processing capable of providing more appropriate information reflecting an evaluation of an action based on an image to a user who views the image including the action Propose methods and programs.
 本開示によれば、アクターのアクションを含むアクション画像に基づいて上記アクションを評価したアクション評価情報を取得する評価情報取得部と、上記アクション画像を視聴するユーザの状況を示すユーザ状況情報を取得する状況情報取得部と、上記アクション画像とともに上記ユーザに提示され、上記アクション評価情報の少なくとも一部を表現する付加情報を、上記ユーザ状況情報に応じて生成する付加情報生成部とを備える情報処理装置が提供される。 According to the present disclosure, an evaluation information acquisition unit that acquires action evaluation information obtained by evaluating the action based on an action image including an action of an actor, and user situation information indicating a situation of a user who views the action image are acquired. An information processing apparatus comprising: a situation information acquisition unit; and an additional information generation unit that generates additional information that is presented to the user together with the action image and expresses at least part of the action evaluation information according to the user situation information Is provided.
 また、本開示によれば、アクターのアクションを含むアクション画像に基づいて上記アクションを評価したアクション評価情報を取得することと、上記アクション画像を視聴するユーザの状況を示すユーザ状況情報を取得することと、上記アクション画像とともに上記ユーザに提示され、上記アクション評価情報の少なくとも一部を表現する付加情報を、上記ユーザ状況情報に応じて生成することとを含む情報処理方法が提供される。 Further, according to the present disclosure, acquiring action evaluation information obtained by evaluating the action based on an action image including an actor's action, and acquiring user situation information indicating a situation of a user viewing the action image And generating additional information that is presented to the user together with the action image and expresses at least a part of the action evaluation information according to the user situation information.
 また、本開示によれば、アクターのアクションを含むアクション画像に基づいて上記アクションを評価したアクション評価情報を取得する機能と、上記アクション画像を視聴するユーザの状況を示すユーザ状況情報を取得する機能と、上記アクション画像とともに上記ユーザに提示され、上記アクション評価情報の少なくとも一部を表現する付加情報を、上記ユーザ状況情報に応じて生成する機能とをコンピュータに実現させるためのプログラムが提供される。 Moreover, according to this indication, the function which acquires the action evaluation information which evaluated the said action based on the action image containing an action of an actor, and the function which acquires the user condition information which shows the condition of the user who views the said action image And a program for causing the computer to realize the function of generating additional information that is presented to the user together with the action image and expresses at least part of the action evaluation information according to the user situation information. .
 画像に基づいてアクションを評価した情報をアクションの画像とともに提供することは、提供される情報と画像の内容とがリンクしているために、ユーザがアクションを理解することを助ける。しかし、どのような情報を提供するかが適切かはユーザの状況によって異なりうるため、ユーザの状況にさらに基づいてアクションの評価に基づく付加情報を生成することによって、より適切な情報をユーザに提供することができる。 Providing information that evaluates an action based on an image together with the action image helps the user understand the action because the provided information is linked to the content of the image. However, since what kind of information is appropriate may vary depending on the user's situation, more appropriate information is provided to the user by generating additional information based on action evaluation based on the user's situation. can do.
 以上説明したように本開示によれば、アクションを含む画像を視聴するユーザに対して、画像に基づくアクションの評価を反映したより適切な情報を提供することができる。 As described above, according to the present disclosure, it is possible to provide more appropriate information reflecting the evaluation of the action based on the image to the user who views the image including the action.
 なお、上記の効果は必ずしも限定的なものではなく、上記の効果とともに、または上記の効果に代えて、本明細書に示されたいずれかの効果、または本明細書から把握され得る他の効果が奏されてもよい。 Note that the above effects are not necessarily limited, and any of the effects shown in the present specification, or other effects that can be grasped from the present specification, together with or in place of the above effects. May be played.
本開示の一実施形態に係るシステムの構成例を示す図である。It is a figure showing an example of composition of a system concerning one embodiment of this indication. 本開示の一実施形態に係る情報処理装置の機能構成例を示す図である。It is a figure showing an example of functional composition of an information processor concerning one embodiment of this indication. 本開示の一実施形態に係る付加情報の第1の例を示す図である。It is a figure showing the 1st example of additional information concerning one embodiment of this indication. 本開示の一実施形態に係る付加情報の第2の例を示す図である。It is a figure showing the 2nd example of additional information concerning one embodiment of this indication. 本開示の一実施形態における情報量が少ない付加情報の例を示す図である。It is a figure which shows the example of additional information with little information content in one Embodiment of this indication. 本開示の一実施形態における情報量が中程度の付加情報の例を示す図である。It is a figure which shows the example of the additional information with medium amount of information in one Embodiment of this indication. 本開示の一実施形態における情報量が多い付加情報の例を示す図である。It is a figure which shows the example of additional information with much information amount in one Embodiment of this indication. 本開示の一実施形態における付加情報の情報量の制御のための処理の例を示すフローチャートである。14 is a flowchart illustrating an example of processing for controlling the amount of additional information according to an embodiment of the present disclosure. 本開示の一実施形態において一人でアクション画像を視聴するユーザ向けの付加情報の例を示す図である。FIG. 5 is a diagram illustrating an example of additional information for a user who views an action image alone in an embodiment of the present disclosure. 本開示の一実施形態において複数でアクション画像を視聴するユーザ向けの付加情報の例を示す図である。FIG. 4 is a diagram illustrating an example of additional information for a user who views a plurality of action images in an embodiment of the present disclosure. 本開示の一実施形態における他の付加情報の第1の例を示す図である。It is a figure showing the 1st example of other additional information in one embodiment of this indication. 本開示の一実施形態における他の付加情報の第2の例を示す図である。It is a figure showing the 2nd example of other additional information in one embodiment of this indication. 本開示の一実施形態における他の付加情報の第3の例を示す図である。It is a figure showing the 3rd example of other additional information in one embodiment of this indication. 本開示の実施形態に係る情報処理装置のハードウェア構成例を示すブロック図である。FIG. 3 is a block diagram illustrating a hardware configuration example of an information processing apparatus according to an embodiment of the present disclosure.
 以下に添付図面を参照しながら、本開示の好適な実施の形態について詳細に説明する。なお、本明細書および図面において、実質的に同一の機能構成を有する構成要素については、同一の符号を付することにより重複説明を省略する。 Hereinafter, preferred embodiments of the present disclosure will be described in detail with reference to the accompanying drawings. In the present specification and drawings, components having substantially the same functional configuration are denoted by the same reference numerals, and redundant description is omitted.
 なお、説明は以下の順序で行うものとする。
 1.システム構成
 2.情報処理装置の機能構成
 3.視覚的な付加情報の例
 4.付加情報の情報量の制御の例
 5.付加情報の内容の制御の例
 6.他の付加情報の例
 7.ハードウェア構成
 8.補足
The description will be made in the following order.
1. System configuration 2. Functional configuration of information processing apparatus Example of visual additional information 4. 4. Example of control of information amount of additional information 5. Example of control of content of additional information 6. Examples of other additional information Hardware configuration Supplement
 (1.システム構成)
 図1は、本開示の一実施形態に係るシステムの構成例を示す図である。図1を参照すると、システム10は、カメラ100と、サーバ200と、出力装置300(テレビ300aおよびレコーダ300b)とを含む。
(1. System configuration)
FIG. 1 is a diagram illustrating a configuration example of a system according to an embodiment of the present disclosure. Referring to FIG. 1, a system 10 includes a camera 100, a server 200, and an output device 300 (a television 300a and a recorder 300b).
 カメラ100は、アクターのアクションを含むアクション画像を撮影する。ここで、アクターは、何らかのアクションを実行する人または物である。アクションは、例えばスポーツのプレー、ダンスなどのパフォーマンス、職業的な技能など、それが評価の対象になりうるあらゆる種類のアクションを含みうる。例えば、アクションがスポーツのプレーである場合には、アクターはプレーヤとも呼ばれる。その他、パフォーマー、ダンサー、試技者など、アクターは例えばアクションの種類に応じて様々な呼び方をされる。カメラ100は、例えば、静止画像または動画像を取得することが可能なデジタルビデオカメラでありうる。カメラ100は、少なくとも1つの撮像素子を有し、アクターを被写体として含む画像を電子データとして記録する。さらに、カメラ100は、画像の電子データをサーバ200および出力装置300に適したフォーマットに変換する処理などを実施する画像処理回路を有してもよい。また、カメラ100は、後述する情報処理装置の機能の一部または全部を実現する情報処理回路を有してもよい。情報処理回路は、例えば後述する情報処理装置のハードウェア構成を応用して構成されうる。 The camera 100 captures an action image including an actor's action. Here, an actor is a person or thing that performs some action. Actions can include all kinds of actions that can be evaluated, such as sports play, performances such as dance, and professional skills. For example, if the action is sport play, the actor is also called a player. In addition, actors such as performers, dancers, and testers are called in various ways depending on the type of action. The camera 100 can be, for example, a digital video camera that can acquire a still image or a moving image. The camera 100 has at least one image sensor and records an image including an actor as a subject as electronic data. Furthermore, the camera 100 may include an image processing circuit that performs processing for converting electronic image data into a format suitable for the server 200 and the output device 300. In addition, the camera 100 may include an information processing circuit that realizes part or all of the functions of the information processing apparatus described later. The information processing circuit can be configured by applying a hardware configuration of an information processing apparatus to be described later, for example.
 サーバ200は、カメラ100からアクション画像を受信する(S11)とともに、アクション画像に基づいてアクションを評価したアクション評価情報を生成または受信する。さらに、ある例において、サーバ200は、アクション画像を視聴するユーザの状況を示すユーザ状況情報を出力装置300から受信し(S13a)、ユーザ状況情報に応じて生成された付加情報を、アクション画像とともに出力装置300に送信する(S12)。ここで、付加情報は、アクション評価情報の少なくとも一部を含む。あるいは、サーバ200は、アクション画像をアクション評価情報とともに、そのまま出力装置300に送信する(S12)。この場合、アクション評価情報を用いた付加情報の生成は、出力装置300で実施される(S13b)。サーバ200は、ネットワーク上の1または複数のサーバ装置によって実現されうる。個々のサーバ装置は、例えば後述する情報処理装置のハードウェア構成を応用して構成されうる。 The server 200 receives an action image from the camera 100 (S11), and generates or receives action evaluation information that evaluates an action based on the action image. Further, in an example, the server 200 receives user status information indicating the status of the user viewing the action image from the output device 300 (S13a), and adds the additional information generated according to the user status information together with the action image. It transmits to the output device 300 (S12). Here, the additional information includes at least a part of the action evaluation information. Alternatively, the server 200 transmits the action image together with the action evaluation information to the output device 300 as it is (S12). In this case, the generation of additional information using the action evaluation information is performed by the output device 300 (S13b). The server 200 can be realized by one or a plurality of server devices on a network. Each server device can be configured by applying a hardware configuration of an information processing device to be described later, for example.
 出力装置300は、サーバ200からアクション画像を受信し(S12)、アクション画像を付加情報とともに出力してユーザに提示する。上記の通り、付加情報は、出力装置300が送信したユーザ状況情報(S13a)に基づいてサーバ200で生成され、アクション画像とともにサーバ200から出力装置300に送信されてもよい(S12)。あるいは、出力装置300は、サーバ200からアクション画像とともにアクション評価情報を受信し(S12)、ユーザ状況情報に基づいてアクション評価情報から付加情報を生成した上で(S13b)、アクション画像および付加情報を出力してもよい。図示された例において、出力装置300は、テレビ300aおよびレコーダ300bとして例示されている。例えば、サーバ200から放送波によって送信されたアクション画像は、レコーダ300bによって受信、記録、および保存され、後にテレビ300aを介して再生および出力される。あるいは、アクション画像は、レコーダ300bを介さずに直接的にテレビ300aによって受信および出力されてもよい。このような場合に、テレビ300aまたはレコーダ300bは、付加情報の生成などに関する情報処理を実施する情報処理回路を有してもよい。情報処理回路は、例えば後述する情報処理装置のハードウェア構成を応用して構成されうる。 The output device 300 receives an action image from the server 200 (S12), outputs the action image together with additional information, and presents it to the user. As described above, the additional information may be generated by the server 200 based on the user situation information (S13a) transmitted by the output device 300, and transmitted from the server 200 to the output device 300 together with the action image (S12). Alternatively, the output device 300 receives the action evaluation information together with the action image from the server 200 (S12), generates additional information from the action evaluation information based on the user situation information (S13b), and then displays the action image and additional information. It may be output. In the illustrated example, the output device 300 is illustrated as a television 300a and a recorder 300b. For example, an action image transmitted from the server 200 by a broadcast wave is received, recorded, and saved by the recorder 300b, and later played and output via the television 300a. Alternatively, the action image may be received and output directly by the television 300a without using the recorder 300b. In such a case, the television 300a or the recorder 300b may include an information processing circuit that performs information processing related to generation of additional information and the like. The information processing circuit can be configured by applying a hardware configuration of an information processing apparatus to be described later, for example.
 なお、出力装置300の例は、図示された例におけるテレビ300aおよびレコーダ300bの組み合わせには限られない。例えば、テレビ300aが単独で出力装置300として用いられてもよい。また、その他にも、受信した画像を出力する機能を有するあらゆる装置が、出力装置300として用いられうる。例えば、スマートフォンやタブレット、パーソナルコンピュータなどが、出力装置300として用いられてもよい。 Note that the example of the output device 300 is not limited to the combination of the television 300a and the recorder 300b in the illustrated example. For example, the television 300a may be used alone as the output device 300. In addition, any device having a function of outputting a received image can be used as the output device 300. For example, a smartphone, a tablet, a personal computer, or the like may be used as the output device 300.
 (2.情報処理装置の機能構成)
 図2は、本開示の一実施形態に係る情報処理装置の機能構成例を示す図である。図2を参照すると、情報処理装置は、画像取得部201、評価情報生成部203、評価情報取得部205、状況情報取得部207、付加情報生成部209、および画面生成部211を含む。情報処理装置の機能構成は、上述のように、システム10においてカメラ100、サーバ200、または出力装置300によって実現される。これらの機能構成は、単一の装置に集積して実現されてもよいし、複数の装置に分散して実現されてもよい。以下で説明する機能構成は、例えば、1または複数の装置のそれぞれのCPUなどのプロセッサがメモリまたはストレージに格納されたプログラムに従って動作することによって実現される。
(2. Functional configuration of information processing apparatus)
FIG. 2 is a diagram illustrating a functional configuration example of the information processing apparatus according to an embodiment of the present disclosure. Referring to FIG. 2, the information processing apparatus includes an image acquisition unit 201, an evaluation information generation unit 203, an evaluation information acquisition unit 205, a situation information acquisition unit 207, an additional information generation unit 209, and a screen generation unit 211. The functional configuration of the information processing apparatus is realized by the camera 100, the server 200, or the output device 300 in the system 10 as described above. These functional configurations may be realized by being integrated in a single device, or may be realized by being distributed to a plurality of devices. The functional configuration described below is realized by, for example, a processor such as a CPU of one or a plurality of devices operating according to a program stored in a memory or storage.
 画像取得部201は、カメラ100によって撮影されたアクション画像を取得する。上述の通り、アクション画像は、アクターによるアクションを含む画像である。画像取得部201は、例えば、ネットワークなどで接続されたカメラ100から有線または無線の通信で画像データを受信する通信装置によって実現されてもよい。また、画像取得部201は、そのような通信装置からデータを取得するドライバプログラムなどのソフトウェア的なインターフェースであってもよい。なお、例えば後述する画面生成部211がカメラ100において実現されるような場合、画像取得部201は、バスなどの内部的な通信経路を介してアクション画像を取得してもよい。 The image acquisition unit 201 acquires an action image taken by the camera 100. As described above, the action image is an image including an action by an actor. The image acquisition unit 201 may be realized by, for example, a communication device that receives image data from the camera 100 connected via a network or the like by wired or wireless communication. The image acquisition unit 201 may be a software interface such as a driver program that acquires data from such a communication device. For example, when a screen generation unit 211 described later is realized in the camera 100, the image acquisition unit 201 may acquire an action image via an internal communication path such as a bus.
 評価情報生成部203は、アクション画像に基づいてアクションを評価したアクション評価情報を生成する。図示された例において、評価情報生成部203は、アクション画像をカメラ100から取得する。上記の画像取得部201と同様に、アクション画像は、装置間の通信によって受信されてもよいし、内部的な通信経路を介して取得されてもよい。例えば、評価情報生成部203は、アクション画像について画像解析を実施し、アクターの変位や回転、姿勢などを推定する。さらに、評価情報生成部203は、推定された情報を評価する。このとき、評価情報生成部203は、アクションの判定または採点の基準を適用することによって、アクションを評価してもよい。判定または採点の基準は、アクションの種類(例えばスポーツやパフォーマンス、職業的な技能など)ごとに定められる。そのような基準は、既に審判または審査員などによって利用されていてもよいし、システム10におけるアクションの評価のために新たに導入されてもよい。後述するように、アクション評価情報、またはこれに基づく付加情報が、審判または審査員に提供されてもよい。必要に応じて、評価情報生成部203は、アクション画像以外の情報に基づいてアクション評価情報を生成してもよい。例えば、評価情報生成部203は、アクション画像に加えて、アクターに装着された慣性センサなどのセンサから得られる検出値などに基づいてアクション評価情報を生成してもよい。 The evaluation information generation unit 203 generates action evaluation information that evaluates an action based on the action image. In the illustrated example, the evaluation information generation unit 203 acquires an action image from the camera 100. Similar to the image acquisition unit 201 described above, the action image may be received by communication between apparatuses, or may be acquired via an internal communication path. For example, the evaluation information generation unit 203 performs image analysis on the action image, and estimates the displacement, rotation, posture, and the like of the actor. Furthermore, the evaluation information generation unit 203 evaluates the estimated information. At this time, the evaluation information generation unit 203 may evaluate the action by applying an action determination or scoring standard. Judgment or scoring standards are determined for each type of action (for example, sports, performance, professional skills, etc.). Such criteria may already be used by referees or judges, or may be newly introduced for the evaluation of actions in the system 10. As will be described later, action evaluation information or additional information based thereon may be provided to a referee or a judge. If necessary, the evaluation information generation unit 203 may generate action evaluation information based on information other than the action image. For example, the evaluation information generation unit 203 may generate action evaluation information based on a detection value obtained from a sensor such as an inertial sensor attached to the actor in addition to the action image.
 評価情報取得部205は、評価情報生成部203によって生成されたアクション評価情報を取得する。例えば、システム10において、サーバ200に含まれる評価情報生成部203と付加情報生成部209とが互いに異なるサーバ装置によって実現される場合、評価情報生成部203がカメラ100において実現される場合、またはシステム10には含まれない外部のサーバでアクション評価情報が生成されるような場合、評価情報取得部205は、装置間の通信によってアクション評価情報を受信する通信装置によって実現されてもよいし、そのような通信装置からデータを取得するソフトウェア的なインターフェースであってもよい。あるいは、評価情報生成部203と付加情報生成部209とが同じサーバ装置(または、カメラ100もしくは出力装置300)によって実現される場合、評価情報取得部205は、バスなどの内部的な通信経路を介してアクション評価情報を取得するソフトウェア的なインターフェースとして実現されうる。 The evaluation information acquisition unit 205 acquires the action evaluation information generated by the evaluation information generation unit 203. For example, in the system 10, when the evaluation information generation unit 203 and the additional information generation unit 209 included in the server 200 are realized by different server devices, the evaluation information generation unit 203 is realized by the camera 100, or the system When the action evaluation information is generated by an external server that is not included in 10, the evaluation information acquisition unit 205 may be realized by a communication device that receives the action evaluation information through communication between the devices. It may be a software interface for acquiring data from such a communication device. Alternatively, when the evaluation information generation unit 203 and the additional information generation unit 209 are realized by the same server device (or the camera 100 or the output device 300), the evaluation information acquisition unit 205 uses an internal communication path such as a bus. It can be realized as a software interface for acquiring action evaluation information through the interface.
 状況情報取得部207は、出力装置300を介してアクション画像を視聴するユーザの状況を示すユーザ状況情報を取得する。ユーザ状況情報は、例えば、ユーザがアクション画像とともに提示される付加情報を参照するための持ち時間、アクション画像に含まれるアクションについてのユーザの造詣の深さ(familiarity)、またはユーザの数などを示す情報を含みうる。上記の画像取得部201と同様に、状況情報取得部207も、付加情報生成部209が出力装置300とは異なる装置において実現される場合には、装置間の通信によってユーザ状況情報を受信する通信装置、またはそのような通信装置からデータを取得するソフトウェア的なインターフェースによって実現される。また、状況情報取得部207は、付加情報生成部209が出力装置300で実現されるような場合には、バスなどの内部的な通信経路を介してユーザ状況情報を取得するソフトウェア的なインターフェースとして実現されうる。 The status information acquisition unit 207 acquires user status information indicating the status of the user who views the action image via the output device 300. The user status information indicates, for example, the time for the user to refer to the additional information presented together with the action image, the user's familiarity regarding the action included in the action image, or the number of users. Can contain information. Similar to the image acquisition unit 201 described above, the situation information acquisition unit 207 also receives user situation information through communication between devices when the additional information generation unit 209 is realized in a device different from the output device 300. It is realized by a device or a software interface for acquiring data from such a communication device. In addition, when the additional information generation unit 209 is realized by the output device 300, the status information acquisition unit 207 is a software interface that acquires user status information via an internal communication path such as a bus. Can be realized.
 ここで、ユーザ状況情報のいくつかの具体的な例について説明する。例えば、アクション画像を視聴するユーザの数は、出力装置300においてディスプレイに取り付けられたカメラが取得した、ディスプレイの前のユーザを含む画像を解析することによって検出される。また、例えば、アクション画像を視聴するユーザは、出力装置300においてディスプレイに取り付けられたカメラが取得した画像からユーザの顔を認識することによって、または、出力装置300がユーザのログインを受け付けている場合にはログインしているユーザの情報によって特定される。特定された個々のユーザについて、サービスに登録されているプロフィールやスケジュール、コンテンツの利用履歴などを参照することによって、アクションについてのユーザの造詣の深さや、アクション画像を視聴するための持ち時間などを推定することができる。 Here, some specific examples of user status information will be described. For example, the number of users who view an action image is detected by analyzing an image including a user in front of the display acquired by a camera attached to the display in the output device 300. Further, for example, a user who views an action image recognizes the user's face from an image acquired by a camera attached to a display in the output device 300 or when the output device 300 accepts a user login. Is specified by information of the logged-in user. For each identified user, by referring to the profile and schedule registered in the service, content usage history, etc., the depth of the user's knowledge about the action, the time to watch the action image, etc. Can be estimated.
 付加情報生成部209は、状況情報取得部207が取得したユーザ状況情報に基づいて、評価情報取得部205が取得したアクション評価情報の少なくとも一部を表現する付加情報を生成する。上述のように、付加情報は、出力装置300においてアクション画像とともにユーザに提示される。なお、付加情報は、アクション評価情報を表現するものであるため、必ずしもアクション評価情報に含まれる情報をそのまま含まなくてもよい。付加情報生成部209は、アクション評価情報に何らかの加工を加えることによって付加情報を生成してもよい。本実施形態において、付加情報は、アクション画像とともに画面に表示される視覚的な情報を含む。より具体的には、例えば、付加情報は、テキストまたはグラフィックを含む。テキストまたはグラフィックは、アクション画像に重畳されてもよいし、アクション画像の近傍に表示されてもよい。なお、付加情報の具体的な例については後述する。付加情報生成部209は、アクション評価情報の少なくとも一部から付加情報を生成するにあたり、ユーザ状況情報に基づいて、付加情報によって表現される情報量や情報の種類を変更する。 The additional information generation unit 209 generates additional information that represents at least a part of the action evaluation information acquired by the evaluation information acquisition unit 205 based on the user situation information acquired by the situation information acquisition unit 207. As described above, the additional information is presented to the user together with the action image in the output device 300. Since the additional information expresses action evaluation information, it does not necessarily include the information included in the action evaluation information as it is. The additional information generation unit 209 may generate additional information by adding some processing to the action evaluation information. In the present embodiment, the additional information includes visual information displayed on the screen together with the action image. More specifically, for example, the additional information includes text or graphics. The text or graphic may be superimposed on the action image or displayed in the vicinity of the action image. A specific example of additional information will be described later. When generating additional information from at least a part of the action evaluation information, the additional information generation unit 209 changes the amount of information and the type of information expressed by the additional information based on the user situation information.
 なお、上記の例では、評価情報生成部203が生成したアクション評価情報から、付加情報生成部209が付加情報に含まれる情報を選択しているが、本開示の実施形態がこのような例には限られない。別の例では、付加情報生成部209がユーザ状況情報に応じて生成することを決定した付加情報の情報量または情報の種類に応じて、評価情報生成部203がアクション評価情報を生成してもよい。つまり、本開示の実施形態において、付加情報は、ユーザの状況に関わらず生成されるアクション評価情報の一部または全部を抽出することによって生成されてもよいし、ユーザに状況に応じて生成されたアクション評価情報に基づいて生成されてもよい。 In the above example, the additional information generation unit 209 selects information included in the additional information from the action evaluation information generated by the evaluation information generation unit 203. However, the embodiment of the present disclosure is such an example. Is not limited. In another example, even if the evaluation information generation unit 203 generates action evaluation information according to the amount of information or the type of information that the additional information generation unit 209 has decided to generate according to the user situation information. Good. That is, in the embodiment of the present disclosure, the additional information may be generated by extracting part or all of the action evaluation information generated regardless of the user's situation, or generated according to the situation by the user. It may be generated based on the action evaluation information.
 画面生成部211は、画像取得部201によって取得されたアクション画像と、付加情報生成部209によって生成された付加情報とを含む画面を生成する。上記の通り、付加情報生成部209は、アクション評価情報に基づいて生成されるテキストまたはグラフィックを含む。画面生成部211は、これらのテキストまたはグラフィックを、アクション画像に重畳されたり、アクション画像の近傍に表示されたりするように配置する。例えば、画面生成部211は、アクション画像および付加情報を含む画面を描画し、該描画された画面の画像データを出力装置300による表示ために提供してもよい。あるいは、画面生成部211は、画面においてアクション画像を描画するための画像データと、付加情報を描画するための画像データとをそれぞれ生成し、出力装置300がこれらの画像データに基づいてアクション画像および付加情報を画面に表示させてもよい。画面生成部211は、例えばサーバ200において実現されてもよいし、出力装置300において実現されてもよい。 The screen generation unit 211 generates a screen including the action image acquired by the image acquisition unit 201 and the additional information generated by the additional information generation unit 209. As described above, the additional information generation unit 209 includes text or graphics generated based on the action evaluation information. The screen generation unit 211 arranges these texts or graphics so as to be superimposed on the action image or displayed in the vicinity of the action image. For example, the screen generation unit 211 may draw a screen including an action image and additional information, and provide the image data of the drawn screen for display by the output device 300. Alternatively, the screen generation unit 211 generates image data for drawing an action image on the screen and image data for drawing additional information, respectively, and the output device 300 generates an action image and an image based on these image data. Additional information may be displayed on the screen. The screen generation unit 211 may be realized, for example, in the server 200 or in the output device 300.
 なお、図2に示された情報処理装置の機能構成は、情報処理装置が主に付加情報生成部209を実現するプロセッサを備える例に対応して説明されている。この場合、情報処理装置は、付加情報生成部209に加えて、アクション評価情報を取得するインターフェースである評価情報取得部205と、ユーザ状況情報を取得するインターフェースである状況情報取得部207とを備える。他の例において、情報処理装置のプロセッサは、さらに、評価情報生成部203を実現してもよい。この場合、評価情報取得部205は、内部的なプログラム上のインターフェースでありうる。また、情報処理装置は、画像取得部201を含んでもよく、情報処理装置のプロセッサは画面生成部211を実現してもよい。さらに別の例において、情報処理装置は、カメラ100または出力装置300に含まれ、図2に示された情報処理装置の機能構成は、カメラ100または出力装置300において内部的に実現されてもよい。 Note that the functional configuration of the information processing apparatus illustrated in FIG. 2 is described corresponding to an example in which the information processing apparatus includes a processor that mainly implements the additional information generation unit 209. In this case, the information processing apparatus includes, in addition to the additional information generation unit 209, an evaluation information acquisition unit 205 that is an interface for acquiring action evaluation information, and a situation information acquisition unit 207 that is an interface for acquiring user situation information. . In another example, the processor of the information processing apparatus may further realize the evaluation information generation unit 203. In this case, the evaluation information acquisition unit 205 can be an internal program interface. Further, the information processing apparatus may include an image acquisition unit 201, and the processor of the information processing apparatus may realize the screen generation unit 211. In still another example, the information processing apparatus is included in the camera 100 or the output apparatus 300, and the functional configuration of the information processing apparatus illustrated in FIG. 2 may be implemented internally in the camera 100 or the output apparatus 300. .
 (3.視覚的な付加情報の例)
 図3は、本開示の一実施形態に係る付加情報の第1の例を示す図である。図3には、アクション画像1000と、アクション画像1000に透過的に重畳されている回転軸1010、回転軸ブレ1020、回転速度1030、回転軸ズレ1040のグラフィック表示とが示されている。アクション画像1000では、フィギュアスケートのスピンがアクションとして捉えられている。評価情報生成部203は、例えば、アクション画像1000から推定されるスケーターの姿勢に基づいて、スピンを評価する。例えば、評価情報生成部203は、画像認識の技術を利用して、RGB画像であるアクション画像1000からスケーターの像を抽出することによって姿勢を推定してもよい。あるいは、評価情報生成部203は、画像認識および空間認識の技術を利用して、RGB画像に深度情報が付加されたアクション画像1000からスケーターの像を抽出してもよい。深度情報に代えて、またはこれとともに、複数の位置から同時に撮影されたRGB画像を利用してもよい。さらに、評価情報生成部203は、スケーターの姿勢を推定するために、スケーターに装着された慣性センサの検出値を利用してもよい。
(3. Examples of additional visual information)
FIG. 3 is a diagram illustrating a first example of additional information according to an embodiment of the present disclosure. FIG. 3 shows an action image 1000 and a graphic display of a rotation axis 1010, a rotation axis blur 1020, a rotation speed 1030, and a rotation axis shift 1040 transparently superimposed on the action image 1000. In the action image 1000, the spin of figure skating is captured as an action. For example, the evaluation information generation unit 203 evaluates the spin based on the attitude of the skater estimated from the action image 1000. For example, the evaluation information generation unit 203 may estimate the posture by extracting a skater image from the action image 1000 that is an RGB image by using an image recognition technique. Alternatively, the evaluation information generation unit 203 may extract a skater image from the action image 1000 in which the depth information is added to the RGB image, using image recognition and space recognition techniques. Instead of or together with the depth information, RGB images taken simultaneously from a plurality of positions may be used. Furthermore, the evaluation information generation unit 203 may use a detection value of an inertial sensor attached to the skater in order to estimate the attitude of the skater.
 例えば上記のようにして推定されたスケーターの姿勢の変化を時系列で解析することによって、評価情報生成部203はスケーターが実行したスピンを評価することができる。図示された例では、評価情報生成部203が、スピンの回転軸(空間内の直線によって表現される)の時系列変化、およびスピンの回転速度を評価している。このような評価を含むアクション評価情報に基づいて、付加情報生成部209は、アクション画像とともに提示する付加情報を生成する。例えば、付加情報生成部209は、スピンの回転軸1010とともに、スケーターの足元(スケート靴の接地点)に対する相対的な軸の偏移を回転軸ブレ1020としてアクション画像1000上に表現する。このとき、付加情報生成部209は、スケーターの足元の位置や、回転軸の位置を特定するために、アクション画像1000に基づいて推定されたスケーターの姿勢、およびアクション画像1000内でのスケーターの位置の情報を利用する。また、付加情報生成部209は、特定された回転軸の位置に合わせて、回転速度1030を表示する。さらに、図示された例において、付加情報生成部209は、スケーターの足元の位置の遷移を、回転軸ズレ1040としてアクション画像1000上に表現する。スケートでは、この他にも、例えばジャンプの高さや距離、エッジの深さなどを示すグラフィックを、アクション画像1000に重畳して表示させてもよい。 For example, by analyzing the change in the attitude of the skater estimated as described above in a time series, the evaluation information generating unit 203 can evaluate the spin executed by the skater. In the illustrated example, the evaluation information generation unit 203 evaluates the time series change of the spin rotation axis (represented by a straight line in space) and the spin rotation speed. Based on the action evaluation information including such evaluation, the additional information generation unit 209 generates additional information to be presented together with the action image. For example, the additional information generation unit 209 expresses, on the action image 1000, the rotation of the axis relative to the foot of the skater (the ground contact point of the skate) along with the rotation axis 1010 of the spin as the rotation axis blur 1020. At this time, the additional information generation unit 209 determines the position of the skater in the action image 1000 and the posture of the skater estimated based on the action image 1000 in order to identify the position of the skater's foot or the position of the rotation axis. Use the information. Further, the additional information generation unit 209 displays the rotation speed 1030 according to the position of the specified rotation axis. Further, in the illustrated example, the additional information generation unit 209 represents the transition of the position of the skater's foot on the action image 1000 as a rotation axis shift 1040. In addition to this, in skating, for example, graphics indicating the jump height and distance, edge depth, and the like may be superimposed on the action image 1000 and displayed.
 図4は、本開示の一実施形態に係る付加情報の第2の例を示す図である。図4には、アクション画像1000と、アクション画像1000とともに表示される回転速度グラフ1050とが示されている。図示された例において、付加情報生成部209は、評価情報生成部203が生成したアクション評価情報に含まれる回転速度に基づいて、グラフ1050を描画する。このとき、付加情報生成部209は、例えば情報処理装置の内部のストレージに記録されている理想的な回転速度の変化を読み出して、アクション画像1000に映っているスケーターのスピンの回転速度と比較するグラフ1050を描画してもよい。なお、理想的な回転速度の情報は、情報処理装置の外部のストレージから読み出されてもよい。ここで、グラフ1050は、必ずしもアクション画像1000に重畳表示されなくてもよい付加情報の例である。上記で図3を参照して説明した回転軸1010、回転軸ブレ1020、回転速度1030、および回転軸ズレ1040が、基本的にアクション画像1000に透過的に重畳され、アクション画像1000に映っているスケーターの像とともに視認される付加情報であったのに対し、グラフ1050は、アクション画像1000に透過的に重畳されてもよいが、図示されているように不透過的に重畳されてもよく、またアクション画像1000に重畳されずにその近傍に別途表示されてもよい。 FIG. 4 is a diagram illustrating a second example of additional information according to an embodiment of the present disclosure. FIG. 4 shows an action image 1000 and a rotation speed graph 1050 displayed together with the action image 1000. In the illustrated example, the additional information generation unit 209 draws the graph 1050 based on the rotation speed included in the action evaluation information generated by the evaluation information generation unit 203. At this time, the additional information generation unit 209 reads, for example, an ideal change in the rotation speed recorded in the internal storage of the information processing apparatus and compares it with the rotation speed of the skater's spin shown in the action image 1000. The graph 1050 may be drawn. Note that information on the ideal rotation speed may be read from a storage external to the information processing apparatus. Here, the graph 1050 is an example of additional information that does not necessarily need to be superimposed on the action image 1000. The rotating shaft 1010, the rotating shaft blur 1020, the rotating speed 1030, and the rotating shaft shift 1040 described above with reference to FIG. 3 are basically transparently superimposed on the action image 1000 and reflected in the action image 1000. While the additional information is visible along with the skater image, the graph 1050 may be transparently superimposed on the action image 1000, but may be opaquely superimposed as illustrated. Further, it may be displayed separately in the vicinity thereof without being superimposed on the action image 1000.
 (4.付加情報の情報量の制御の例)
 次に、本開示の一実施形態における付加情報の情報量の制御の例について説明する。本実施形態では、上記で説明したように、アクション画像とともに付加情報が提示される。このとき、付加情報に含まれる情報量が多いほど、ユーザはアクション画像に含まれるアクション(上記の例ではフィギュアスケートにおけるアクション)に関する充実した情報を得ることができる。その一方で、例えばユーザが生中継でアクション画像を視聴している場合や、録画されたアクション画像を視聴していてもユーザにあまり時間がない場合、多すぎる情報量は、却ってアクション画像の視聴の妨げになる。そこで、本実施形態は、付加情報生成部209が、ユーザ状況情報に基づく付加情報の生成において、ユーザの状況に応じて付加情報の情報量を制御する構成を含みうる。以下では、そのような構成の一例について説明する。
(4. Example of controlling the amount of additional information)
Next, an example of controlling the amount of additional information according to an embodiment of the present disclosure will be described. In the present embodiment, as described above, the additional information is presented together with the action image. At this time, as the amount of information included in the additional information increases, the user can obtain more enriched information regarding the action included in the action image (in the above example, the action in figure skating). On the other hand, for example, when the user is watching an action image in a live broadcast, or when the user does not have much time even when watching a recorded action image, the amount of information is too large. It becomes an obstacle. Therefore, the present embodiment may include a configuration in which the additional information generation unit 209 controls the information amount of the additional information according to the user situation when generating the additional information based on the user situation information. Hereinafter, an example of such a configuration will be described.
 図5は、本開示の一実施形態における情報量が少ない付加情報の例を示す図である。図5には、アクション画像1100と、アクション画像1100に透過的に重畳されているジャンプ軌跡1110とが示されている。アクション画像1100では、フィギュアスケートのジャンプがアクションとして捉えられている。評価情報生成部203は、例えば、アクション画像1100から推定されるスケーターの姿勢に基づいて、ジャンプを評価する。評価情報生成部203によるスケーターの姿勢の推定は、例えば上記で図3などを参照して説明したスピンの例と同様に実施されうる。このようにして推定されたスケーターの姿勢の変化を時系列で解析することによって、評価情報生成部203はスケーターが実行したジャンプを評価することができる。図示された例では、評価情報生成部203が、ジャンプの踏み切り時(スケート靴が氷面から離れた瞬間)および着地時(スケート靴が氷面に接地した瞬間)におけるアクション画像1100内でのスケート靴の座標を特定する。付加情報生成部209は、ジャンプの踏み切りと着地との間におけるアクション画像1100内でのスケート靴の軌跡を、ジャンプ軌跡1110として描画する。 FIG. 5 is a diagram illustrating an example of additional information with a small amount of information according to an embodiment of the present disclosure. FIG. 5 shows an action image 1100 and a jump locus 1110 that is transparently superimposed on the action image 1100. In the action image 1100, a figure skating jump is captured as an action. For example, the evaluation information generation unit 203 evaluates the jump based on the skater posture estimated from the action image 1100. The estimation of the skater posture by the evaluation information generation unit 203 can be performed, for example, in the same manner as the spin example described above with reference to FIG. By analyzing the pose change of the skater estimated in this way in time series, the evaluation information generating unit 203 can evaluate the jump executed by the skater. In the illustrated example, the evaluation information generation unit 203 performs skating in the action image 1100 at the time of jump crossing (the moment when the skates are separated from the ice surface) and at the time of landing (the moment when the skates are touching the ice surface). Identify shoe coordinates. The additional information generation unit 209 draws the trajectory of the skate shoes in the action image 1100 between the jump crossing and landing as the jump trajectory 1110.
 ここで、図5に示されたジャンプ軌跡1110は、現にアクション画像1100に映っているスケーターのジャンプ軌跡1111と、同じスケーターによる昨年のジャンプ軌跡1112とを含む。付加情報生成部209は、昨年のジャンプ軌跡1112を、例えばストレージから読み出された情報に基づいて描画する。昨年のジャンプ軌跡1112を描画するための情報は、例えば参考情報として情報処理装置の内部の、または外部のストレージに格納されている。ジャンプ軌跡1111,1112の近傍には、それぞれ「TODAY」、「LAST YEAR」というテキストと、それぞれのジャンプの飛距離とを示すテキストが表示されている。図5に示された例でアクション画像1100とともに表示される付加情報では、ジャンプ軌跡1110のグラフィックおよびテキストによって、アクション画像1100に捉えられたジャンプに関する情報が簡潔に表現されている。 Here, the jump trajectory 1110 shown in FIG. 5 includes the skater's jump trajectory 1111 currently shown in the action image 1100 and the last year's jump trajectory 1112 by the same skater. The additional information generation unit 209 draws the last year's jump trajectory 1112 based on, for example, information read from the storage. Information for drawing last year's jump trajectory 1112 is stored in, for example, a storage inside or outside the information processing apparatus as reference information. In the vicinity of the jump trajectories 1111 and 1112, texts “TODAY” and “LAST YEAR” and texts indicating the jump distances of the respective jumps are displayed. In the additional information displayed together with the action image 1100 in the example shown in FIG. 5, the jump-related information captured in the action image 1100 is simply expressed by the graphic and text of the jump trajectory 1110.
 図6は、本開示の一実施形態における情報量が中程度の付加情報の例を示す図である。図6には、アクション画像1100と、ジャンプ軌跡1110と、ジャンプ情報1120とが示されている。図6に示される例では、ジャンプ軌跡1110に、アクション画像1100に映っているスケーター(仮にJane Smithとする)の今回のジャンプ軌跡1111と、Janeの昨年のジャンプ軌跡1112と、今回の優勝者(残念ながらJaneではなかった)のジャンプ軌跡1113とが含まれる。さらに、ジャンプ情報1120は、それぞれのジャンプ軌跡1110に関連付けて表示され、ジャンプの飛距離、高さ、およびジャンプ中の回転速度を示すテキストを含む。図6に示された例でアクション画像1100とともに表示される付加情報では、ジャンプ軌跡1110のグラフィック、およびジャンプ情報1120のテキストによって、アクション画像1100に捉えられたジャンプに関するより充実した情報が表現されている。 FIG. 6 is a diagram illustrating an example of additional information having a medium information amount according to an embodiment of the present disclosure. FIG. 6 shows an action image 1100, a jump trajectory 1110, and jump information 1120. In the example shown in FIG. 6, the jump trajectory 1110 includes the current jump trajectory 1111 of the skater (assumed to be Jane Smith) shown in the action image 1100, the jump trajectory 1112 last year of Jane, and the current winner ( Unfortunately not Jane) jump trajectory 1113. Furthermore, the jump information 1120 is displayed in association with each jump trajectory 1110 and includes text indicating the jump flight distance, height, and rotation speed during the jump. In the additional information displayed together with the action image 1100 in the example shown in FIG. 6, more complete information regarding the jump captured in the action image 1100 is expressed by the graphic of the jump trajectory 1110 and the text of the jump information 1120. Yes.
 図7は、本開示の一実施形態における情報量が多い付加情報の例を示す図である。図7には、アクション画像1100と、ジャンプ軌跡1110と、ジャンプ情報1120と、グラフ1130とが示されている。図7に示される例では、ジャンプ軌跡1110およびジャンプ情報1120が、Janeの今回のジャンプ、Janeの昨年のジャンプ、および今回の優勝者のジャンプの中から選択できるように表示されている。例えば、ユーザは、出力装置300において、タッチパネルやマウスなどのポインティングデバイスを用いて任意のジャンプ情報1120を選択することによって、いずれかのジャンプ軌跡1110およびジャンプ情報1120の組を表示させることができる。グラフ1130は、表示されているジャンプ軌跡1110およびジャンプ情報1120の組に対応し、当該ジャンプにおける回転速度の変化をグラフとして表示する。グラフ1130は、表示開始時には折り畳まれて(巻き取られて)おり、例えばユーザがポインティングデバイスでグラフ1130のアイコンを選択することによって、非表示部分1131が展開される。 FIG. 7 is a diagram illustrating an example of additional information with a large amount of information according to an embodiment of the present disclosure. FIG. 7 shows an action image 1100, a jump locus 1110, jump information 1120, and a graph 1130. In the example shown in FIG. 7, jump trajectory 1110 and jump information 1120 are displayed so that they can be selected from Jane's current jump, Jane's last year's jump, and current winner's jump. For example, the user can display any set of jump trajectory 1110 and jump information 1120 by selecting arbitrary jump information 1120 using a pointing device such as a touch panel or a mouse on the output device 300. A graph 1130 corresponds to the displayed set of jump trajectory 1110 and jump information 1120, and displays a change in rotational speed in the jump as a graph. The graph 1130 is folded (wound up) at the start of display, and the non-display portion 1131 is expanded when the user selects the icon of the graph 1130 with a pointing device, for example.
 上記の図7に示される例において、アクション画像1100とともに表示される付加情報は、図5および図6に示された例よりも多くの情報を含む。また、図5および図6に示された例とは異なり、図7に示された例では、アクション画像1100を視聴するユーザが内容を選択可能な付加情報が生成される。提供される情報量が多い場合、すべての情報を一度に表示するよりも、例えば同じ種類の情報についてはユーザの選択に応じて表示させる方が、視認性が向上しうる。従って、必ずしも情報量が最も多い付加情報の場合でなくても(例えば図6の示された例の場合でも)、選択可能な付加情報を生成することは有用でありうる。ただし、選択のためのユーザ操作の受け付けや、操作に応じた付加情報の内容の変更にはある程度の時間がかかるため、選択可能な付加情報が生成されるのは、ユーザがアクション画像1100を視聴するのにかける時間が長い場合、より具体的にはユーザがアクション画像1100をジャンプの区間で一時的に静止させる、または繰り返して再生させるような場合であることが望ましい。従って、本実施形態において、付加情報生成部209は、例えばユーザがアクション画像を参照するための持ち時間がある程度以上長いと推定される場合に、選択可能な付加情報を表示させる。 In the example shown in FIG. 7 above, the additional information displayed together with the action image 1100 includes more information than the examples shown in FIGS. In addition, unlike the examples shown in FIGS. 5 and 6, in the example shown in FIG. 7, additional information that allows the user viewing the action image 1100 to select the content is generated. When the amount of information to be provided is large, visibility can be improved, for example, by displaying the same type of information according to the user's selection, rather than displaying all information at once. Accordingly, it may be useful to generate selectable additional information even if it is not necessarily the case of additional information having the largest amount of information (for example, in the case of the example shown in FIG. 6). However, since it takes a certain amount of time to accept a user operation for selection and change the content of additional information according to the operation, selectable additional information is generated when the user views the action image 1100. When the time taken to do this is long, more specifically, it is desirable that the user temporarily stops the action image 1100 in the jump section or reproduces the action image 1100 repeatedly. Therefore, in the present embodiment, the additional information generation unit 209 displays selectable additional information, for example, when it is estimated that the user has a long time for referring to the action image to some extent.
 図8は、本開示の一実施形態における付加情報の情報量の制御のための処理の例を示すフローチャートである。図8を参照すると、まず、評価情報取得部205および状況情報取得部207が、アクション評価情報およびユーザ状況情報を取得する(S101)。次に、付加情報生成部209が、ユーザ状況情報から、ユーザがアクション画像とともに提供される付加情報を参照するための持ち時間を推定する。 FIG. 8 is a flowchart illustrating an example of processing for controlling the amount of additional information according to an embodiment of the present disclosure. Referring to FIG. 8, first, the evaluation information acquisition unit 205 and the situation information acquisition unit 207 acquire action evaluation information and user situation information (S101). Next, the additional information generation unit 209 estimates the time for the user to refer to the additional information provided together with the action image from the user situation information.
 ここで、持ち時間は、例えば、ユーザがアクション画像を生中継で視聴しているか、録画されたアクション画像を視聴しているかに基づいて推定されてもよい。例えば、アクターがアクション画像に含まれるアクションを実行した時刻と、ユーザがアクション画像を視聴する時刻との差が閾値を下回る場合、つまりユーザが生中継に近い状態でアクション画像を視聴している場合、ユーザが付加情報を参照するための持ち時間は短いと推定されうる。逆に、アクターがアクションを実行した時刻とユーザがアクション画像を視聴する時刻との差が閾値を上回る場合、つまりユーザが録画されたアクション画像を事後的に視聴している場合、ユーザが付加情報を参照するための持ち時間は長いと推定されうる。 Here, the holding time may be estimated based on, for example, whether the user is viewing the action image live or viewing the recorded action image. For example, when the difference between the time when the actor performed the action included in the action image and the time when the user views the action image is below the threshold, that is, when the user is viewing the action image in a state close to live broadcast The time for the user to refer to the additional information can be estimated to be short. Conversely, if the difference between the time when the actor performs the action and the time when the user views the action image exceeds the threshold, that is, if the user is viewing the recorded action image afterwards, the user will receive additional information. It can be estimated that the holding time for referring to is long.
 あるいは、ユーザが録画されたアクション画像を視聴する場合、ユーザが付加情報を参照するための持ち時間は、ユーザのスケジュールや、アクション画像の視聴にあたって設定された再生時間などに基づいて推定されてもよい。また、例えば、出力装置300において、ユーザが付加情報を参照するための持ち時間、またはユーザがアクション画像を視聴するにあたって適切と考える付加情報の量を示す情報が、ユーザによって入力されてもよい。 Alternatively, when the user views the recorded action image, the time for the user to refer to the additional information may be estimated based on the user's schedule, the playback time set for viewing the action image, or the like. Good. In addition, for example, in the output device 300, information indicating the time for the user to refer to the additional information or the amount of additional information that the user considers appropriate for viewing the action image may be input by the user.
 上記のように持ち時間を推定した付加情報生成部209は、推定された持ち時間が、第1の閾値th1を下回るか否かを判定する(S103)。ここで、持ち時間が閾値th1を下回る場合(YES)、付加情報生成部209は、例えば図5に示した例のような、情報量が少ない簡単な付加情報を生成する(S105)。あるいは、S105のように持ち時間がより短いと判定される場合、付加情報生成部209は、付加情報を生成しないことを決定してもよい。 The additional information generating unit 209 that has estimated the holding time as described above determines whether or not the estimated holding time is less than the first threshold th1 (S103). Here, when the possession time is less than the threshold th1 (YES), the additional information generation unit 209 generates simple additional information with a small amount of information, such as the example illustrated in FIG. 5 (S105). Alternatively, when it is determined that the holding time is shorter as in S105, the additional information generation unit 209 may determine not to generate additional information.
 S103の判定において、ユーザが付加情報を参照するための持ち時間が閾値th1を下回らない場合、さらに、付加情報生成部209は、推定された持ち時間が、第2の閾値th2を超えるか否かを判定する(S107)。ここで、持ち時間が閾値th2を超える場合(YES)、付加情報生成部209は、例えば図7に示した例のような、情報量がより多い、選択可能な付加情報を生成する(S109)。それ以外の場合(S107のNO)、付加情報生成部209は、例えば図6に示した例のような、情報量が中程度の付加情報を生成する(S111)。図示された例では、この場合の付加情報が通常の付加情報とされているが、通常の付加情報が定義されていなくてもよい。 If it is determined in step S103 that the time for the user to refer to the additional information does not fall below the threshold th1, the additional information generation unit 209 further determines whether or not the estimated time remaining exceeds the second threshold th2. Is determined (S107). Here, when the holding time exceeds the threshold th2 (YES), the additional information generation unit 209 generates selectable additional information with a larger amount of information, such as the example illustrated in FIG. 7 (S109). . In other cases (NO in S107), the additional information generation unit 209 generates additional information with a medium amount of information, such as the example illustrated in FIG. 6 (S111). In the illustrated example, the additional information in this case is normal additional information, but the normal additional information may not be defined.
 以上で説明したような付加情報の情報量の制御によれば、例えばユーザが付加情報を参照するための持ち時間に応じて、適切な量の情報を提示することができる。ユーザは、アクション画像を生中継で視聴する場合もあれば、録画されたアクション画像を視聴する場合もあり、またスケジュールなどのユーザの状況によっても適切な付加情報の量は変化しうるが、本実施形態ではそのような変化に柔軟に対応し、ユーザに適切な量の情報を提供することができうる。 According to the control of the information amount of the additional information as described above, an appropriate amount of information can be presented according to, for example, the time for the user to refer to the additional information. The user may view the action image live, or may view the recorded action image. The appropriate amount of additional information may vary depending on the user's situation such as a schedule. In the embodiment, it is possible to flexibly cope with such a change and provide an appropriate amount of information to the user.
 (5.付加情報の内容の制御の例)
 次に、本開示の一実施形態における付加情報の内容の制御の例について説明する。本実施形態では、上記で説明したように、アクション情報とともに付加情報が提示される。このとき、ユーザが付加情報としてどのような情報を必要とするかは、ユーザの状況、より具体的にはアクション画像に含まれるアクションについてのユーザの造詣の深さや、アクション画像を視聴しているユーザの数などによって異なりうる。そこで、本実施形態は、付加情報生成部209が、ユーザ状況情報に基づく付加情報の生成において、ユーザの状況に応じて付加情報の内容を制御する構成を含みうる。以下では、そのような構成の一例について説明する。
(5. Example of control of content of additional information)
Next, an example of control of the content of additional information according to an embodiment of the present disclosure will be described. In the present embodiment, as described above, additional information is presented together with action information. At this time, what kind of information the user needs as additional information depends on the user's situation, more specifically, the depth of the user's knowledge about the action included in the action image and the action image. It may vary depending on the number of users. Therefore, the present embodiment may include a configuration in which the additional information generation unit 209 controls the content of the additional information in accordance with the user situation when generating the additional information based on the user situation information. Hereinafter, an example of such a configuration will be described.
 図9は、本開示の一実施形態において一人でアクション画像を視聴するユーザ向けの付加情報の例を示す図である。図9には、アクション画像1100と、アクション画像1100に透過的に重畳されているジャンプ軌跡1110と、ジャンプ情報1120とが示されている。図9に示された例において、ジャンプ軌跡1110およびジャンプ情報1120は、アクション画像1100に映っているスケーター(Jane Smith)のジャンプ、アクション画像1100を視聴しているユーザが仮にジャンプした場合の想定されるジャンプ(「YOU」とラベルされている)、ユーザの同年代(30's)の想定されるジャンプ、およびユーザの若いころ(YOU IN YOUTH)の想定されるジャンプについて表示されている。このようなジャンプ軌跡1110およびジャンプ情報1120は、例えば、Jane Smithのジャンプに関するアクション評価情報と、ユーザの基礎体力データ、年齢、および年代ごとの平均的な基礎体力データに基づいて描画される。 FIG. 9 is a diagram illustrating an example of additional information for a user who views an action image alone in an embodiment of the present disclosure. FIG. 9 shows an action image 1100, a jump trajectory 1110 that is transparently superimposed on the action image 1100, and jump information 1120. In the example shown in FIG. 9, the jump trajectory 1110 and the jump information 1120 are assumed to be a jump of a skater (Jane Smith) shown in the action image 1100 and a user who is watching the action image 1100 jumps temporarily. Jumps (labeled “YOU”), assumed jumps of the user's same age (30's), and assumed jumps of the user's youth (YOU IN YOUTH). Such jump trajectory 1110 and jump information 1120 are drawn based on, for example, action evaluation information related to Jane Smith's jump and average basic physical strength data for each user's basic physical strength data, age, and age.
 図10は、本開示の一実施形態において複数でアクション画像を視聴するユーザ向けの付加情報の例を示す図である。図10に示す例において、付加情報は、アクション画像1100に透過的に重畳されているジャンプ軌跡1110と、ジャンプ情報1120と、ジャンプに関するランキング1140とを含む。図10に示された例において、ジャンプ軌跡1110およびジャンプ情報1120は、アクション画像1100に映っているスケーター(Jane Smith)のジャンプ、およびアクション画像1100を視聴している複数のユーザ(JOHNNYとDAD)のそれぞれが仮にジャンプした場合の想定されるジャンプについて表示されている。このようなジャンプ軌跡1110およびジャンプ情報1120は、例えば、Jane Smithのジャンプに関するアクション評価情報と、それぞれのユーザの基礎体力データに基づいて描画される。また、ランキング1140には、JOHNNYとDADのそれぞれが仮にジャンプした場合の想定されるジャンプについて、想定される飛距離、回転速度、および総合のランキングが表示されている。 FIG. 10 is a diagram illustrating an example of additional information for a user who views a plurality of action images according to an embodiment of the present disclosure. In the example illustrated in FIG. 10, the additional information includes a jump locus 1110 that is transparently superimposed on the action image 1100, jump information 1120, and a ranking 1140 related to the jump. In the example shown in FIG. 10, the jump trajectory 1110 and the jump information 1120 are a skater (Jane Smith) jump shown in the action image 1100 and a plurality of users (JOHNNY and DAD) who are watching the action image 1100. It is displayed about the assumed jump when each jumps temporarily. Such jump trajectory 1110 and jump information 1120 are drawn based on, for example, action evaluation information regarding Jane Smith's jump and basic physical strength data of each user. The ranking 1140 displays the assumed flying distance, rotational speed, and overall ranking for the assumed jumps when JOHNNY and DAD each jump.
 上記で図9および図10に示された例では、付加情報生成部209が、アクション画像を視聴しているユーザの数によって、付加情報の内容を制御している。より具体的には、付加情報生成部209は、出力装置300でアクション画像を視聴している一人のユーザが検出された場合には、スケーター(Jane Smith)に加えて、一人のユーザ自身について、ユーザ自身、同年代、および若いころの仮想的なジャンプの情報を含む付加情報を生成する。一方、付加情報生成部209は、出力装置300でアクション画像を視聴している複数のユーザが検出された場合には、Jane Smithに加えて、複数のユーザの仮想的なジャンプを比較したり、ランキング化したりする付加情報を生成する。これによって、一人でアクション画像を視聴しているユーザは、自分自身に関する身近な情報が多く提供されることによってJane Smithのジャンプのレベルを実感することができる。また、複数でアクション画像を視聴しているユーザは、それぞれのユーザに関する仮想的なジャンプの情報が提供され、さらにそれらを競わせるようなランキングなどの情報が提供されることによって、例えば互いの間での話題を得ることができる。 In the example shown in FIGS. 9 and 10 above, the additional information generation unit 209 controls the content of the additional information according to the number of users viewing the action image. More specifically, when one user who is viewing an action image is detected by the output device 300, the additional information generation unit 209, in addition to the skater (Jane Smith), Additional information including virtual jump information of the user himself, the same age, and young age is generated. On the other hand, when a plurality of users viewing an action image is detected by the output device 300, the additional information generation unit 209 compares virtual jumps of a plurality of users in addition to Jane Smith, Generate additional information to rank. As a result, the user who is watching the action image alone can feel the jump level of Jane Smith by providing a lot of familiar information about himself. In addition, users who are viewing action images in a plurality are provided with virtual jump information about each user, and further provided with information such as ranking that allows them to compete, for example, between each other. You can get a topic at.
 なお、上記で図9および図10に示した例のような付加情報は、いずれも、例えばアクション(図示された例ではフィギュアスケート)に関する造詣があまり深くないユーザのための付加情報でありうる。例えば、ユーザのプロフィールや、画像などのコンテンツの視聴履歴などから、ユーザのフィギュアスケートに関する造詣が深いと推定される場合、付加情報生成部209は、例えば上記で図7を参照して説明した例のように、他のスケーターや過去の記録などとの比較を可能にする付加情報を提供してもよい。この場合、フィギュアスケートに造詣が深いユーザは、競技として、様々なスケーターや過去の記録との比較に興味があると想定されている。一方、造詣があまり深くないユーザは、自分自身や近くにいるユーザとの比較から、スケーターのレベルの高さを実感することに興味があると想定されている。なお、このような想定は一例にすぎず、様々な想定に基づいて付加情報を生成することが可能である。 It should be noted that the additional information such as the example shown in FIGS. 9 and 10 can be additional information for a user who is not deeply familiar with the action (figure skating in the illustrated example), for example. For example, when it is estimated that the user's knowledge about figure skating is deep from the user's profile, viewing history of content such as images, the additional information generation unit 209 is, for example, the example described above with reference to FIG. As described above, additional information that enables comparison with other skaters or past records may be provided. In this case, it is assumed that users who are familiar with figure skating are interested in comparison with various skaters and past records as competitions. On the other hand, users who are not very proficient are assumed to be interested in realizing a high level of skater based on comparison with themselves and nearby users. Note that such an assumption is merely an example, and additional information can be generated based on various assumptions.
 (6.他の付加情報の例)
 図11は、本開示の一実施形態における他の付加情報の第1の例を示す図である。図11には、アクション画像1200と、アクション画像1200に表示される合成画像1210とが示されている。アクション画像1200では、シンクロナイズドスイミングの演技がアクションとして捉えられている。評価情報生成部203は、例えば、アクション画像1200から推定されるスイマーの姿勢に基づいて、シンクロナイズドスイミングの演技を評価する。より具体的には、評価情報生成部203は、それぞれのスイマーの動きがどれだけ同期しているかによって演技を評価する。例えば、評価情報生成部203は、それぞれのスイマーの動きを時系列で比較し、動き出しのタイミングや、動きの大きさ、動きの速さなどを比較する。
(6. Examples of other additional information)
FIG. 11 is a diagram illustrating a first example of other additional information according to an embodiment of the present disclosure. FIG. 11 shows an action image 1200 and a composite image 1210 displayed on the action image 1200. In the action image 1200, the performance of synchronized swimming is captured as an action. For example, the evaluation information generation unit 203 evaluates the performance of synchronized swimming based on the posture of the swimmer estimated from the action image 1200. More specifically, the evaluation information generation unit 203 evaluates the performance according to how much the movements of the respective swimmers are synchronized. For example, the evaluation information generation unit 203 compares the movements of the respective swimmers in time series, and compares the movement start timing, the magnitude of movement, the speed of movement, and the like.
 一方、付加情報生成部209は、アクション画像1200から抽出されたスイマーの画像を透過的に互いに重畳させて、合成画像1210を生成する。より具体的には、付加情報生成部209は、評価情報生成部203によって推定されたスイマーの姿勢に基づいて、それぞれのスイマーの画像1211~1214の位置を合わせ、透過的に重畳させることによって合成画像1210を生成する。これによって、アクション画像1200を視聴するユーザは、それぞれのスイマーの動きがどれだけ同期しているかを容易に把握することができる。 On the other hand, the additional information generation unit 209 transparently superimposes the swimmer images extracted from the action image 1200 to generate a composite image 1210. More specifically, the additional information generation unit 209 combines the positions of the images 1211 to 1214 of the respective swimmers based on the postures of the swimmers estimated by the evaluation information generation unit 203 and transparently superimposes them. An image 1210 is generated. Thus, the user who views the action image 1200 can easily grasp how much the movements of the respective swimmers are synchronized.
 図12は、本開示の一実施形態における他の付加情報の第2の例を示す図である。図12には、アクション画像1300と、アクション画像1300に表示されるパンチアイコン1310とが示されている。アクション画像1300では、ボクシングにおけるパンチがアクションとして捉えられている。評価情報生成部203は、例えば、グローブに内蔵された衝撃センサの検出値に基づいて、パンチの強さを推定する。さらに、評価情報生成部203は、アクション画像1300から推定されるボクサーの姿勢に基づいて、パンチの時のグローブとボクサーの体のコンタクト位置、つまりパンチが当たった位置を推定する。例えば、評価情報生成部203は、パンチが当たった位置とパンチの強さとに基づいて、パンチが有効打であるか否かを判定する。また、評価情報生成部203は、有効打でないものも含むパンチの数、および当たった位置の傾向から、ボクサーの積極性を評価してもよい。 FIG. 12 is a diagram illustrating a second example of other additional information according to an embodiment of the present disclosure. FIG. 12 shows an action image 1300 and a punch icon 1310 displayed on the action image 1300. In the action image 1300, a punch in boxing is captured as an action. For example, the evaluation information generation unit 203 estimates the strength of the punch based on the detection value of the impact sensor built in the glove. Furthermore, the evaluation information generation unit 203 estimates the contact position between the glove and the boxer body at the time of punching, that is, the position where the punch hits, based on the posture of the boxer estimated from the action image 1300. For example, the evaluation information generation unit 203 determines whether or not the punch is an effective hit based on the position where the punch hits and the strength of the punch. Further, the evaluation information generation unit 203 may evaluate the aggressiveness of the boxer from the number of punches including those that are not effective hits and the tendency of the hit positions.
 一方、付加情報生成部209は、評価情報生成部203によって推定されたパンチが当たった位置と、パンチの強さとに基づいて、アクション画像1300にパンチアイコン1310を表示させる。より具体的には、付加情報生成部209は、推定されたパンチが当たった位置に応じてパンチアイコン1310の表示位置を決定し、推定されたパンチの強さに基づいてパンチアイコン1310の大きさや色を決定する。また、付加情報生成部209は、対戦しているボクサーのうちのどちらのパンチかによって、パンチアイコン1310の色を変化させてもよい。これによって、アクション画像1300を視聴するユーザは、それぞれのボクサーが放ったパンチがどこに当たったか、およびどのくらいの強さであったかを、直観的に把握することができる。 On the other hand, the additional information generation unit 209 displays a punch icon 1310 on the action image 1300 based on the punch hit position estimated by the evaluation information generation unit 203 and the punch strength. More specifically, the additional information generation unit 209 determines the display position of the punch icon 1310 according to the estimated position where the punch hits, and determines the size of the punch icon 1310 based on the estimated punch strength. Determine the color. Further, the additional information generation unit 209 may change the color of the punch icon 1310 depending on which punch of the fighting boxers is punched. As a result, the user who views the action image 1300 can intuitively understand where the punches emitted by the respective boxers hit and how strong they were.
 図13は、本開示の一実施形態における他の付加情報の第3の例を示す図である。図13には、アクション画像1400と、アクション画像1400に表示される接地領域1410とが示されている。アクション画像1400では、柔道における投げ技がアクションとして捉えられている。評価情報生成部203は、例えば、アクション画像1400から推定される競技者姿勢と、アクション画像1400内で認識されている床面との関係に基づいて、投げ技の有効性を評価する。例えば、投げられた競技者が肩から床に落ちたか、背中から落ちたかによって、投げ技の有効性は異なる。この場合、評価情報生成部203は、競技者の骨格モデルも参照して、投げられた競技者の体と床とのコンタクト位置が、肩であったか背中であったかを判定する。これによって、投げ技の有効性、およびそれに基づいて与えられるポイントを判定することができる。 FIG. 13 is a diagram illustrating a third example of other additional information according to an embodiment of the present disclosure. FIG. 13 shows an action image 1400 and a ground area 1410 displayed on the action image 1400. In the action image 1400, a throwing technique in judo is captured as an action. For example, the evaluation information generation unit 203 evaluates the effectiveness of the throwing technique based on the relationship between the player posture estimated from the action image 1400 and the floor surface recognized in the action image 1400. For example, the effectiveness of throwing techniques varies depending on whether the thrown athlete fell from the shoulder to the floor or from the back. In this case, the evaluation information generation unit 203 also refers to the athlete's skeleton model to determine whether the contact position between the thrown athlete's body and the floor is the shoulder or the back. Thereby, the effectiveness of the throwing technique and the points given based on it can be determined.
 一方、付加情報生成部209は、評価情報生成部203によって推定された、投げられた競技者の体と床とのコンタクト位置に基づいて、アクション画像1400に接地領域1410を表示させる。より具体的には、付加情報生成部209は、推定されたコンタクト位置に対応する領域に、所定の色のグラフィックを表示させる。また、付加情報生成部209は、例えばコンタクト位置が肩であったか背中であったかによって、接地領域1410の色などを変化させてもよい。これによって、アクション画像1400を視聴するユーザは、投げ技にポイントが与えられた、または与えられなかったのはなぜかを、直観的に把握することができる。 On the other hand, the additional information generation unit 209 causes the action image 1400 to display the ground contact area 1410 based on the contact position between the body of the thrown player and the floor estimated by the evaluation information generation unit 203. More specifically, the additional information generation unit 209 displays a graphic of a predetermined color in an area corresponding to the estimated contact position. Further, the additional information generation unit 209 may change the color or the like of the ground contact area 1410 depending on, for example, whether the contact position is the shoulder or the back. Accordingly, the user who views the action image 1400 can intuitively understand why the points are given or not given to the throwing technique.
 例えば、上記で図11~図13を参照して説明したような付加情報、および付加情報の基になるアクション評価情報は、アクションの判定または採点に関連する情報(シンクロナイズドスイミングにおけるスイマーの動きの同期、ボクシングにおけるパンチのヒット位置と強さ、および柔道における投げられた競技者の接地位置)を含む。従って、上記の例において、アクション評価情報は、アクションを判定または採点した情報を含み、付加情報は、判定または採点の経過または結果を示すともいえる。なお、上記の例では付加情報が採点の経過を示しているが、同様にして結果(与えられたポイントなど)が示されてもよい。このような情報は、例えば娯楽としてアクション画像を視聴するユーザには限らず、例えば審判または審査員のように、実際にアクションに対して評価を下すユーザに、評価の補助のために提供されてもよい。 For example, the additional information as described above with reference to FIGS. 11 to 13 and the action evaluation information based on the additional information include information related to action determination or scoring (synchronization of swimmer movements in synchronized swimming). , The hit position and strength of the punch in boxing, and the grounded position of the thrown athlete in judo). Therefore, in the above example, the action evaluation information includes information on determining or scoring the action, and the additional information can be said to indicate the progress or result of the determination or scoring. In the above example, the additional information indicates the progress of scoring, but the result (eg, given points) may be indicated in the same manner. Such information is not limited to users who watch action images as entertainment, for example, but is provided to assist users who actually evaluate actions, such as referees or judges. Also good.
 以上で説明したような他の付加情報とユーザの状況との関係について、一例をあげると、付加情報生成部209は、アクション評価情報に基づいて上記のような付加情報を生成するか否かを、アクション画像に捉えられた競技についてのユーザの造詣の深さに応じて決定してもよい。例えば、ユーザがその競技について深い造詣をもつ場合、上記の例のような付加情報は、むしろ観戦の邪魔になる可能性もある。その一方で、ユーザがその競技について詳しくない場合、上記の例のような付加情報は、例えば審判の判定によって与えられるポイントがどのような理由で与えられたかをユーザが理解するために有用でありうる。そこで、付加情報生成部209は、ユーザの競技についての造詣の深さに応じて、造詣が深い場合には付加情報を生成しない、または判定の結果のみを表示する簡単な付加情報を生成し、そうでない場合には上記の例のようなグラフィックを伴うわかりやすい付加情報を生成してもよい。 As an example of the relationship between the other additional information and the user situation as described above, the additional information generation unit 209 determines whether or not to generate the additional information as described above based on the action evaluation information. The determination may be made according to the depth of the user's knowledge about the competition captured in the action image. For example, if the user has a deep knowledge of the game, the additional information as in the above example may rather interfere with watching the game. On the other hand, if the user is not familiar with the game, the additional information as in the above example is useful for the user to understand why the points given by the judgment of the referee were given, for example. sell. Therefore, the additional information generation unit 209 generates simple additional information that does not generate additional information when the knowledge is deep, or displays only the determination result, according to the depth of knowledge about the user's competition, Otherwise, easy-to-understand additional information with graphics as in the above example may be generated.
 別の例として、付加情報生成部209は、上記で図5~図7を参照して説明したような付加情報と、図9および図10を参照して説明したような付加情報とを、ユーザの造詣の深さに応じて選択的に生成してもよい。例えば、ユーザがフィギュアスケートについて深い造詣をもつ場合、図5~図7を参照して説明したような、純粋に競技の内容に関する付加情報が適切でありうる。一方、ユーザがフィギュアスケートについて詳しくない場合、図9および図10を参照して説明したような、ユーザ自身などの身近な例とアクション画像に映っているスケーターの演技とを比較する付加情報が適切でありうる。このように、付加情報生成部209は、付加情報の内容を、アクションについてのユーザの造詣の深さに応じて変更してもよい。 As another example, the additional information generation unit 209 uses the additional information as described above with reference to FIGS. 5 to 7 and the additional information as described with reference to FIGS. It may be selectively generated according to the depth of the slag. For example, if the user has a deep knowledge of figure skating, additional information regarding the content of the competition, as described with reference to FIGS. 5 to 7, may be appropriate. On the other hand, if the user is not familiar with figure skating, additional information for comparing a familiar example such as the user himself with the skater performance shown in the action image as described with reference to FIGS. 9 and 10 is appropriate. It can be. As described above, the additional information generation unit 209 may change the content of the additional information according to the depth of the user's knowledge about the action.
 (7.ハードウェア構成)
 次に、図14を参照して、本開示の実施形態に係る情報処理装置のハードウェア構成について説明する。図14は、本開示の実施形態に係る情報処理装置のハードウェア構成例を示すブロック図である。図示された情報処理装置900は、例えば、上記の実施形態におけるサーバ装置、カメラ、または出力装置を実現しうる。
(7. Hardware configuration)
Next, a hardware configuration of the information processing apparatus according to the embodiment of the present disclosure will be described with reference to FIG. FIG. 14 is a block diagram illustrating a hardware configuration example of the information processing apparatus according to the embodiment of the present disclosure. The illustrated information processing apparatus 900 can realize, for example, the server apparatus, camera, or output apparatus in the above-described embodiment.
 情報処理装置900は、CPU(Central Processing unit)901、ROM(Read Only Memory)903、およびRAM(Random Access Memory)905を含む。また、情報処理装置900は、ホストバス907、ブリッジ909、外部バス911、インターフェース913、入力装置915、出力装置917、ストレージ装置919、ドライブ921、接続ポート923、通信装置925を含んでもよい。さらに、情報処理装置900は、必要に応じて、撮像装置933、およびセンサ935を含んでもよい。情報処理装置900は、CPU901に代えて、またはこれとともに、DSP(Digital Signal Processor)、ASIC(Application Specific Integrated Circuit)、またはFPGA(Field-Programmable Gate Array)などの処理回路を有してもよい。 The information processing apparatus 900 includes a CPU (Central Processing unit) 901, a ROM (Read Only Memory) 903, and a RAM (Random Access Memory) 905. The information processing apparatus 900 may include a host bus 907, a bridge 909, an external bus 911, an interface 913, an input device 915, an output device 917, a storage device 919, a drive 921, a connection port 923, and a communication device 925. Furthermore, the information processing apparatus 900 may include an imaging device 933 and a sensor 935 as necessary. The information processing apparatus 900 may include a processing circuit such as a DSP (Digital Signal Processor), an ASIC (Application Specific Integrated Circuit), or an FPGA (Field-Programmable Gate Array) instead of or in addition to the CPU 901.
 CPU901は、演算処理装置および制御装置として機能し、ROM903、RAM905、ストレージ装置919、またはリムーバブル記録媒体927に記録された各種プログラムに従って、情報処理装置900内の動作全般またはその一部を制御する。ROM903は、CPU901が使用するプログラムや演算パラメータなどを記憶する。RAM905は、CPU901の実行において使用するプログラムや、その実行において適宜変化するパラメータなどを一次記憶する。CPU901、ROM903、およびRAM905は、CPUバスなどの内部バスにより構成されるホストバス907により相互に接続されている。さらに、ホストバス907は、ブリッジ909を介して、PCI(Peripheral Component Interconnect/Interface)バスなどの外部バス911に接続されている。 The CPU 901 functions as an arithmetic processing device and a control device, and controls all or a part of the operation in the information processing device 900 according to various programs recorded in the ROM 903, the RAM 905, the storage device 919, or the removable recording medium 927. The ROM 903 stores programs and calculation parameters used by the CPU 901. The RAM 905 primarily stores programs used in the execution of the CPU 901, parameters that change as appropriate during the execution, and the like. The CPU 901, the ROM 903, and the RAM 905 are connected to each other by a host bus 907 configured by an internal bus such as a CPU bus. Further, the host bus 907 is connected to an external bus 911 such as a PCI (Peripheral Component Interconnect / Interface) bus via a bridge 909.
 入力装置915は、例えば、マウス、キーボード、タッチパネル、ボタン、スイッチおよびレバーなど、ユーザによって操作される装置である。入力装置915は、例えば、赤外線やその他の電波を利用したリモートコントロール装置であってもよいし、情報処理装置900の操作に対応した携帯電話などの外部接続機器929であってもよい。入力装置915は、ユーザが入力した情報に基づいて入力信号を生成してCPU901に出力する入力制御回路を含む。ユーザは、この入力装置915を操作することによって、情報処理装置900に対して各種のデータを入力したり処理動作を指示したりする。 The input device 915 is a device operated by the user, such as a mouse, a keyboard, a touch panel, a button, a switch, and a lever. The input device 915 may be, for example, a remote control device that uses infrared rays or other radio waves, or may be an external connection device 929 such as a mobile phone that supports the operation of the information processing device 900. The input device 915 includes an input control circuit that generates an input signal based on information input by the user and outputs the input signal to the CPU 901. The user operates the input device 915 to input various data and instruct processing operations to the information processing device 900.
 出力装置917は、取得した情報をユーザに対して視覚や聴覚、触覚などの感覚を用いて通知することが可能な装置で構成される。出力装置917は、例えば、LCD(Liquid Crystal Display)または有機EL(Electro-Luminescence)ディスプレイなどの表示装置、スピーカまたはヘッドフォンなどの音声出力装置、もしくはバイブレータなどでありうる。出力装置917は、情報処理装置900の処理により得られた結果を、テキストもしくは画像などの映像、音声もしくは音響などの音声、またはバイブレーションなどとして出力する。 The output device 917 is configured by a device capable of notifying the acquired information to the user using a sense such as vision, hearing, or touch. The output device 917 can be, for example, a display device such as an LCD (Liquid Crystal Display) or an organic EL (Electro-Luminescence) display, an audio output device such as a speaker or headphones, or a vibrator. The output device 917 outputs the result obtained by the processing of the information processing device 900 as video such as text or image, sound such as sound or sound, or vibration.
 ストレージ装置919は、情報処理装置900の記憶部の一例として構成されたデータ格納用の装置である。ストレージ装置919は、例えば、HDD(Hard Disk Drive)などの磁気記憶部デバイス、半導体記憶デバイス、光記憶デバイス、または光磁気記憶デバイスなどにより構成される。ストレージ装置919は、例えばCPU901が実行するプログラムや各種データ、および外部から取得した各種のデータなどを格納する。 The storage device 919 is a data storage device configured as an example of a storage unit of the information processing device 900. The storage device 919 includes, for example, a magnetic storage device such as an HDD (Hard Disk Drive), a semiconductor storage device, an optical storage device, or a magneto-optical storage device. The storage device 919 stores, for example, programs executed by the CPU 901 and various data, and various data acquired from the outside.
 ドライブ921は、磁気ディスク、光ディスク、光磁気ディスク、または半導体メモリなどのリムーバブル記録媒体927のためのリーダライタであり、情報処理装置900に内蔵、あるいは外付けされる。ドライブ921は、装着されているリムーバブル記録媒体927に記録されている情報を読み出して、RAM905に出力する。また、ドライブ921は、装着されているリムーバブル記録媒体927に記録を書き込む。 The drive 921 is a reader / writer for a removable recording medium 927 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory, and is built in or externally attached to the information processing apparatus 900. The drive 921 reads information recorded on the attached removable recording medium 927 and outputs the information to the RAM 905. In addition, the drive 921 writes a record in the attached removable recording medium 927.
 接続ポート923は、機器を情報処理装置900に接続するためのポートである。接続ポート923は、例えば、USB(Universal Serial Bus)ポート、IEEE1394ポート、SCSI(Small Computer System Interface)ポートなどでありうる。また、接続ポート923は、RS-232Cポート、光オーディオ端子、HDMI(登録商標)(High-Definition Multimedia Interface)ポートなどであってもよい。接続ポート923に外部接続機器929を接続することで、情報処理装置900と外部接続機器929との間で各種のデータが交換されうる。 The connection port 923 is a port for connecting a device to the information processing apparatus 900. The connection port 923 can be, for example, a USB (Universal Serial Bus) port, an IEEE 1394 port, a SCSI (Small Computer System Interface) port, or the like. The connection port 923 may be an RS-232C port, an optical audio terminal, an HDMI (registered trademark) (High-Definition Multimedia Interface) port, or the like. By connecting the external connection device 929 to the connection port 923, various types of data can be exchanged between the information processing apparatus 900 and the external connection device 929.
 通信装置925は、例えば、通信ネットワーク931に接続するための通信デバイスなどで構成された通信インターフェースである。通信装置925は、例えば、LAN(Local Area Network)、Bluetooth(登録商標)、Wi-Fi、またはWUSB(Wireless USB)用の通信カードなどでありうる。また、通信装置925は、光通信用のルータ、ADSL(Asymmetric Digital Subscriber Line)用のルータ、または、各種通信用のモデムなどであってもよい。通信装置925は、例えば、インターネットや他の通信機器との間で、TCP/IPなどの所定のプロトコルを用いて信号などを送受信する。また、通信装置925に接続される通信ネットワーク931は、有線または無線によって接続されたネットワークであり、例えば、インターネット、家庭内LAN、赤外線通信、ラジオ波通信または衛星通信などを含みうる。 The communication device 925 is a communication interface configured with, for example, a communication device for connecting to the communication network 931. The communication device 925 can be, for example, a communication card for LAN (Local Area Network), Bluetooth (registered trademark), Wi-Fi, or WUSB (Wireless USB). The communication device 925 may be a router for optical communication, a router for ADSL (Asymmetric Digital Subscriber Line), or a modem for various communication. The communication device 925 transmits and receives signals and the like using a predetermined protocol such as TCP / IP with the Internet and other communication devices, for example. The communication network 931 connected to the communication device 925 is a network connected by wire or wireless, and may include, for example, the Internet, a home LAN, infrared communication, radio wave communication, satellite communication, or the like.
 撮像装置933は、例えば、CMOS(Complementary Metal Oxide Semiconductor)またはCCD(Charge Coupled Device)などの撮像素子、および撮像素子への被写体像の結像を制御するためのレンズなどの各種の部材を用いて実空間を撮像し、撮像画像を生成する装置である。撮像装置933は、静止画を撮像するものであってもよいし、また動画を撮像するものであってもよい。 The imaging device 933 uses various members such as an imaging element such as a CMOS (Complementary Metal Oxide Semiconductor) or a CCD (Charge Coupled Device), and a lens for controlling the formation of a subject image on the imaging element. It is an apparatus that images a real space and generates a captured image. The imaging device 933 may capture a still image or may capture a moving image.
 センサ935は、例えば、加速度センサ、角速度センサ、地磁気センサ、照度センサ、温度センサ、気圧センサ、または音センサ(マイクロフォン)などの各種のセンサである。センサ935は、例えば情報処理装置900の筐体の姿勢など、情報処理装置900自体の状態に関する情報や、情報処理装置900の周辺の明るさや騒音など、情報処理装置900の周辺環境に関する情報を取得する。また、センサ935は、GPS(Global Positioning System)信号を受信して装置の緯度、経度および高度を測定するGPS受信機を含んでもよい。 The sensor 935 is various sensors such as an acceleration sensor, an angular velocity sensor, a geomagnetic sensor, an illuminance sensor, a temperature sensor, an atmospheric pressure sensor, or a sound sensor (microphone). The sensor 935 acquires information about the state of the information processing apparatus 900 itself, such as the posture of the information processing apparatus 900, and information about the surrounding environment of the information processing apparatus 900, such as brightness and noise around the information processing apparatus 900, for example. To do. The sensor 935 may include a GPS receiver that receives a GPS (Global Positioning System) signal and measures the latitude, longitude, and altitude of the device.
 以上、情報処理装置900のハードウェア構成の一例を示した。上記の各構成要素は、汎用的な部材を用いて構成されていてもよいし、各構成要素の機能に特化したハードウェアにより構成されていてもよい。かかる構成は、実施する時々の技術レベルに応じて適宜変更されうる。 Heretofore, an example of the hardware configuration of the information processing apparatus 900 has been shown. Each component described above may be configured using a general-purpose member, or may be configured by hardware specialized for the function of each component. Such a configuration can be appropriately changed according to the technical level at the time of implementation.
 (8.補足)
 本開示の実施形態は、例えば、上記で説明したような情報処理装置、システム、情報処理装置またはシステムで実行される情報処理方法、情報処理装置を機能させるためのプログラム、およびプログラムが記録された一時的でない有形の媒体を含みうる。
(8. Supplement)
In the embodiment of the present disclosure, for example, an information processing apparatus, a system, an information processing method executed by the information processing apparatus or system, a program for causing the information processing apparatus to function, and a program are recorded. It may include tangible media that is not temporary.
 以上、添付図面を参照しながら本開示の好適な実施形態について詳細に説明したが、本開示の技術的範囲はかかる例に限定されない。本開示の技術分野における通常の知識を有する者であれば、請求の範囲に記載された技術的思想の範疇内において、各種の変更例または修正例に想到し得ることは明らかであり、これらについても、当然に本開示の技術的範囲に属するものと了解される。 The preferred embodiments of the present disclosure have been described in detail above with reference to the accompanying drawings, but the technical scope of the present disclosure is not limited to such examples. It is obvious that a person having ordinary knowledge in the technical field of the present disclosure can come up with various changes or modifications within the scope of the technical idea described in the claims. Of course, it is understood that it belongs to the technical scope of the present disclosure.
 また、本明細書に記載された効果は、あくまで説明的または例示的なものであって限定的ではない。つまり、本開示に係る技術は、上記の効果とともに、または上記の効果に代えて、本明細書の記載から当業者には明らかな他の効果を奏しうる。 In addition, the effects described in this specification are merely illustrative or illustrative, and are not limited. That is, the technology according to the present disclosure can exhibit other effects that are apparent to those skilled in the art from the description of the present specification in addition to or instead of the above effects.
 なお、以下のような構成も本開示の技術的範囲に属する。
(1)アクターのアクションを含むアクション画像に基づいて前記アクションを評価したアクション評価情報を取得する評価情報取得部と、
 前記アクション画像を視聴するユーザの状況を示すユーザ状況情報を取得する状況情報取得部と、
 前記アクション画像とともに前記ユーザに提示され、前記アクション評価情報の少なくとも一部を表現する付加情報を、前記ユーザ状況情報に応じて生成する付加情報生成部と
 を備える情報処理装置。
(2)前記付加情報は、視覚的な情報を含む、前記(1)に記載の情報処理装置。
(3)前記付加情報は、前記アクション画像に重畳される、または前記アクション画像の近傍に表示されるテキストまたはグラフィックを含む、前記(2)に記載の情報処理装置。
(4)前記アクション評価情報は、前記アクション画像内での前記アクターの体の部分の座標を示す情報を含み、
 前記付加情報は、前記部分の軌跡を表現するグラフィックを含む、前記(3)に記載の情報処理装置。
(5)前記アクション評価情報は、前記アクションを判定または採点した情報を含み、
 前記付加情報は、前記判定または採点の経過または結果を示すテキストまたはグラフィックを含む、前記(3)または(4)に記載の情報処理装置。
(6)前記ユーザ状況情報は、前記付加情報を参照するための前記ユーザの持ち時間の長さを推定するための情報を含み、
 前記付加情報生成部は、前記持ち時間の長さに応じて前記付加情報を生成する、前記(1)~(5)のいずれか1項に記載の情報処理装置。
(7)前記付加情報生成部は、前記持ち時間が長いほど前記付加情報の情報量を増加させる、前記(6)に記載の情報処理装置。
(8)前記付加情報生成部は、前記持ち時間が長い場合には前記ユーザが内容を選択可能な前記付加情報を生成する、前記(6)または(7)に記載の情報処理装置。
(9)前記ユーザ状況情報は、前記アクターが前記アクションを実行した第1の時刻と前記ユーザが前記アクション画像を視聴する第2の時刻との関係を示す情報を含み、
 前記付加情報生成部は、前記第1の時刻と前記第2の時刻との差が閾値を下回る場合に前記時間が短いと判定する、前記(6)~(8)のいずれか1項に記載の情報処理装置。
(10)前記ユーザ状況情報は、前記ユーザの前記アクションについての造詣の深さを推定するための情報を含み、
 前記付加情報生成部は、前記造詣の深さに応じて前記付加情報を生成する、前記(1)~(9)のいずれか1項に記載の情報処理装置。
(11)前記ユーザ状況情報は、前記ユーザの数を示す情報を含み、
 前記付加情報生成部は、前記ユーザの数に応じて前記付加情報を生成する、前記(1)~(10)のいずれか1項に記載の情報処理装置。
(12)前記アクション評価情報は、前記アクション画像に基づいて推定された、前記アクターの姿勢を評価した情報を含む、前記(1)~(11)のいずれか1項に記載の情報処理装置。
(13)前記アクション評価情報は、前記アクション画像に基づいて推定された、前記アクターと他のオブジェクトとのコンタクト位置を評価した情報を含む、前記(1)~(12)のいずれか1項に記載の情報処理装置。
(14)前記アクション画像に基づいて前記アクション評価情報を生成する評価情報生成部をさらに備える、前記(1)~(13)のいずれか1項に記載の情報処理装置。
(15)アクターのアクションを含むアクション画像に基づいて前記アクションを評価したアクション評価情報を取得することと、
 前記アクション画像を視聴するユーザの状況を示すユーザ状況情報を取得することと、
 前記アクション画像とともに前記ユーザに提示され、前記アクション評価情報の少なくとも一部を表現する付加情報を、前記ユーザ状況情報に応じて生成することと
 を含む情報処理方法。
(16)アクターのアクションを含むアクション画像に基づいて前記アクションを評価したアクション評価情報を取得する機能と、
 前記アクション画像を視聴するユーザの状況を示すユーザ状況情報を取得する機能と、
 前記アクション画像とともに前記ユーザに提示され、前記アクション評価情報の少なくとも一部を表現する付加情報を、前記ユーザ状況情報に応じて生成する機能と
 をコンピュータに実現させるためのプログラム。
The following configurations also belong to the technical scope of the present disclosure.
(1) An evaluation information acquisition unit that acquires action evaluation information obtained by evaluating the action based on an action image including an action of an actor;
A situation information acquisition unit that acquires user situation information indicating a situation of a user who views the action image;
An information processing apparatus comprising: an additional information generation unit that generates additional information that is presented to the user together with the action image and expresses at least part of the action evaluation information according to the user situation information.
(2) The information processing apparatus according to (1), wherein the additional information includes visual information.
(3) The information processing apparatus according to (2), wherein the additional information includes text or graphics that are superimposed on the action image or displayed in the vicinity of the action image.
(4) The action evaluation information includes information indicating coordinates of a body part of the actor in the action image,
The information processing apparatus according to (3), wherein the additional information includes a graphic representing a locus of the portion.
(5) The action evaluation information includes information for determining or scoring the action,
The information processing apparatus according to (3) or (4), wherein the additional information includes text or graphics indicating a progress or result of the determination or scoring.
(6) The user status information includes information for estimating the length of time the user has for referring to the additional information,
6. The information processing apparatus according to any one of (1) to (5), wherein the additional information generation unit generates the additional information according to the length of time.
(7) The information processing apparatus according to (6), wherein the additional information generation unit increases the information amount of the additional information as the holding time increases.
(8) The information processing apparatus according to (6) or (7), wherein the additional information generation unit generates the additional information from which the user can select contents when the holding time is long.
(9) The user status information includes information indicating a relationship between a first time when the actor performs the action and a second time when the user views the action image,
The additional information generation unit according to any one of (6) to (8), wherein the time is determined to be short when a difference between the first time and the second time is less than a threshold value. Information processing device.
(10) The user situation information includes information for estimating a depth of knowledge about the action of the user,
The information processing apparatus according to any one of (1) to (9), wherein the additional information generation unit generates the additional information according to a depth of the sculpting.
(11) The user status information includes information indicating the number of users,
The information processing apparatus according to any one of (1) to (10), wherein the additional information generation unit generates the additional information according to the number of users.
(12) The information processing apparatus according to any one of (1) to (11), wherein the action evaluation information includes information obtained by evaluating the posture of the actor estimated based on the action image.
(13) In any one of (1) to (12), the action evaluation information includes information obtained by evaluating a contact position between the actor and another object estimated based on the action image. The information processing apparatus described.
(14) The information processing apparatus according to any one of (1) to (13), further including an evaluation information generation unit that generates the action evaluation information based on the action image.
(15) obtaining action evaluation information obtained by evaluating the action based on an action image including an actor's action;
Obtaining user situation information indicating a situation of a user viewing the action image;
Generating additional information that is presented to the user together with the action image and expresses at least a part of the action evaluation information according to the user situation information.
(16) a function of acquiring action evaluation information obtained by evaluating the action based on an action image including an action of an actor;
A function of acquiring user situation information indicating a situation of a user who views the action image;
A program for causing a computer to realize a function of generating, according to the user situation information, additional information that is presented to the user together with the action image and expresses at least a part of the action evaluation information.
 10  システム
 100  カメラ
 200  サーバ
 300  出力装置
 300a  テレビ
 300b  レコーダ
 201  画像取得部
 203  評価情報生成部
 205  評価情報取得部
 207  状況情報取得部
 209  付加情報生成部
 211  画面生成部
DESCRIPTION OF SYMBOLS 10 System 100 Camera 200 Server 300 Output device 300a Television 300b Recorder 201 Image acquisition part 203 Evaluation information generation part 205 Evaluation information acquisition part 207 Status information acquisition part 209 Additional information generation part 211 Screen generation part

Claims (16)

  1.  アクターのアクションを含むアクション画像に基づいて前記アクションを評価したアクション評価情報を取得する評価情報取得部と、
     前記アクション画像を視聴するユーザの状況を示すユーザ状況情報を取得する状況情報取得部と、
     前記アクション画像とともに前記ユーザに提示され、前記アクション評価情報の少なくとも一部を表現する付加情報を、前記ユーザ状況情報に応じて生成する付加情報生成部と
     を備える情報処理装置。
    An evaluation information acquisition unit for acquiring action evaluation information for evaluating the action based on an action image including an action of an actor;
    A situation information acquisition unit that acquires user situation information indicating a situation of a user who views the action image;
    An information processing apparatus comprising: an additional information generation unit that generates additional information that is presented to the user together with the action image and expresses at least part of the action evaluation information according to the user situation information.
  2.  前記付加情報は、視覚的な情報を含む、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the additional information includes visual information.
  3.  前記付加情報は、前記アクション画像に重畳される、または前記アクション画像の近傍に表示されるテキストまたはグラフィックを含む、請求項2に記載の情報処理装置。 The information processing apparatus according to claim 2, wherein the additional information includes a text or a graphic superimposed on the action image or displayed in the vicinity of the action image.
  4.  前記アクション評価情報は、前記アクション画像内での前記アクターの体の部分の座標を示す情報を含み、
     前記付加情報は、前記部分の軌跡を表現するグラフィックを含む、請求項3に記載の情報処理装置。
    The action evaluation information includes information indicating the coordinates of the body part of the actor in the action image,
    The information processing apparatus according to claim 3, wherein the additional information includes a graphic representing a locus of the part.
  5.  前記アクション評価情報は、前記アクションを判定または採点した情報を含み、
     前記付加情報は、前記判定または採点の経過または結果を示すテキストまたはグラフィックを含む、請求項3に記載の情報処理装置。
    The action evaluation information includes information for determining or scoring the action,
    The information processing apparatus according to claim 3, wherein the additional information includes text or a graphic indicating a progress or result of the determination or scoring.
  6.  前記ユーザ状況情報は、前記付加情報を参照するための前記ユーザの持ち時間の長さを推定するための情報を含み、
     前記付加情報生成部は、前記持ち時間の長さに応じて前記付加情報を生成する、請求項1に記載の情報処理装置。
    The user status information includes information for estimating the length of time the user has for referring to the additional information,
    The information processing apparatus according to claim 1, wherein the additional information generation unit generates the additional information in accordance with the length of the possession time.
  7.  前記付加情報生成部は、前記持ち時間が長いほど前記付加情報の情報量を増加させる、請求項6に記載の情報処理装置。 The information processing apparatus according to claim 6, wherein the additional information generation unit increases the information amount of the additional information as the holding time increases.
  8.  前記付加情報生成部は、前記持ち時間が長い場合には前記ユーザが内容を選択可能な前記付加情報を生成する、請求項6に記載の情報処理装置。 The information processing apparatus according to claim 6, wherein the additional information generation unit generates the additional information from which the user can select contents when the holding time is long.
  9.  前記ユーザ状況情報は、前記アクターが前記アクションを実行した第1の時刻と前記ユーザが前記アクション画像を視聴する第2の時刻との関係を示す情報を含み、
     前記付加情報生成部は、前記第1の時刻と前記第2の時刻との差が閾値を下回る場合に前記時間が短いと判定する、請求項6に記載の情報処理装置。
    The user status information includes information indicating a relationship between a first time when the actor performs the action and a second time when the user views the action image,
    The information processing apparatus according to claim 6, wherein the additional information generation unit determines that the time is short when a difference between the first time and the second time is less than a threshold value.
  10.  前記ユーザ状況情報は、前記ユーザの前記アクションについての造詣の深さを推定するための情報を含み、
     前記付加情報生成部は、前記造詣の深さに応じて前記付加情報を生成する、請求項1に記載の情報処理装置。
    The user situation information includes information for estimating a depth of knowledge about the action of the user,
    The information processing apparatus according to claim 1, wherein the additional information generation unit generates the additional information according to a depth of the sculpting.
  11.  前記ユーザ状況情報は、前記ユーザの数を示す情報を含み、
     前記付加情報生成部は、前記ユーザの数に応じて前記付加情報を生成する、請求項1に記載の情報処理装置。
    The user status information includes information indicating the number of users,
    The information processing apparatus according to claim 1, wherein the additional information generation unit generates the additional information according to the number of users.
  12.  前記アクション評価情報は、前記アクション画像に基づいて推定された、前記アクターの姿勢を評価した情報を含む、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the action evaluation information includes information obtained by evaluating the posture of the actor estimated based on the action image.
  13.  前記アクション評価情報は、前記アクション画像に基づいて推定された、前記アクターと他のオブジェクトとのコンタクト位置を評価した情報を含む、請求項1に記載の情報処理装置。 2. The information processing apparatus according to claim 1, wherein the action evaluation information includes information obtained by evaluating a contact position between the actor and another object estimated based on the action image.
  14.  前記アクション画像に基づいて前記アクション評価情報を生成する評価情報生成部をさらに備える、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, further comprising an evaluation information generation unit that generates the action evaluation information based on the action image.
  15.  アクターのアクションを含むアクション画像に基づいて前記アクションを評価したアクション評価情報を取得することと、
     前記アクション画像を視聴するユーザの状況を示すユーザ状況情報を取得することと、
     前記アクション画像とともに前記ユーザに提示され、前記アクション評価情報の少なくとも一部を表現する付加情報を、前記ユーザ状況情報に応じて生成することと
     を含む情報処理方法。
    Obtaining action evaluation information evaluating the action based on an action image including an actor's action;
    Obtaining user situation information indicating a situation of a user viewing the action image;
    Generating additional information that is presented to the user together with the action image and expresses at least a part of the action evaluation information according to the user situation information.
  16.  アクターのアクションを含むアクション画像に基づいて前記アクションを評価したアクション評価情報を取得する機能と、
     前記アクション画像を視聴するユーザの状況を示すユーザ状況情報を取得する機能と、
     前記アクション画像とともに前記ユーザに提示され、前記アクション評価情報の少なくとも一部を表現する付加情報を、前記ユーザ状況情報に応じて生成する機能と
     をコンピュータに実現させるためのプログラム。
    A function of acquiring action evaluation information obtained by evaluating the action based on an action image including an actor's action;
    A function of acquiring user situation information indicating a situation of a user who views the action image;
    A program for causing a computer to realize a function of generating, according to the user situation information, additional information that is presented to the user together with the action image and expresses at least a part of the action evaluation information.
PCT/JP2015/076633 2014-12-08 2015-09-18 Information processing device, information processing method, and program WO2016092933A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2014-248046 2014-12-08
JP2014248046 2014-12-08

Publications (1)

Publication Number Publication Date
WO2016092933A1 true WO2016092933A1 (en) 2016-06-16

Family

ID=56107128

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2015/076633 WO2016092933A1 (en) 2014-12-08 2015-09-18 Information processing device, information processing method, and program

Country Status (1)

Country Link
WO (1) WO2016092933A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018069985A1 (en) * 2016-10-11 2018-04-19 富士通株式会社 Scoring support program, scoring support apparatus, and scoring support method
WO2019146024A1 (en) * 2018-01-24 2019-08-01 富士通株式会社 Screen generation program, screen generation method and information processing device
JP2021078783A (en) * 2019-11-19 2021-05-27 富士通株式会社 Screen output program, screen output method, information processor, screen generation program, screen generation method, and monitoring system
WO2021171470A1 (en) * 2020-02-27 2021-09-02 富士通株式会社 Information processing program, device, and method
WO2021186645A1 (en) * 2020-03-18 2021-09-23 富士通株式会社 Information processing program, device, and method
WO2022215116A1 (en) * 2021-04-05 2022-10-13 富士通株式会社 Information processing program, device, and method

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001299975A (en) * 2000-04-27 2001-10-30 Hiromi Hamabe Bodily sensing device and bodily sensing system
JP2008170685A (en) * 2007-01-11 2008-07-24 Yamaha Corp Voice evaluation device and karaoke device
WO2014162788A1 (en) * 2013-04-02 2014-10-09 Necソリューションイノベータ株式会社 Facial-expression assessment device, dance assessment device, karaoke device, and game device

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001299975A (en) * 2000-04-27 2001-10-30 Hiromi Hamabe Bodily sensing device and bodily sensing system
JP2008170685A (en) * 2007-01-11 2008-07-24 Yamaha Corp Voice evaluation device and karaoke device
WO2014162788A1 (en) * 2013-04-02 2014-10-09 Necソリューションイノベータ株式会社 Facial-expression assessment device, dance assessment device, karaoke device, and game device

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPWO2018069985A1 (en) * 2016-10-11 2019-04-25 富士通株式会社 Scoring support program, scoring support device and scoring support method
US10936860B2 (en) 2016-10-11 2021-03-02 Fujitsu Limited Marking assist device, and marking assist method
WO2018069985A1 (en) * 2016-10-11 2018-04-19 富士通株式会社 Scoring support program, scoring support apparatus, and scoring support method
US11446561B2 (en) 2018-01-24 2022-09-20 Fujitsu Limited Non-transitory computer readable recording medium, screen generating method, and information processing apparatus
WO2019146024A1 (en) * 2018-01-24 2019-08-01 富士通株式会社 Screen generation program, screen generation method and information processing device
CN111867687A (en) * 2018-01-24 2020-10-30 富士通株式会社 Screen generation program, screen generation method, and information processing device
JPWO2019146024A1 (en) * 2018-01-24 2021-01-28 富士通株式会社 Screen generation program, screen generation method and information processing device
CN111867687B (en) * 2018-01-24 2021-10-22 富士通株式会社 Storage medium, screen generating method, and information processing apparatus
JP2021078783A (en) * 2019-11-19 2021-05-27 富士通株式会社 Screen output program, screen output method, information processor, screen generation program, screen generation method, and monitoring system
JP7363403B2 (en) 2019-11-19 2023-10-18 富士通株式会社 Screen output program, screen output method, information processing device, screen generation program, screen generation method, and monitoring system
EP4093023A4 (en) * 2020-02-27 2023-03-01 Fujitsu Limited Information processing program, device, and method
WO2021171470A1 (en) * 2020-02-27 2021-09-02 富士通株式会社 Information processing program, device, and method
JP7400937B2 (en) 2020-02-27 2023-12-19 富士通株式会社 Information processing program, device, and method
WO2021186645A1 (en) * 2020-03-18 2021-09-23 富士通株式会社 Information processing program, device, and method
EP4093024A4 (en) * 2020-03-18 2023-03-01 Fujitsu Limited Information processing program, device, and method
JP7444238B2 (en) 2020-03-18 2024-03-06 富士通株式会社 Information processing program, device, and method
WO2022215116A1 (en) * 2021-04-05 2022-10-13 富士通株式会社 Information processing program, device, and method
EP4300948A4 (en) * 2021-04-05 2024-05-22 Fujitsu Ltd Information processing program, device, and method

Similar Documents

Publication Publication Date Title
WO2016092933A1 (en) Information processing device, information processing method, and program
US11247114B2 (en) Information processing device, storage medium, and information processing method
JP5994306B2 (en) Information processing apparatus, information processing system, and program
JP6610689B2 (en) Information processing apparatus, information processing method, and recording medium
US8885979B2 (en) Apparatus and associated methodology for analyzing subject motion in images
US10873777B2 (en) Information processing device and information processing method to calculate score for evaluation of action
EP3632512B1 (en) Information processing device, information processing method, and program
WO2016111069A1 (en) Information processing device, information processing method, and program
EP2985059B1 (en) Apparatus and method of user interaction
EP3186599B1 (en) Feedback provision system
WO2018203453A1 (en) Information processing device, information processing method, and program
US9004998B2 (en) Social interaction during online gaming
CN110753267B (en) Display control method and device and display
WO2019187493A1 (en) Information processing device, information processing method, and program
US11606608B1 (en) Gamification of video content presented to a user
US20180261123A1 (en) Method for educating sports fans
JP2017022727A (en) Information processing device, information processing system, and program
JP2020095699A (en) Visually impaired person-purpose information presentation system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15867467

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

NENP Non-entry into the national phase

Ref country code: JP

122 Ep: pct application non-entry in european phase

Ref document number: 15867467

Country of ref document: EP

Kind code of ref document: A1