WO2022102550A1 - Information processing device and information processing method - Google Patents

Information processing device and information processing method Download PDF

Info

Publication number
WO2022102550A1
WO2022102550A1 PCT/JP2021/040879 JP2021040879W WO2022102550A1 WO 2022102550 A1 WO2022102550 A1 WO 2022102550A1 JP 2021040879 W JP2021040879 W JP 2021040879W WO 2022102550 A1 WO2022102550 A1 WO 2022102550A1
Authority
WO
WIPO (PCT)
Prior art keywords
information
user
information processing
event
unit
Prior art date
Application number
PCT/JP2021/040879
Other languages
French (fr)
Japanese (ja)
Inventor
茜 近藤
賢次 杉原
広 岩瀬
文彦 飯田
Original Assignee
ソニーグループ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニーグループ株式会社 filed Critical ソニーグループ株式会社
Publication of WO2022102550A1 publication Critical patent/WO2022102550A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis

Definitions

  • This disclosure relates to an information processing device and an information processing method.
  • the excitement of live performances and sporting events and the effects of staging are subjective, and there is no effective means to quantitatively measure the effects, and the actual situation is that the effects are evaluated based on the subjectivity and experience of the person in charge. ..
  • the production of the event is in line with a static scenario, and it is practically difficult to produce a production that immediately responds to the internal state of the audience.
  • Patent Document 1 the degree of excitement of the customer is estimated in real time by the seat sensor installed in the audience seat and the wearable sensor worn by the audience, and the staging content is provided to the audience at an effective timing according to the degree of excitement. Is disclosed.
  • Patent Document 1 only the degree of excitement is estimated, and it is not possible to infer the internal state of each spectator or to produce an effect according to the attributes and tastes of the spectators.
  • Patent Document 1 is intended only for the audience who actually visited the event venue, and does not provide any effect to the audience who participates in the event from the remote environment via the network. Not taken into account. Recently, the number of users who participate in an event in a remote environment is increasing, and the importance of event production in consideration of users who participate in an event in a remote environment is increasing.
  • the present disclosure provides an information processing device and an information processing method capable of grasping the internal state of a user participating in an event and reflecting it in the production of the event.
  • a feature amount extraction unit that extracts feature amounts based on the user's sensing information
  • a first estimation unit that estimates at least one of the user's attributes and behavior based on the feature amount
  • a clustering unit that classifies the user or a group consisting of a plurality of users into a plurality of clusters based on the estimation by the first estimation unit.
  • An information processing apparatus including an information processing unit that performs predetermined information processing based on at least one of estimation by the first estimation unit and classification by the clustering unit is provided.
  • the sensing information includes an image captured by the image pickup device.
  • the clustering unit may be classified into the plurality of clusters based on the analysis result of the captured image.
  • the captured image includes the image of the user, and the captured image includes the image of the user.
  • the feature amount extraction unit may extract the feature amount including at least one of the user's face, posture, body movement, and skeletal information.
  • the feature amount extraction unit may extract the feature amount based on at least one of acoustic data, object recognition, and frequency analysis information.
  • the first estimation unit may estimate at least one of the attributes and actions of the user who participates in the event based on the feature amount and the progress information of the event.
  • a tagging unit for adding tag information may be provided in units of the user or the group.
  • the information processing unit may provide information based on the tag information to the user or the group to which the same tag information is given.
  • the information processing unit may provide and exchange information according to at least one of the attributes and actions of the user.
  • a situation image generation unit that generates a situation image in which an identifier indicating the internal state of the user, which is determined based on the sensing information, is added to the image taken by the user.
  • the situation image generation unit may generate the situation image including the progress information of the event in which the user participates and the information regarding at least one of the degree of excitement and the degree of concentration of the user.
  • the first estimation unit estimates an internal state including at least one of the degree of excitement and the degree of concentration of the user based on the sensing information.
  • the clustering unit may be classified into the plurality of clusters based on the feature amount and the internal state.
  • the clustering unit may be classified into the plurality of clusters based on the change in the internal state according to the progress information of the event.
  • sensing information acquisition unit that acquires the sensing information regarding at least one of the user at the event venue and the user who participates in the event from the remote environment.
  • the clustering unit may be classified into the plurality of clusters in units of users who participate in the event from a remote environment or a group consisting of a plurality of users.
  • a second estimation unit that estimates an internal state including at least one of the degree of excitement and concentration of users who participate in the event from a remote environment based on the sensing information.
  • a display control unit that adjusts the size of the display area for displaying information about the event based on the internal state estimated by the second estimation unit may be provided.
  • the display control unit is one with the audience seats of the venue of the event according to the increase in at least one of the degree of excitement and the degree of concentration of the user on the display unit viewed by the user who participates in the event from the remote environment.
  • An image that enhances the experience may be displayed.
  • the display control unit can provide an information providing image and a visual sense according to the predetermined condition within a range visible to the user. At least one of the effect images may be displayed.
  • the visual effect image may be a virtual person image of another user who participates in the venue of the event and whose internal state satisfies the predetermined condition.
  • a user who participates in the event from a remote environment is provided with an information exchange unit that exchanges information with the other person corresponding to the virtual person image via the virtual person image when the predetermined condition is satisfied. May be good.
  • a step of extracting a feature amount based on a user's sensing information and A step of estimating at least one of the user's attributes and behavior based on the feature amount, and Based on the estimation, the step of classifying into a plurality of clusters in units of the user or a group consisting of a plurality of users, and An information processing method comprising a step of performing predetermined information processing based on at least one of the estimation and the plurality of clusters.
  • the block diagram which shows the schematic structure of the information processing system provided with the information processing apparatus by one Embodiment of this disclosure.
  • the flowchart which shows the processing operation of the information processing apparatus by this embodiment.
  • the figure which shows the type of "meaningful attribute" while a game is interrupted.
  • FIG. 3 is a diagram showing a processing operation for automatically verifying tag information of each spectator generated by executing the fan tagging program of FIG. 4.
  • the figure which shows the video data including a correct answer tag The figure which illustrated the main application example of the information processing apparatus by this embodiment.
  • the figure following FIG. 17G. The figure which showed the similarity between various sports and basketball.
  • FIG. 1 is a block diagram showing a schematic configuration of an information processing system 2 provided with an information processing device 1 according to an embodiment of the present disclosure.
  • the information processing system 2 of FIG. 1 performs various information processing based on video data obtained by photographing the spectators of the stadium 3 with the camera 4 installed on the court (stadium) 3 where the sporting event is held.
  • a sporting event an example of watching a basketball game will be mainly described, but the type of sport does not matter.
  • the present embodiment can be widely applied to various events other than sports (for example, live music performance, event of entertainment, etc.).
  • the event is not limited to the event held at a specific venue such as a stadium, and may be an event delivered by live distribution described later. The event may be such that a user who participates in the event can participate in the event from the remote environment of the venue and the live distribution destination.
  • the information processing system 2 of FIG. 1 includes a plurality of cameras 4 fixed by a jig 4a at positions where the spectators of the court (stadium) 3 can be photographed, a network device 5, a processing server 6, and a database (hereinafter, DB). It is equipped with a server 7 (abbreviated as). A device other than that shown in FIG. 1 may be connected to the information processing system 2 according to the present embodiment. In addition, it is assumed that there is a separate camera for shooting the game to be held on the court (stadium) 3.
  • the network device 5 controls to transmit the video data taken by the plurality of cameras 4 to the processing server 6 via the network.
  • the network may be a public line such as the Internet or a dedicated line. Further, the network may be either wireless or wired.
  • the processing server 6 receives video data captured by a plurality of cameras 4 via the network device 5 and performs various information processing. For example, the processing server 6 performs distortion correction, color correction, camera 4 control processing for normalizing a plurality of video data shot by the plurality of cameras 4 for a plurality of video data shot by the plurality of cameras 4. After that, various information processing is performed.
  • the DB server 7 stores game progress information of the sporting event being held, stats information such as the results of participating athletes, and video data processed by the processing server 6.
  • the processing server 6 and the DB server 7 may be integrated into one server, or at least one of the processing server 6 and the DB server 7 may be divided into two or more servers.
  • the information processing system 2 of FIG. 1 may include a distribution server 9 that distributes information to a mobile terminal 8 or the like possessed by the spectators of the stadium 3.
  • the distribution server 9 controls the distribution information for the spectators transmitted from the processing server 6 to be transmitted to the corresponding spectator's mobile terminal 8 or the like.
  • the spectator's mobile terminal 8 is, for example, a smartphone, a wearable device such as a watch, a penlight possessed by the spectator for supporting an event, or the like.
  • the information processing apparatus 1 includes at least a processing server 6, and may also include a DB server 7, a distribution server 9, and the like.
  • Audiences watching sporting events do not always watch the game on the court (stadium) 3, but may watch it through TV, PC, mobile terminal 8 at home, etc., or in public viewing.
  • wireless networks that wirelessly communicate large volumes of data at high speed and at low cost will rapidly become widespread, and it is expected that the number of spectators watching games outside the stadium 3 will increase.
  • participating in an event other than the event venue such as stadium 3 is referred to as event participation in a remote environment or online watching (participation).
  • the information processing system 2 of FIG. 1 may include an operator server (not shown) for distributing various information to the event operator's PC or the like.
  • the operator server In the case of a sporting event, the operator server generates video data in which an identifier is added to the video data of the spectators so that the degree of excitement and concentration of the spectators can be grasped, and displays the video data on the operator tool screen described later. ..
  • the processing server 6 and the distribution server 9 may have the function of the operator server.
  • FIG. 2 is a block diagram showing a schematic configuration of an information processing system 2 provided with an information processing device 1 corresponding to event participation in a remote environment.
  • the information processing system 2 of FIG. 2 has a system configuration in which an spectator watches a sporting event at a TV 10a or PC 10b or a public viewing venue 10c at a place other than the court (stadium) 3 (for example, at home). Shows. It is assumed that the TV 10a and the PC 10b for watching the sporting event are equipped with a camera 4 for photographing the spectators. Further, it is assumed that the camera 4 for photographing the spectators watching the game in the public viewing is installed in the public viewing venue 10c.
  • the information processing system 2 of FIG. 2 includes a processing server 6, a DB server 7, and a distribution server 9 that acquire video data from the above-mentioned camera 4 and perform various information processing.
  • the basic processing of these servers is the same as that of each server shown in FIG.
  • the processing server 6 and the distribution server 9 can be devised in various ways to give the spectators watching the game in the remote environment the same sense of presence as the spectators watching the game in the stadium 3. .. A specific example thereof will be described later.
  • the processing server 6 of FIG. 1 may be integrated with the processing server 6 of FIG. 2, and similarly, the DB server 7 of FIG. 1 may be integrated with the DB server 7 of FIG. 2, and similarly, the distribution of FIG. 1 may be integrated.
  • the server 9 may be integrated with the distribution server 9 of FIG. Further, at least two or more of the processing server 6, the DB server 7, and the distribution server 9 of FIGS. 1 and 2 may be integrated, or conversely, various information processing may be performed by distributing them to more servers. It may be distributed and executed. That is, the server configurations of FIGS. 1 and 2 are only examples.
  • FIG. 3 is a functional block diagram of the information processing apparatus 1 according to the present embodiment.
  • FIG. 3 mainly blocks the functions of the processing server 6, the DB server 7, and the distribution server 9 of FIG. 1 or 2.
  • the information processing device 1 in FIG. 3 includes a sensing information acquisition unit 11, a feature amount extraction unit 12, a first estimation unit 13, a clustering unit 14, and an information processing unit 15.
  • the sensing information acquisition unit 11 acquires sensing information.
  • Sensing information is detection information detected by various sensors.
  • a typical example of sensing information is an image taken by an image pickup device.
  • the sensing information may be video data captured by an image sensor. It should be noted that the captured image of the image pickup device and the video data obtained by the image sensor are not necessarily essential sensing information.
  • the sensing information may include acoustic data of an event venue such as a stadium 3. Acoustic data is useful for determining the degree of excitement of an event.
  • the sensing information may include the detection information of the vibration sensor installed in the audience seats of the event venue such as the stadium 3.
  • the sensing information may include detection information such as an acceleration sensor or a gyro sensor built in the mobile terminal 8 possessed by the spectator or the penlight for cheering. As described above, as long as the sensing information can be used to determine the internal state such as the degree of excitement and the degree of concentration of the audience, the specific type thereof does not matter.
  • the sensing information acquisition unit 11 acquires sensing information periodically, irregularly, or continuously from the start of the match to the end of the match.
  • the feature amount extraction unit 12 extracts the feature amount based on the sensing information sensed by the user. For example, the feature amount extraction unit 12 extracts a feature amount including at least one of the face, posture, body movement, and skeletal information (also referred to as bone data) of a user (for example, an event participant).
  • the sensing information includes video data
  • the feature amount extraction unit 12 analyzes the video data to identify each spectator spectating at the stadium 3, the degree of smile of each identified spectator, and the wrist.
  • the feature amount can be extracted according to the amount of movement, the amount of movement of the head, the degree of opening of the eyes, the direction of the line of sight, and the like. Further, the feature amount extraction unit 12 may extract the feature amount based on at least one of the acoustic data, the object recognition, and the frequency analysis information.
  • the first estimation unit 13 estimates at least one of the user's attributes and actions based on the feature amount extracted by the feature amount extraction unit 12.
  • the specific estimation contents of the first estimation unit 13 will be described later, but in the case of a sporting event, for example, the degree of excitement and concentration of each spectator in the game are estimated.
  • the clustering unit 14 classifies the clusters into a plurality of clusters based on the estimation by the first estimation unit 13 with the user or a group consisting of a plurality of users as a unit.
  • the type of cluster is arbitrary. For example, in the case of a sporting event, attention may be paid to the cheering method, and spectators who have the same cheering method may be assigned to the same cluster. Alternatively, each spectator may be classified into one of a plurality of clusters according to the degree of interest in the game, focusing on the degree of interest in the game.
  • the information processing unit 15 performs predetermined information processing based on at least one of the estimation by the first estimation unit 13 and the classification by the clustering unit 14.
  • the specific content of information processing performed by the information processing unit 15 is arbitrary.
  • the information processing unit 15 may perform information processing for providing information suitable for each spectator to the clustered spectators.
  • the information processing unit 15 may provide information to the spectators who are watching the game in the remote environment to control the degree of presence according to the degree of excitement of the spectators.
  • the information processing apparatus 1 may include an event information acquisition unit 16.
  • the event information acquisition unit 16 acquires progress information of an event in which the user participates. For example, in the case of a sporting event, the event information acquisition unit 16 acquires progress information such as who scored a score on which team several minutes after the start of the match.
  • the first estimation unit 13 may estimate at least one of the attributes and actions of the user who participates in the event based on the feature amount and the progress information of the event.
  • the information processing apparatus 1 may include a tagging unit 17.
  • the tagging unit 17 adds tag information in units of users or groups participating in the event based on the estimation by the first estimation unit 13.
  • the tag information may be information that identifies a home fan who supports the home team, an away fan who supports the away team, and a beginner who watches a game.
  • the tagging unit 17 tags the spectators who are excited when the home team scores as home fans, and the spectators who are excited when the away team scores points are away fans. Audiences who do not get excited no matter which team scores the score may be tagged as a beginner.
  • the information processing unit 15 may provide information based on the tag information to the user or group to which the same tag information is given. Further, the information processing unit 15 may provide at least one of information provision and information exchange according to at least one of the user's attributes and actions.
  • Information provision is, for example, the provision of information related to an event and which the user may be interested in.
  • Information exchange is, for example, conversation with other spectators participating in the event.
  • the function of the information processing unit 15 can be built in, for example, the distribution server 9 of FIG.
  • the distribution server 9 transmits various information related to the tag information to the tagged spectator's mobile terminal 8 or the like based on the instruction from the processing server 6. As an example, spectators tagged as beginners may be provided with information about the rules of the match, and spectators who are rooting for the team that scored may be provided with stats information for the players who scored. ..
  • the information processing apparatus 1 may include a situation image generation unit 18.
  • the situation image generation unit 18 generates a situation image in which an identifier indicating the internal state of the user is added to the image of the user in the video data obtained by shooting the venue of the event.
  • the situation image is used as an image for the operator for the operator to confirm the event situation.
  • the situation image generation unit 18 may generate a situation image (operator image) including information on the progress of the event and information on at least one of the degree of excitement and the degree of concentration of the user.
  • a specific example of the situation image (image for the operator) will be described later.
  • the operator who directs the event can confirm whether or not the effect produced the intended effect by the situation image (image for the operator).
  • the internal state of the user is, for example, the degree of excitement and the degree of concentration of the user.
  • the identifier indicating the internal state is, for example, a circle having a diameter size according to the degree of excitement of the user, and this circle may be superimposed on the face image of the user.
  • the information processing apparatus 1 may include a second estimation unit 19 and a display control unit 20.
  • the second estimation unit 19 estimates the internal state including at least one of the degree of excitement and the degree of concentration of the users who participate in the event from the remote environment based on the sensing information.
  • the display control unit 20 adjusts the size of the display area for displaying information about the event based on the internal state estimated by the second estimation unit 19. Further, the display control unit 20 may display an image in which the degree of similarity with the actual venue of the event is adjusted according to the internal state on the display unit viewed by the user who participates in the event from the remote environment. Further, when the user who participates in the event from the remote environment satisfies the predetermined condition, the display control unit 20 has at least one of the information providing image and the visual effect image according to the predetermined condition within the range visible to the user. May be displayed.
  • the visual effect image may include a virtual person image of another user who is participating in the venue of the event and whose internal state satisfies a predetermined condition.
  • the information processing apparatus 1 may include an information exchange unit 21.
  • the information exchange unit 21 exchanges information with another person corresponding to the avatar via the avatar.
  • FIG. 4 is a block diagram showing a software configuration of the information processing apparatus 1 according to the present embodiment.
  • the software configuration of FIG. 4 is executed by the processing server 6, the DB server 7, and the distribution server 9.
  • the software configuration of the information processing apparatus 1 includes a real-time processing 31, a prediction processing 32, and a data distribution processing 33.
  • the real-time processing 31 and the prediction processing 32 are mainly executed by the processing server 6 and the DB server 7.
  • the data distribution process 33 is mainly executed by the distribution server 9.
  • the real-time processing 31 performs processing for extracting features in real time based on sensing information (for example, video data).
  • the real-time processing 31 includes a face recognition execution program 31a, a bone data extraction program 31b, an acoustic data extraction program 31c, a match live data extraction program 31d, and an extraction result storage process 31e.
  • the face recognition execution program 31a performs face recognition processing based on the video data, and estimates the degree of smile of each spectator, the amount of movement of the wrist, the amount of movement of the head, the degree of opening of the eyes, the direction of the line of sight, and the like.
  • the bone data extraction program 31b extracts the bone data of each customer based on the video data.
  • the acoustic data extraction program 31c extracts the acoustic data of the event venue.
  • the extracted acoustic data contains, for example, information about the direction and volume of sound.
  • the game live data extraction program 31d acquires the game live data, for example, when the game live data is distributed on the official site of the host team.
  • the extraction result storage process 31e is extracted by the output data of the face recognition execution program 31a, the data extracted by the bone data extraction program 31b, the data extracted by the acoustic data extraction program 31c, and the match live data extraction program 31d. Control is performed to store the collected data in the DB server 7 as a feature amount.
  • the prediction process 32 performs a process of assigning tag information to each spectator, a process of classifying each spectator into a plurality of clusters, and a process of determining the behavior of each spectator. More specifically, the prediction process 32 includes a prediction data creation program 32a, a fan tagging program 32b, a fan clustering program 32c, a fan behavior determination program 32d, and a prediction result storage process 32e.
  • the prediction data creation program 32a is a program for creating a fan tagging program 32b, a fan clustering program 32c, and a fan behavior determination program 32d.
  • the fan tagging program 32b adds tag information to users who participate in the event (fans in the case of sports events) based on the feature amount obtained by the real-time processing 31.
  • the tag information is, for example, a home fan, an away fan, a beginner, and the like.
  • the fan clustering program 32c classifies users (fans, etc.) who participate in the event into a plurality of clusters based on the feature amount obtained by the real-time processing 31. As described above, a plurality of clusters may be classified into a plurality of clusters according to the degree of excitement of the game, or may be classified into a plurality of clusters according to the degree of concentration in the game, or a combination of several conditions. It may be classified into a plurality of clusters.
  • the fan behavior determination program 32d determines the behavior of users (fans, etc.) who participate in the event. For example, it is determined that an enthusiastic fan who is staring at the game, a fan who eats and drinks or has a conversation without watching the game, a fan who seems to be bored by looking only at a smartphone, and the like.
  • the prediction result storage process 32e performs a process of storing the prediction result obtained by executing the fan tagging program 32b, the fan clustering program 32c, and the fan behavior determination program 32d in a predetermined database.
  • the data distribution process 33 distributes information for distribution to the audience and the operator to the distribution server 9 and the operator server (tool selection server) 30 based on the prediction result data obtained by the prediction process 32. Generate. Information for distribution for spectators is transmitted from the distribution server 9 to the corresponding spectator's mobile terminal 8. Information for distribution for the operator is transmitted from the operator server 30 to the operator's PC or the like. The video data of the spectators' seats and the video data of the match stored in the cloud storage 22 are input to the operator server 30.
  • FIG. 5 is a flowchart showing the processing operation of the information processing apparatus 1 according to the present embodiment. This flow chart is performed regularly, irregularly, or continuously during the duration of the event.
  • user information such as the names of users participating in the event is acquired (step S1).
  • this registration information may be acquired.
  • the user information may be acquired including the seat information of the spectator seats. If the user's seat information can be acquired, it is possible to grasp who is sitting in which seat and improve the reliability of information distribution.
  • the sensing information is acquired, and the feature amount of the user is extracted based on the sensing information (step S2).
  • the sensing information includes video data of the audience seat
  • the face recognition execution program 31a and the bone data extraction program 31b described above are executed to include the facial expression, posture, body movement, skeletal information, etc. of the user.
  • Feature quantity can be extracted.
  • the feature amount including the movement amount of the user can be extracted by the optical flow.
  • the sensing information includes acoustic data
  • the feature amount including the user's voice can be extracted.
  • step S3 acquire the event environment information and content information.
  • the game progress information may be acquired by the game live data extraction program 31d shown in FIG.
  • the stats information of the players participating in the match and the players of the team to which they belong may be acquired from the DB server 7 or the like.
  • environmental information such as the number of participants in the event, the venue, the weather, the temperature, the season, and the time of the event may be acquired.
  • attribute determination and action determination are performed for each user or group participating in the event, and it is determined whether or not the clustering process has been normally performed based on the determination result (step S4).
  • the attribute determination and the behavior determination of each user are performed using the result of learning with a teacher using an existing database for the feature amount extracted based on the video data.
  • a clustering process is performed in which a plurality of participants having similar attributes are grouped and classified into a plurality of clusters.
  • Step S5 If the clustering process cannot be performed normally, such as when learning is not performed correctly, or if the reliability is low even if it can be performed, the analysis results are registered in the data retention DB without classifying into multiple clusters. (Step S5).
  • step S6 If the attribute determination and the action determination are performed in step S4 and the clustering process can be performed normally, the classification results for a plurality of clusters are registered in the result DB (step S6).
  • step S7 output processing is performed based on the processing results of steps S5 and S6 (step S7). If the clustering process can be performed normally, the information corresponding to each cluster is provided to the users classified into each cluster, and the classification result of a plurality of clusters is provided to the operator. Alternatively, the entire or part of the venue of the event may be directed to liven up the event. As described above, the specific content of the output processing is arbitrary.
  • FIG. 6A is a diagram showing an example of bone data 34 extracted by the bone data extraction program 31b.
  • FIG. 6A shows the bone data 34 superimposed on the video data obtained by photographing the audience seats of the stadium 3.
  • the bone data 34 is composed of a circle and a polygonal line.
  • the circle indicates the position of the spectator's face, and the polygonal line connected to the circle indicates the movement of the body. Audiences who stand up and cheer for them have a long line. Further, whether or not the person is applauding can also be detected by the bone data 34. For example, an spectator whose polygonal line becomes longer at the timing of increasing the score can be presumed to be a fan supporting the team that increased the score.
  • FIG. 6B is a diagram showing an example of the result of attribute determination. As mentioned above, it is possible to determine whether the player is a home fan, an away fan, or a beginner based on whether or not the body moves a lot when the score is increased.
  • FIG. 6B shows an area 35b in the home area 35a where home fans are gathered, an area 35c where away fans are gathered, and an area 35d where beginners are gathered by analyzing video data obtained by photographing the audience seats. Is shown as an example of automatic determination.
  • the information processing apparatus 1 prepares "meaningful attributes" in advance according to the situation of the event in order to estimate the internal state of the user participating in the event and determine the attributes.
  • the information processing apparatus 1 performs a process of applying to any of the "meaningful attributes" prepared in advance based on the result of analyzing the video data. This makes it possible to easily determine the attributes of the user.
  • FIG. 7A, 7B and 7C are diagrams showing an example of a type of "meaningful attribute” for estimating the internal state of the spectator of a sporting event and determining the attribute.
  • FIG. 7A shows the types of "meaningful attributes” during a match (also called during a period)
  • FIG. 7B shows the "meaning” during a match being interrupted (also called a non-period).
  • the type of "certain attribute” is shown
  • FIG. 7C shows the type of "meaningful attribute” for all time zones during the match.
  • the attributes for enlivening the match include the reaction to the match development and whether or not the match is being watched.
  • the reaction to the game development can be judged by the degree of smile, the amount of wrist movement, and the degree of eye opening. Whether or not you are watching the game can be judged by the degree of smile and whether or not you are watching the court (stadium) 3.
  • the attributes of other candidates include the presence or absence of a reaction (unintentional reaction), the intentional reaction, the reaction before and after the score, the audio volume of the entire event venue, and commentary. ..
  • the degree of concentration during play can be determined by whether or not the player is looking at the court (stadium) 3, and the degree of eye opening can be used to determine whether or not the player is staring at the court.
  • the intentional reaction can be determined to be intentionally moving the wrist based on the amount of movement of the wrist, and can be determined to be applauding based on the amount of movement of the optical flow in the area under the face.
  • the amount of sound in the entire venue can be judged from the size of the acoustic data. Also, depending on how you open your mouth, you can judge whether you are explaining the game or listening to the explanation.
  • the degree of participation in the spectator involvement type performance and the degree of interest in the non-spectator involvement type performance are the attributes for the responsiveness to the performance.
  • the degree of participation in the audience involvement type performance the degree of participation in the giveaway can be determined by looking at the court (stadium) 3 where the performance is performed, and the degree of participation in the audience involvement type cheerleader can be determined by the optical flow.
  • the degree of non-audience involvement type performance participation is the degree of cheerleader interest, staff dance interest, free throw interest, commemorative photo interest, and control interest, depending on the average degree of eye opening and watching court (stadium) 3. Can be judged.
  • the degree of interest of DJ can be determined from the amount of movement of the face, the amount of movement of the skeleton, and the amount of movement of the optical flow.
  • the attribute of whether or not the person is away from the seat can be determined depending on the presence or absence of data.
  • the attribute of whether or not both hands are raised can be determined by whether or not the wrist is above the face, and the attribute of whether or not the arms are crossed is determined by whether or not the X direction of the wrist is crossed. It can be judged, and the attribute of whether or not the face is touched can be judged from the distance between the face and the wrist.
  • the attribute of actions that may be related to basketball such as whether or not you are watching a smartphone.
  • the attribute of calmness can be determined by whether or not the body is moving regardless of the game development.
  • the attributes of facial expressions can be determined by the degree of smile.
  • the attribute of whether or not you are looking at the pamphlet the attribute of whether or not you have many conversations
  • the attribute of whether or not you are taking pictures match photography, selfie, directing photography
  • the court stadium
  • an operator tool screen is prepared for the operator to objectively verify the effect of the effect performed by the operator.
  • the above-mentioned situation image (image for the operator) is displayed on the tool screen for the operator.
  • FIG. 8 is a diagram showing an example of the operator tool screen 40.
  • the operator tool screen 40 of FIG. 8 can be displayed on a PC or the like owned by the operator.
  • the operator can verify in detail the reaction of the event participants to the production performed during the event by the situation image (image for the operator) displayed on the operator tool screen 40.
  • the operator tool screen 40 in FIG. 8 shows an example in which a performance or the like is produced during a basketball game.
  • the operator tool screen 40 of FIG. 8 displays the first area 40a for displaying the video data of the game, the second area 40b for displaying the reaction of the spectators, the score change of the game, the degree of excitement of the spectators, and the degree of concentration. It has a third area 40c and the like.
  • the images displayed in the second area 40b and the third area 40c are collectively referred to as a situation image (operator image).
  • FIG. 9A is an enlarged view of the image displayed in the second area 40b of FIG.
  • a circle (identifier) 41 indicating the degree of excitement is superimposed on the face of the audience. The greater the degree of excitement, the larger the radius of the circle 41.
  • a plurality of buttons 42 for selecting the range of spectator seats to be displayed in the second area 40b are arranged.
  • the range of Tokyo bench is selected.
  • FIG. 9B is an enlarged view of a part of the image displayed in the third area 40c of FIG.
  • the gray range 43 of FIG. 9B is the match duration, and the black range 44 between the grays is the match interruption period.
  • the horizontal axis is time, and the vertical solid line shows the scores of the opposing teams.
  • two polygonal lines are shown, one of which shows the degree of excitement of the audience and the other of which shows the degree of concentration of the audience.
  • the numbers in the rectangular frame 45 indicate the score and the score difference of each team at that time.
  • the operator should examine in detail how the degree of excitement and concentration of the spectators changed with the progress of the game. Can be done. At a sporting event, the home fan and the away fan show different reactions depending on the scoring situation, but on the operator tool screen 40 of FIG. 8, the range in which the home fan and the away fan are sitting can be examined in detail separately.
  • FIG. 10 is a diagram showing an example of a processing operation for generating the operator tool screen 40.
  • the processing operation of FIG. 10 may be performed by the processing server 6 or the operator server 30.
  • the process of generating the operator tool screen 40 is referred to as an exciting visualization tool.
  • the excitement visualization tool has a match analysis process 46a, an excitement circle drawing program 46b, an analysis data processing program 46c, a web display program 46d, and CSS (Cascading Style Sheets) 46e.
  • the match analysis process 46a outputs information on the concentration, excitement, and fan attributes of each spectator as analysis items for each spectator based on the result of analyzing the video data.
  • the match analysis process 46a outputs information on the score, player substitution, match suspension period, performance content performed during the match interruption, and performance time as match analysis items.
  • the game analysis process 46a outputs information on the degree of concentration and the degree of excitement for each frame of the video data and for each fan.
  • the exciting circle drawing program 46b superimposes a circle proportional to the degree of excitement of each spectator on the face image of the spectator based on the video data of the spectator seat and the output data of the game analysis process 46a.
  • the analysis data processing program 46c extracts only the information required by the tool from the CSV data after analysis. Convert pickle data to json format after analysis.
  • the web display program 46d is based on the video image of the game, the image generated by the circle drawing program 46b, and the CSS46e that describes the layout definition of the tool, and the game is displayed in the first area 40a of the tool screen 40 for the operator.
  • the process of reproducing the video image is performed, and the images of the second area 40b and the third area 40c are generated.
  • the WebAP server 47 may perform the processing of the web display program 46d and the CSS46e.
  • FIG. 11 is a diagram showing details of the game analysis process 46a of FIG.
  • the game analysis process 46a includes a spectator video data processing program 48a, a game data processing program 48b, a graph display program 48c, an index evaluation program 48d, a simple CSV creation program 48e for designers, and a setting change program 48f.
  • Data related to the feature amount generated by the feature quantification process 48g is input to the audience video data processing program 48a.
  • the feature quantity processing 48g the feature quantity is extracted by executing the face recognition execution program 31a, the bone data extraction program 31b, the optical flow processing, and the like on the video of the spectator seat and the video of the game.
  • the detection data representing the feature quantity and the fan attribute information are output.
  • the attributes of the fans are determined from the behavior of the spectators and the development of the game.
  • the spectator video data processing program 48a performs personal identification, addition of analysis items, video data integration of a plurality of cameras 4, correction between cameras 4, etc. based on the video data of the spectator seats, and the degree of concentration and excitement of each spectator. , And fan attributes, etc. are output.
  • Match progress data is input to the match data processing program 48b via, for example, the Internet.
  • the match progress data includes information such as performance information performed during the match and video time of the match, in addition to the match progress information.
  • the match data processing program 48b performs match data processing, personal item integration, performance information processing, manual inconsistency detection, etc. based on the output data of the spectator video data processing program 48a and the match progress data, and each spectator. In addition to outputting the degree of concentration, the degree of excitement, and the fan attributes, information on points, player substitution, game suspension period, performance content and period is output as analysis items for the game.
  • the graph display program 48c generates a graph 49 showing the time change of the degree of concentration and the degree of excitement based on the data obtained by executing the game data processing program 48b.
  • FIG. 12 is a diagram showing an example of the graph 49 generated by the graph display program 48c, and is an enlarged view of the graph 49 shown in FIG.
  • the horizontal axis shows the elapsed time of the game, and the vertical axis shows the degree of excitement and concentration of the spectators.
  • FIG. 12 shows a graph of the degree of excitement and a graph of the degree of concentration of the audience.
  • the gray period 49a in FIG. 12 is the game continuation period, and the dark gray period 49b indicates the period in which some kind of production is performed. By presenting this graph to the operator, it is possible to grasp at a glance whether or not the audience was excited by the production.
  • FIG. 13 is a diagram showing an example of the index 50 generated by the index evaluation program 48d, and is an enlarged view of the one shown in FIG.
  • the correlation between the home team scoring 1 to 3 points, the correlation with the home team score, the away team scoring 1 to 3 points, and the away team For each of the correlations with the score of, the index of the total number of spectators whose degree of excitement increased, the index of the total number of spectators whose degree of excitement decreased, the index of the rate of increase, and the degree of excitement within a predetermined period. An index for the average value is shown.
  • the simple CSV creation program 48e for designers in FIG. 11 creates a CSV file containing information on the degree of concentration and excitement for each frame of video and for each fan based on the execution result of the game data processing program 48b.
  • the setting change program 48f in FIG. 11 switches the input / output folder for storing the files obtained by executing the game data processing program 48b, the graph display program 48c, the index evaluation program 48d, and the simple CSV creation program 48e for the designer. Further, the setting change program 48f changes the color and display position of the graph generated by the graph display program 48c. Further, the setting change program 48f switches the video frame for calculating the index generated by the index evaluation program 48d.
  • FIG. 14 is a diagram showing a processing operation for automatically verifying the tag information of each spectator generated by executing the fan tagging program 32b of FIG.
  • the process of automatically verifying tag information is referred to as a tag information automatic verification tool.
  • the processing of the tag information automatic verification tool may be performed by the processing server 6 of FIG. 1, the operator server 30, or the like.
  • the tag information automatic verification tool has a score information file generation program 51a, a bone data cleaning program 51b, and a fan tagging / evaluation program 51c.
  • the score information file generation program 51a corrects the time lag between the match video and the spectator video based on the data related to the feature obtained by the feature quantification process 48g, and the time file in the scoring team and the spectator video (hereinafter, score information). File) is generated.
  • the bone data cleaning program 51b extracts the bone coordinates of the spectator when the home team and the away team raise the score based on the score information file, and outputs the movement amount of the spectator at the time of scoring based on the bone coordinates.
  • the fan tagging / evaluation program 51c compares the changes in bone data with the average value of bone data at the time of scoring when the home team scores and when the away team scores, and tags the audience. Judgment of attachment is made. The determination is made, for example, by rule-based or deep learning (DNN: Deep Neural Network).
  • DNN Deep Neural Network
  • FIG. 15A shows video data 52a including tag information (also referred to as estimated tag) generated by the fan tagging / evaluation program 51c
  • FIG. 15B shows video data 52b including correct tag information (also referred to as correct tag).
  • tag information also referred to as estimated tag
  • correct tag information also referred to as correct tag
  • the result of comparing the correct answer tag and the estimated tag is presented in a table format as shown in FIG. 14, and the correct answer rate of the tag information is presented numerically.
  • FIG. 16 is a diagram illustrating a main application example of the information processing apparatus 1 according to the present embodiment.
  • the information processing apparatus 1 according to the present embodiment can be applied to applications other than those shown in FIG.
  • the information processing apparatus 1 may have a function of automatically archiving the timing at which an event rises. As a result, it is possible to easily acquire a video of the timing when the event is exciting.
  • the information processing apparatus 1 may have a function of performing an effect or performing a cleaning work in an area on the court (stadium) 3 that is not attracting attention.
  • the audience's line of sight can be directed to the area that is not attracting attention.
  • the cleaning work in an area that is not attracting attention the coat can be cleaned without bothering the eyes of the spectators.
  • the information processing device 1 may change the production according to the cheering style of the audience.
  • the type of production can be switched according to the degree of excitement and concentration of the audience. For example, for a spectator who is speaking out, the degree of excitement may be further increased by vibrating the spectator seats.
  • the information processing device 1 may provide advertising information or food and drink information when the concentration of the spectators to the game is reduced. For example, by providing advertising information and food and drink information in anticipation of a period during which the game is temporarily interrupted, it is possible to enhance the advertising effect and improve the sales of restaurants and the like in the stadium 3.
  • the information processing device 1 may have a function of distributing information on the goods of the attention player to the smartphone of the spectator or the like.
  • the information processing device 1 may have a function of giving benefits such as virtual currency and points to the spectator according to the support style of the spectator.
  • benefits such as virtual currency and points
  • information on the number of times the event has been attended may be acquired for each spectator, and benefits such as points may be given to the spectators who have participated frequently.
  • the information processing device 1 may have a function of visualizing the excitement of the entire event venue or between teams and feeding back to the organizer team. As a result, it is possible to increase the spirit of each player in the organizer team.
  • the information processing device 1 may have a function of delivering a photographed image of the spectator himself and a friend at the time of excitement to the corresponding spectator's smartphone.
  • the information processing apparatus 1 may have an attribute with less excitement and a function of delivering a predetermined sound source only to the audience in the area.
  • the information processing device 1 may have a function of creating an image of a virtual stadium 3 by a fake cloud that captures the characteristics of each attribute. For example, when the degree of excitement of the audience increases, a virtual person (avatar) is displayed so that the audience can cheer with the avatar, or when the degree of excitement of the audience increases further, a conversation with the avatar will occur. It may be provided with a function such as enabling it or enabling high five. Specific examples of this function will be described with reference to FIGS. 17D and 17E described later.
  • the information processing device 1 transmits the degree of excitement of the spectators who are watching the game online (watching the game in a remote environment) to the venue of the event, and the degree of excitement at the venue of the event is watched online. It may have a function of transmitting to a place. For example, when the total amount of excitement at the online watching place reaches the standard amount, the information that the standard amount has been reached is transmitted to the event venue via the network, and the effect sound and visual effect image are transmitted at the event venue. You may perform the production by such as. On the contrary, when the amount of sound or the like at the venue of the event reaches the reference amount, some visual effect may be produced at the online watching place.
  • the information processing apparatus 1 may have a function of matching spectators who are reacting at the same timing with a high synchronization rate.
  • the spectators are assumed to be far from each other, for example, matching a spectator at the event venue with a spectator watching online. For example, an avatar cheering at another place that matches the cheering style of the spectator watching online may be displayed so that conversation and high five can be performed. Specific examples of this function will also be described with reference to FIGS. 17D and 17E described later.
  • FIGS. 17A to 17H are diagrams showing an example of participating in an event in a remote environment, that is, watching an online game.
  • FIGS. 17A to 17H show a state in which the spectators are guided to the place where the online spectator is to be watched, the spectators are explained how to watch the spectators, and then the spectators are actually watched. It is assumed that this place has a function to display the video of the game on the wall instead of the spectator's home, and a function to display 3D video such as avatar using AR (Augmented Reality) technology. ..
  • AR Augmented Reality
  • a function to display the audience cheering at other places such as the event venue, virtual people, etc. in a part of the image displayed on the wall surface. May be provided.
  • 17A to 17F show an example of watching a soccer game online, but the type of sporting event to watch online does not matter.
  • FIG. 17A shows how the guide guides the spectators to the spectator place.
  • the spectators are guided to a spectator place called a space equipped with a multi-faceted projector system (hereinafter referred to as "warp square").
  • warp square a multi-faceted projector system
  • There is a sofa in Warp Square and the image of Stadium 3 is displayed on the wall in front of the sofa.
  • the display on the wall surface is performed by, for example, a projector.
  • a camera 4 (not shown) for capturing an image of the audience sitting on the sofa is provided, and the degree of excitement and concentration of the audience are analyzed based on the video data captured by the camera 4.
  • the guide gives the audience a megaphone and explains that they can use the megaphone to cheer loudly.
  • the spectators in Warp Square can cheer in the same cheering style as the spectators in Stadium 3.
  • the spectator sits on the couch in Warp Square and presses the execute button on the remote controller, or the user's actions (eg, sitting on the couch, looking at the wall, etc.).
  • the image of the stadium 3 is projected on the front wall surface by the automatic recognition behavior sensing.
  • a 3D image that makes the avatar happy may be displayed.
  • an example is shown in which an avatar makes a gesture for a high five via a megaphone.
  • an effect sound may be produced or a visual effect image may be displayed at the place where the high five is performed.
  • FIGS. 17C to 17F the same images as those in FIGS. 17C to 17F described above are displayed.
  • the method of displaying the image is not limited to that shown in FIGS. 17C to 17F.
  • the screen is divided into two in the vertical direction, the video of the game is displayed on the upper side, and the audience seats of the stadium 3 are on the lower side. You may display a panoramic image as if you were doing it.
  • the silhouette of the avatar of the spectator in the stadium 3 may be displayed and the voice of the avatar cheering may be heard as shown in the figure on the right side of FIG. 17G.
  • FIG. 18 is a diagram showing the similarity between various sports and basketball.
  • the horizontal axis of FIG. 18 indicates the degree to which the shooting conditions are different from those of basketball, and the right side indicates that the shooting conditions are more different.
  • the shooting conditions are indoor / outdoor, dark / bright, wide / narrow, and large / small number of people.
  • the vertical axis of FIG. 18 shows the degree to which the behaviors of the basketball and the spectators are different, and the lower side shows that the behaviors of the spectators are more different.
  • FIG. 18 when the shooting conditions and the behavior of the spectators are comprehensively considered, ice hockey is most similar to basketball, and golf is most different from basketball. Therefore, for golf, it may be necessary to drastically change the processes shown in FIGS. 5, 10, 11 and 14 described above. Further, for sports other than golf, it may be necessary to change the processing of FIGS. 5, 10, 11 and 14 as necessary.
  • FIG. 19 is a diagram showing the similarity between various non-sports events and basketball. Similar to FIG. 18, the horizontal axis of FIG. 19 indicates the degree to which the shooting conditions differ from that of basketball, and the vertical axis indicates the degree to which the behavior of the basketball and the spectator differ.
  • FIG. 19 when the shooting conditions and the behavior of the audience are comprehensively considered, the music (live) event is most similar to basketball, and watching a movie in a movie theater is the most different from basketball. Therefore, for movie theaters, it may be necessary to drastically change the processes shown in FIGS. 5, 10, 11, 14, and the like described above. Further, for events other than movie theaters, it may be necessary to change the processing of FIGS. 5, 10, 11 and 14 as necessary.
  • the present embodiment since the internal state of the user participating in the event is estimated based on the sensing information, at least one of the user's attributes and behavior can be estimated, and the user is classified into a plurality of clusters. be able to. Further, according to the present embodiment, it is possible to tag a user or a group of users based on the estimation result of the attributes and behaviors of the users participating in the event and the classification result of a plurality of clusters, and the user or the group of users can be tagged. Can provide information suitable for. As a result, for example, it is possible to provide information suitable for each of the fans of the home team, the fans of the away team, and the beginners, and it is possible to further improve the attractiveness of watching sports.
  • the effect of the effect that could be judged only subjectively in the past can be achieved. It can be analyzed and evaluated in detail and objectively, and can be useful for producing more attractive productions.
  • the production method can be changed according to the degree of excitement of the user in the remote environment to make the degree of excitement higher, or the effect as if at the event venue can be performed, or the user at the event venue can be displayed as an avatar.
  • At least a part of the information processing apparatus described in the above-described embodiment may be configured by hardware or software.
  • a program that realizes at least a part of the functions of the information processing apparatus may be stored in a recording medium such as a flexible disk or a CD-ROM, read by a computer, and executed.
  • the recording medium is not limited to a removable one such as a magnetic disk or an optical disk, and may be a fixed recording medium such as a hard disk device or a memory.
  • a program that realizes at least a part of the functions of the information processing device may be distributed via a communication line (including wireless communication) such as the Internet. Further, the program may be encrypted, modulated, compressed, and distributed via a wired line or a wireless line such as the Internet, or stored in a recording medium.
  • a communication line including wireless communication
  • the program may be encrypted, modulated, compressed, and distributed via a wired line or a wireless line such as the Internet, or stored in a recording medium.
  • the present technology can have the following configurations.
  • a feature amount extraction unit that extracts a feature amount based on the user's sensing information
  • a first estimation unit that estimates at least one of the user's attributes and behavior based on the feature amount
  • a clustering unit that classifies the user or a group consisting of a plurality of users into a plurality of clusters based on the estimation by the first estimation unit.
  • An information processing apparatus including an information processing unit that performs predetermined information processing based on at least one of estimation by the first estimation unit and classification by the clustering unit.
  • the sensing information includes an image captured by the image pickup device. The information processing apparatus according to (1), wherein the clustering unit is classified into the plurality of clusters based on the analysis result of the captured image.
  • the captured image includes the image of the user.
  • the information processing apparatus according to (2), wherein the feature amount extraction unit extracts the feature amount including at least one of the user's face, posture, body movement, and skeletal information.
  • Processing device Further provided with an event information acquisition unit for acquiring progress information of the event in which the user participates.
  • the first estimation unit estimates at least one of the attributes and behaviors of the user participating in the event based on the feature amount and the progress information of the event, any one of (1) to (4).
  • the information processing device. (10) The situation image generation unit, according to (9), which generates the situation image including the progress information of the event in which the user participates and the information about at least one of the degree of excitement and the degree of concentration of the user.
  • the first estimation unit estimates an internal state including at least one of the degree of excitement and the degree of concentration of the user based on the sensing information.
  • the information processing apparatus according to any one of (1) to (10), wherein the clustering unit is classified into the plurality of clusters based on the feature amount and the internal state. (12) The information processing apparatus according to (11), wherein the clustering unit is classified into the plurality of clusters based on the change in the internal state according to the progress information of the event. (13) Any one of (1) to (12) further comprising a sensing information acquisition unit that acquires the sensing information regarding at least one of the user at the event venue and the user participating in the event from the remote environment. The information processing device described in the section. (14) The clustering unit classifies the event into the plurality of clusters in units of users who participate in the event from a remote environment or a group consisting of a plurality of users.
  • a second estimation unit that estimates an internal state including at least one of the degree of excitement and the degree of concentration of users who participate in the event from a remote environment based on the sensing information.
  • the display control unit is a seat of the audience at the venue of the event in response to an increase in at least one of the degree of excitement and the degree of concentration of the user on the display unit viewed by a user who participates in the event from a remote environment.
  • the information processing apparatus according to (15), which displays an image that enhances a sense of unity with.
  • the display control unit provides information according to the predetermined condition within a range visible to the user.
  • the information processing apparatus according to any one of (15) to (16), which displays at least one of an image and a visual effect image.
  • the visual effect image is a virtual person image of another user who participates in the venue of the event and whose internal state satisfies the predetermined condition.
  • An information exchange unit in which a user who participates in the event from a remote environment exchanges information with the other person corresponding to the virtual person image via the virtual person image when the predetermined condition is satisfied.
  • 1 Information processing device 2 Information processing system, 3 Court (stadium), 4 Camera, 5 Network equipment, 6 Processing server, 7 DB server, 8 Mobile terminal, 9 Distribution server, 11 Sensing information acquisition unit, 12 Feature quantity extraction unit , 13 1st estimation unit, 14 clustering unit, 15 information processing unit, 16 event information acquisition unit, 17 tagging unit, 18 situation image generation unit, 19 second estimation unit, 20 display control unit, 21 information exchange unit, 22 Cloud storage, 30 operator server (server for tool selection)

Abstract

[Problem] To ascertain the internal state of a participant of an event, and reflect the internal state to the production of the event. [Solution] This information processing device comprises: a feature amount extraction unit which extracts a feature amount on the basis of user sensing information; a first estimation unit which estimates at least one among the attribute and behavior of the user; a clustering unit which performs classification into a plurality of clusters in units of the user or a group composed of a plurality of users on the basis of the estimation by the first estimation unit; and an information processing unit which performs a prescribed information process on the basis of at least one among the estimation by the first estimation unit and the classification by the clustering unit.

Description

情報処理装置及び情報処理方法Information processing equipment and information processing method
 本開示は、情報処理装置及び情報処理方法に関する。 This disclosure relates to an information processing device and an information processing method.
 ライブやスポーツイベントの盛り上がりや演出の効果は主観的なものであり、その効果を定量的に測定する有効な手段はなく、担当者の主観や経験により効果を評価しているのが実情である。また、イベントの演出は、静的なシナリオに沿ったものであり、観客の内的状態に即応した演出を行うのは現実的には困難である。 The excitement of live performances and sporting events and the effects of staging are subjective, and there is no effective means to quantitatively measure the effects, and the actual situation is that the effects are evaluated based on the subjectivity and experience of the person in charge. .. In addition, the production of the event is in line with a static scenario, and it is practically difficult to produce a production that immediately responds to the internal state of the audience.
 特許文献1には、観客席に設置されたシートセンサと観客が装着するウェアラブルセンサにて顧客の盛り上がり度合いをリアルタイムに推定し、盛り上がり度合いに応じて演出コンテンツを効果的なタイミングで観客に提供することが開示されている。 In Patent Document 1, the degree of excitement of the customer is estimated in real time by the seat sensor installed in the audience seat and the wearable sensor worn by the audience, and the staging content is provided to the audience at an effective timing according to the degree of excitement. Is disclosed.
特開2019-144882号公報Japanese Unexamined Patent Publication No. 2019-144882
 しかしながら、特許文献1では、盛り上がり度合いを推定するだけであり、個々の観客の内的状態を推測したり、観客の属性や趣向に合わせた演出を行うことはできない。 However, in Patent Document 1, only the degree of excitement is estimated, and it is not possible to infer the internal state of each spectator or to produce an effect according to the attributes and tastes of the spectators.
 また、特許文献1は、実際にイベント会場に足を運んだ観客に対する演出だけを念頭に置いており、リモート環境からネットワーク経由でイベントに参加する観客に対して、何らかの演出を提供することは全く考慮に入れていない。最近は、リモート環境でイベントに参加するユーザが増えており、リモート環境でイベントに参加するユーザに配慮したイベント演出の重要性が高まっている。 In addition, Patent Document 1 is intended only for the audience who actually visited the event venue, and does not provide any effect to the audience who participates in the event from the remote environment via the network. Not taken into account. Recently, the number of users who participate in an event in a remote environment is increasing, and the importance of event production in consideration of users who participate in an event in a remote environment is increasing.
 そこで、本開示では、イベントに参加するユーザの内的状態を把握して、イベントの演出に反映させることができる情報処理装置及び情報処理方法を提供するものである。 Therefore, the present disclosure provides an information processing device and an information processing method capable of grasping the internal state of a user participating in an event and reflecting it in the production of the event.
 上記の課題を解決するために、本開示の一態様によれば、
 ユーザのセンシング情報に基づいて特徴量を抽出する特徴量抽出部と、
 前記特徴量に基づいて、前記ユーザの属性及び行動の少なくとも一方を推定する第1推定部と、
 前記第1推定部による推定に基づいて、前記ユーザ又は複数のユーザからなるグループを単位として、複数のクラスタに分類するクラスタリング部と、
 前記第1推定部による推定と前記クラスタリング部による分類との少なくとも一方に基づいて、所定の情報処理を行う情報処理部と、を備える、情報処理装置が提供される。
In order to solve the above problems, according to one aspect of the present disclosure,
A feature amount extraction unit that extracts feature amounts based on the user's sensing information,
A first estimation unit that estimates at least one of the user's attributes and behavior based on the feature amount,
A clustering unit that classifies the user or a group consisting of a plurality of users into a plurality of clusters based on the estimation by the first estimation unit.
An information processing apparatus including an information processing unit that performs predetermined information processing based on at least one of estimation by the first estimation unit and classification by the clustering unit is provided.
 前記センシング情報は、撮像装置の撮像画像を含んでおり、
 前記クラスタリング部は、前記撮像画像の解析結果に基づいて前記複数のクラスタに分類してもよい。
The sensing information includes an image captured by the image pickup device.
The clustering unit may be classified into the plurality of clusters based on the analysis result of the captured image.
 前記撮像画像は、前記ユーザの画像を含んでおり、
 前記特徴量抽出部は、前記ユーザの顔、姿勢、体の動き、及び骨格情報の少なくとも一つを含む前記特徴量を抽出してもよい。
The captured image includes the image of the user, and the captured image includes the image of the user.
The feature amount extraction unit may extract the feature amount including at least one of the user's face, posture, body movement, and skeletal information.
 前記特徴量抽出部は、音響データ、物体認識、及び周波数解析情報の少なくとも一つに基づいて前記特徴量を抽出してもよい。 The feature amount extraction unit may extract the feature amount based on at least one of acoustic data, object recognition, and frequency analysis information.
 前記ユーザが参加するイベントの経過情報を取得するイベント情報取得部を備え、
 前記第1推定部は、前記特徴量と前記イベントの経過情報とに基づいて、前記イベントに参加するユーザの属性及び行動の少なくとも一方を推定してもよい。
It is provided with an event information acquisition unit that acquires progress information of the event in which the user participates.
The first estimation unit may estimate at least one of the attributes and actions of the user who participates in the event based on the feature amount and the progress information of the event.
 前記第1推定部による推定に基づいて、前記ユーザ又は前記グループを単位として、タグ情報を付与するタグ付け部を備えてもよい。 Based on the estimation by the first estimation unit, a tagging unit for adding tag information may be provided in units of the user or the group.
 前記情報処理部は、同一の前記タグ情報が付与された前記ユーザ又は前記グループに対して、前記タグ情報に基づいて情報を提供してもよい。 The information processing unit may provide information based on the tag information to the user or the group to which the same tag information is given.
 前記情報処理部は、前記ユーザの属性及び行動の少なくとも一方に応じた情報提供及び情報交換をの少なくとも一方を行ってもよい。 The information processing unit may provide and exchange information according to at least one of the attributes and actions of the user.
 前記ユーザを撮影した画像に、前記センシング情報に基づいて決定される前記ユーザの内的状態を示す識別子を付加した状況画像を生成する状況画像生成部を備えてもよい。 It may be provided with a situation image generation unit that generates a situation image in which an identifier indicating the internal state of the user, which is determined based on the sensing information, is added to the image taken by the user.
 前記状況画像生成部は、前記ユーザが参加するイベントの経過情報と、前記ユーザの盛り上がり度合い及び集中度の少なくとも一方に関する情報とを含む前記状況画像を生成してもよい。 The situation image generation unit may generate the situation image including the progress information of the event in which the user participates and the information regarding at least one of the degree of excitement and the degree of concentration of the user.
 前記第1推定部は、前記センシング情報に基づいて、前記ユーザの盛り上がり度合い及び集中度の少なくとも一方を含む内的状態を推定し、
 前記クラスタリング部は、前記特徴量及び前記内的状態に基づいて、前記複数のクラスタに分類してもよい。
The first estimation unit estimates an internal state including at least one of the degree of excitement and the degree of concentration of the user based on the sensing information.
The clustering unit may be classified into the plurality of clusters based on the feature amount and the internal state.
 前記クラスタリング部は、イベントの経過情報に応じた前記内的状態の変化に基づいて、前記複数のクラスタに分類してもよい。 The clustering unit may be classified into the plurality of clusters based on the change in the internal state according to the progress information of the event.
 イベントの会場にいるユーザと、リモート環境から前記イベントに参加するユーザとの少なくとも一方に関する前記センシング情報を取得するセンシング情報取得部を備えていてもよい。 It may be provided with a sensing information acquisition unit that acquires the sensing information regarding at least one of the user at the event venue and the user who participates in the event from the remote environment.
 前記クラスタリング部は、前記イベントにリモート環境から参加するユーザ又は複数のユーザからなるグループを単位として、前記複数のクラスタに分類してもよい。 The clustering unit may be classified into the plurality of clusters in units of users who participate in the event from a remote environment or a group consisting of a plurality of users.
 前記センシング情報に基づいて、前記イベントにリモート環境から参加するユーザの盛り上がり度合い及び集中度の少なくとも一つを含む内的状態を推定する第2推定部と、
 前記第2推定部で推定された前記内的状態に基づいて、前記イベントに関する情報を表示する表示領域のサイズを調整する表示制御部と、備えてもよい。
A second estimation unit that estimates an internal state including at least one of the degree of excitement and concentration of users who participate in the event from a remote environment based on the sensing information.
A display control unit that adjusts the size of the display area for displaying information about the event based on the internal state estimated by the second estimation unit may be provided.
 前記表示制御部は、前記イベントにリモート環境から参加するユーザが見る表示部に、前記ユーザの盛り上がり度合い及び集中度の少なくとも一方が高まることに応じて、前記イベントの開催場所の観客席との一体感を高める映像を表示させてもよい。 The display control unit is one with the audience seats of the venue of the event according to the increase in at least one of the degree of excitement and the degree of concentration of the user on the display unit viewed by the user who participates in the event from the remote environment. An image that enhances the experience may be displayed.
 前記表示制御部は、前記イベントにリモート環境から参加するユーザの内的状態が所定の条件を満たすときに、前記ユーザが視認可能な範囲内に、前記所定の条件に応じた情報提供画像及び視覚効果画像の少なくとも一方を表示させてもよい。 When the internal state of a user who participates in the event from a remote environment satisfies a predetermined condition, the display control unit can provide an information providing image and a visual sense according to the predetermined condition within a range visible to the user. At least one of the effect images may be displayed.
 前記視覚効果画像は、前記イベントの会場に参加しており、内的状態が前記所定の条件を満たす別のユーザの仮想人物画像であってもよい。 The visual effect image may be a virtual person image of another user who participates in the venue of the event and whose internal state satisfies the predetermined condition.
 前記イベントにリモート環境から参加するユーザは、前記所定の条件を満たすときに、前記仮想人物画像を介して、前記仮想人物画像に対応する前記別の人物と情報交換を行う情報交換部を備えてもよい。 A user who participates in the event from a remote environment is provided with an information exchange unit that exchanges information with the other person corresponding to the virtual person image via the virtual person image when the predetermined condition is satisfied. May be good.
 本開示の他の一態様によれば、ユーザのセンシング情報に基づいて特徴量を抽出するステップと、
 前記特徴量に基づいて、前記ユーザの属性及び行動の少なくとも一方を推定するステップと、
 前記推定に基づいて、前記ユーザ又は複数のユーザからなるグループを単位として、複数のクラスタに分類するステップと、
 前記推定と前記複数のクラスタとの少なくとも一方に基づいて、所定の情報処理を行うステップと、を備える、情報処理方法が提供される。
According to another aspect of the present disclosure, a step of extracting a feature amount based on a user's sensing information and
A step of estimating at least one of the user's attributes and behavior based on the feature amount, and
Based on the estimation, the step of classifying into a plurality of clusters in units of the user or a group consisting of a plurality of users, and
An information processing method is provided comprising a step of performing predetermined information processing based on at least one of the estimation and the plurality of clusters.
本開示の一実施形態による情報処理装置を備えた情報処理システムの概略構成を示すブロック図。The block diagram which shows the schematic structure of the information processing system provided with the information processing apparatus by one Embodiment of this disclosure. リモート環境でのイベント参加に対応した情報処理装置を備えた情報処理システムの概略構成を示すブロック図。A block diagram showing a schematic configuration of an information processing system equipped with an information processing device that supports event participation in a remote environment. 本実施形態による情報処理装置の機能ブロック図。The functional block diagram of the information processing apparatus by this embodiment. 本実施形態による情報処理装置のソフトウェア構成を示すブロック図。The block diagram which shows the software structure of the information processing apparatus by this embodiment. 本実施形態による情報処理装置の処理動作を示すフローチャート。The flowchart which shows the processing operation of the information processing apparatus by this embodiment. ボーンデータ抽出プログラムにて抽出されたボーンデータの一例を示す図。The figure which shows an example of the bone data extracted by a bone data extraction program. 属性判定を行った結果の一例を示す図。The figure which shows an example of the result of having performed the attribute judgment. 試合が行われている最中の「意味のある属性」の種別を示す図。A diagram showing the types of "meaningful attributes" during a match. 試合が中断している最中の「意味のある属性」の種別を示す図。The figure which shows the type of "meaningful attribute" while a game is interrupted. 試合中の全時間帯の「意味のある属性」の種別を示す図。A diagram showing the types of "meaningful attributes" for all time zones during a match. 運営者用ツール画面の一例を示す図。The figure which shows an example of the tool screen for an operator. 図8の第2エリアに表示される画像を拡大した図。The enlarged view of the image displayed in the 2nd area of FIG. 図8の第3エリア表示される画像の一部を拡大した図。The third area of FIG. 8 is an enlarged view of a part of the displayed image. 運営者用ツール画面を生成する処理動作の一例を示す図。The figure which shows an example of the processing operation which generates the tool screen for an operator. 図10の試合解析処理の詳細を示す図。The figure which shows the detail of the game analysis processing of FIG. グラフ表示プログラムが生成するグラフの一例を示す図。The figure which shows an example of the graph generated by the graph display program. 指標評価プログラムで生成された指標の一例を示す図。The figure which shows an example of the index generated by the index evaluation program. 図4のファンタグ付けプログラムの実行により生成された各観客のタグ情報を自動検証する処理動作を示す図。FIG. 3 is a diagram showing a processing operation for automatically verifying tag information of each spectator generated by executing the fan tagging program of FIG. 4. 推定タグを含む映像データを示す図。The figure which shows the video data including the estimation tag. 正解タグを含む映像データを示す図。The figure which shows the video data including a correct answer tag. 本実施形態による情報処理装置の主要な応用例を例示した図。The figure which illustrated the main application example of the information processing apparatus by this embodiment. オンライン観戦の一例を示す図。The figure which shows an example of watching an online game. 図17Aに続く図。The figure following FIG. 17A. 図17Bに続く図。The figure following FIG. 17B. 図17Cに続く図。The figure following FIG. 17C. 図17Dに続く図。The figure following FIG. 17D. 図17Eに続く図。The figure following FIG. 17E. 図17Fに続く図。The figure following FIG. 17F. 図17Gに続く図。The figure following FIG. 17G. 種々のスポーツとバスケットボールとの類似性を示した図。The figure which showed the similarity between various sports and basketball. スポーツ以外の種々のイベントとバスケットボールとの類似性を示した図。The figure which showed the similarity between various non-sports events and basketball.
 以下、図面を参照して、情報処理装置及び情報処理方法の実施形態について説明する。以下では、情報処理装置及び情報処理方法の主要な構成部分を中心に説明するが、情報処理装置及び情報処理方法には、図示又は説明されていない構成部分や機能が存在しうる。以下の説明は、図示又は説明されていない構成部分や機能を除外するものではない。 Hereinafter, embodiments of the information processing apparatus and the information processing method will be described with reference to the drawings. In the following, the main components of the information processing device and the information processing method will be mainly described, but the information processing device and the information processing method may have components and functions not shown or described. The following description does not exclude components or functions not shown or described.
 図1は本開示の一実施形態による情報処理装置1を備えた情報処理システム2の概略構成を示すブロック図である。図1の情報処理システム2は、スポーツイベントが開催されるコート(スタジアム)3に設置されたカメラ4でスタジアム3の観客を撮影した映像データに基づいて種々の情報処理を行うものである。以下では、スポーツイベントの一例として、バスケットボールの試合を観戦する例を主に説明するが、スポーツの種類は問わない。また、本実施形態は、スポーツ以外の種々のイベント(例えば、音楽のライブ演奏、催し物のイベントなど)に幅広く適用可能である。また、イベントはスタジアムなどの特定の会場で開催されるものに限定されず、後述するライブ配信で配信されるイベントであってもよい。イベントに参加するユーザが会場およびライブ配信先のリモート環境からイベントに参加可能なイベントであってもよい。 FIG. 1 is a block diagram showing a schematic configuration of an information processing system 2 provided with an information processing device 1 according to an embodiment of the present disclosure. The information processing system 2 of FIG. 1 performs various information processing based on video data obtained by photographing the spectators of the stadium 3 with the camera 4 installed on the court (stadium) 3 where the sporting event is held. In the following, as an example of a sporting event, an example of watching a basketball game will be mainly described, but the type of sport does not matter. Further, the present embodiment can be widely applied to various events other than sports (for example, live music performance, event of entertainment, etc.). Further, the event is not limited to the event held at a specific venue such as a stadium, and may be an event delivered by live distribution described later. The event may be such that a user who participates in the event can participate in the event from the remote environment of the venue and the live distribution destination.
 図1の情報処理システム2は、コート(スタジアム)3の観客を撮影可能な位置に治具4aで固定された複数のカメラ4と、ネットワーク機器5と、処理サーバ6と、データベース(以下、DBと略する)サーバ7と、を備えている。なお、本実施形態による情報処理システム2には、図1に示した以外の機器が接続されていてもよい。なお、コート(スタジアム)3で行われる試合を撮影するカメラは別にあるものとする。 The information processing system 2 of FIG. 1 includes a plurality of cameras 4 fixed by a jig 4a at positions where the spectators of the court (stadium) 3 can be photographed, a network device 5, a processing server 6, and a database (hereinafter, DB). It is equipped with a server 7 (abbreviated as). A device other than that shown in FIG. 1 may be connected to the information processing system 2 according to the present embodiment. In addition, it is assumed that there is a separate camera for shooting the game to be held on the court (stadium) 3.
 ネットワーク機器5は、複数のカメラ4で撮影された映像データをネットワーク経由で処理サーバ6に送信する制御を行う。ネットワークは、インターネット等の公衆回線でもよいし、専用回線でもよい。また、ネットワークは、無線と有線のどちらでもよい。 The network device 5 controls to transmit the video data taken by the plurality of cameras 4 to the processing server 6 via the network. The network may be a public line such as the Internet or a dedicated line. Further, the network may be either wireless or wired.
 処理サーバ6は、複数のカメラ4で撮影された映像データを、ネットワーク機器5を介して受信し、種々の情報処理を行う。例えば、処理サーバ6は、複数のカメラ4で撮影された複数の映像データに対して、歪み補正、色補正、複数のカメラ4で撮影された複数の映像データを正規化するカメラ4制御処理などを行った上で、種々の情報処理を行う。 The processing server 6 receives video data captured by a plurality of cameras 4 via the network device 5 and performs various information processing. For example, the processing server 6 performs distortion correction, color correction, camera 4 control processing for normalizing a plurality of video data shot by the plurality of cameras 4 for a plurality of video data shot by the plurality of cameras 4. After that, various information processing is performed.
 DBサーバ7は、開催中のスポーツイベントの試合経過情報や、出場選手の成績等のスタッツ情報などを格納する他、処理サーバ6が加工した映像データを格納する。処理サーバ6とDBサーバ7は一つのサーバに統合してもよいし、処理サーバ6とDBサーバ7の少なくとも一方を2以上のサーバに分割してもよい。 The DB server 7 stores game progress information of the sporting event being held, stats information such as the results of participating athletes, and video data processed by the processing server 6. The processing server 6 and the DB server 7 may be integrated into one server, or at least one of the processing server 6 and the DB server 7 may be divided into two or more servers.
 この他、図1の情報処理システム2は、スタジアム3の観客が所持する携帯端末8等に情報を配信する配信サーバ9を備えていてもよい。配信サーバ9は、処理サーバ6から送信された観客向けの配信情報を、対応する観客の携帯端末8等に送信する制御を行う。観客の携帯端末8は、例えば、スマートフォン、時計等のウェアラブルデバイス、イベント応援用に観客が所持するペンライトなどである。 In addition, the information processing system 2 of FIG. 1 may include a distribution server 9 that distributes information to a mobile terminal 8 or the like possessed by the spectators of the stadium 3. The distribution server 9 controls the distribution information for the spectators transmitted from the processing server 6 to be transmitted to the corresponding spectator's mobile terminal 8 or the like. The spectator's mobile terminal 8 is, for example, a smartphone, a wearable device such as a watch, a penlight possessed by the spectator for supporting an event, or the like.
 本実施形態による情報処理装置1は、処理サーバ6を少なくとも備えており、この他に、DBサーバ7や配信サーバ9などを備えていてもよい。 The information processing apparatus 1 according to the present embodiment includes at least a processing server 6, and may also include a DB server 7, a distribution server 9, and the like.
 スポーツイベントを観戦する観客は、必ずしも、コート(スタジアム)3で観戦するとは限らず、自宅等のTVやPC、携帯端末8を通して観戦したり、パブリックビューイングで観戦する場合がある。特に、今後は、大容量のデータを高速かつ低コストで無線通信する無線ネットワークが急速に普及するため、スタジアム3以外で観戦する観客が増えると予想されている。本明細書では、スタジアム3等のイベント会場以外でイベントに参加することをリモート環境でのイベント参加、又はオンライン観戦(参加)と呼ぶ。 Audiences watching sporting events do not always watch the game on the court (stadium) 3, but may watch it through TV, PC, mobile terminal 8 at home, etc., or in public viewing. In particular, in the future, wireless networks that wirelessly communicate large volumes of data at high speed and at low cost will rapidly become widespread, and it is expected that the number of spectators watching games outside the stadium 3 will increase. In this specification, participating in an event other than the event venue such as stadium 3 is referred to as event participation in a remote environment or online watching (participation).
 図1の情報処理システム2は、イベントの運営者のPC等に各種情報を配信するための不図示の運営者サーバを備えていてもよい。運営者サーバは、スポーツイベントの場合には、観客の映像データに、観客の盛り上がり度合いと集中度が把握できるような識別子を付加した映像データを生成して、後述する運営者ツール画面に表示する。なお、処理サーバ6や配信サーバ9が運営者サーバの機能を備えていてもよい。 The information processing system 2 of FIG. 1 may include an operator server (not shown) for distributing various information to the event operator's PC or the like. In the case of a sporting event, the operator server generates video data in which an identifier is added to the video data of the spectators so that the degree of excitement and concentration of the spectators can be grasped, and displays the video data on the operator tool screen described later. .. The processing server 6 and the distribution server 9 may have the function of the operator server.
 図2はリモート環境でのイベント参加に対応した情報処理装置1を備えた情報処理システム2の概略構成を示すブロック図である。図2の情報処理システム2は、観客がコート(スタジアム)3とは別の場所(例えば、自宅など)のTV10a又はPC10b、あるいはパブリックビューイング会場10cにてスポーツイベントを観戦する場合のシステム構成を示している。スポーツイベントを観戦するTV10aやPC10bには、観客を撮影するカメラ4が設置されているものとする。また、パブリックビューイング会場10cには、パブリックビューイングで観戦する観客を撮影するカメラ4が設置されているものとする。 FIG. 2 is a block diagram showing a schematic configuration of an information processing system 2 provided with an information processing device 1 corresponding to event participation in a remote environment. The information processing system 2 of FIG. 2 has a system configuration in which an spectator watches a sporting event at a TV 10a or PC 10b or a public viewing venue 10c at a place other than the court (stadium) 3 (for example, at home). Shows. It is assumed that the TV 10a and the PC 10b for watching the sporting event are equipped with a camera 4 for photographing the spectators. Further, it is assumed that the camera 4 for photographing the spectators watching the game in the public viewing is installed in the public viewing venue 10c.
 図2の情報処理システム2は、上述したカメラ4からの映像データを取得して種々の情報処理を行う処理サーバ6と、DBサーバ7と、配信サーバ9とを備えている。これらサーバの基本的な処理は、図1に示した各サーバと同様である。ただし、処理サーバ6と配信サーバ9は、リモート環境で観戦している観客に対して、スタジアム3で観戦している観客と同様の臨場感を与えるための種々の工夫を施すことも可能である。その具体例については後述する。 The information processing system 2 of FIG. 2 includes a processing server 6, a DB server 7, and a distribution server 9 that acquire video data from the above-mentioned camera 4 and perform various information processing. The basic processing of these servers is the same as that of each server shown in FIG. However, the processing server 6 and the distribution server 9 can be devised in various ways to give the spectators watching the game in the remote environment the same sense of presence as the spectators watching the game in the stadium 3. .. A specific example thereof will be described later.
 図1の処理サーバ6は図2の処理サーバ6と統合してもよく、同様に、図1のDBサーバ7は図2のDBサーバ7と統合してもよく、同様に、図1の配信サーバ9は図2の配信サーバ9と統合してもよい。また、図1や図2の処理サーバ6、DBサーバ7及び配信サーバ9の少なくとも2つ以上を統合してもよいし、その逆に、より多くのサーバに分散させて、種々の情報処理を分散して実行してもよい。すなわち、図1及び図2のサーバ構成は一例に過ぎない。 The processing server 6 of FIG. 1 may be integrated with the processing server 6 of FIG. 2, and similarly, the DB server 7 of FIG. 1 may be integrated with the DB server 7 of FIG. 2, and similarly, the distribution of FIG. 1 may be integrated. The server 9 may be integrated with the distribution server 9 of FIG. Further, at least two or more of the processing server 6, the DB server 7, and the distribution server 9 of FIGS. 1 and 2 may be integrated, or conversely, various information processing may be performed by distributing them to more servers. It may be distributed and executed. That is, the server configurations of FIGS. 1 and 2 are only examples.
 図3は本実施形態による情報処理装置1の機能ブロック図である。図3は主に、図1又は図2の処理サーバ6、DBサーバ7及び配信サーバ9が備える機能をブロック化したものである。 FIG. 3 is a functional block diagram of the information processing apparatus 1 according to the present embodiment. FIG. 3 mainly blocks the functions of the processing server 6, the DB server 7, and the distribution server 9 of FIG. 1 or 2.
 図3の情報処理装置1は、センシング情報取得部11と、特徴量抽出部12と、第1推定部13と、クラスタリング部14と、情報処理部15とを備えている。 The information processing device 1 in FIG. 3 includes a sensing information acquisition unit 11, a feature amount extraction unit 12, a first estimation unit 13, a clustering unit 14, and an information processing unit 15.
 センシング情報取得部11は、センシング情報を取得する。センシング情報とは、種々のセンサで検知された検知情報である。センシング情報の代表例は、撮像装置の撮影画像である。より具体的な一例としては、センシング情報は、イメージセンサで撮像された映像データであってもよい。なお、撮像装置の撮像画像やイメージセンサによる映像データは必ずしも必須のセンシング情報ではない。センシング情報は、スタジアム3等のイベント会場の音響データを含んでいてもよい。音響データは、イベントの盛り上がり度合いを判断するために有効である。センシング情報は、スタジアム3等のイベント会場の観客席に設置された振動センサの検知情報を含んでいてもよい。あるいは、センシング情報は、観客が所持する携帯端末8や応援用のペンライトに内蔵された加速度センサやジャイロセンサ等の検知情報を含んでいてもよい。このように、センシング情報は、観客の盛り上がり度合いや集中度等の内的状態を判断するのに用いることができれば、その具体的な種類は問わない。 The sensing information acquisition unit 11 acquires sensing information. Sensing information is detection information detected by various sensors. A typical example of sensing information is an image taken by an image pickup device. As a more specific example, the sensing information may be video data captured by an image sensor. It should be noted that the captured image of the image pickup device and the video data obtained by the image sensor are not necessarily essential sensing information. The sensing information may include acoustic data of an event venue such as a stadium 3. Acoustic data is useful for determining the degree of excitement of an event. The sensing information may include the detection information of the vibration sensor installed in the audience seats of the event venue such as the stadium 3. Alternatively, the sensing information may include detection information such as an acceleration sensor or a gyro sensor built in the mobile terminal 8 possessed by the spectator or the penlight for cheering. As described above, as long as the sensing information can be used to determine the internal state such as the degree of excitement and the degree of concentration of the audience, the specific type thereof does not matter.
 スポーツイベントは、試合の経過に従って、観客の盛り上がり度合いや集中度が変化する。このため、センシング情報取得部11は、試合開始から試合終了までの間、定期的又は不定期的に、あるいは継続的に、センシング情報を取得する。 For sporting events, the degree of excitement and concentration of the spectators changes as the game progresses. Therefore, the sensing information acquisition unit 11 acquires sensing information periodically, irregularly, or continuously from the start of the match to the end of the match.
 特徴量抽出部12は、ユーザをセンシングしたセンシング情報に基づいて特徴量を抽出する。例えば、特徴量抽出部12は、ユーザ(例えば、イベントの参加者)の顔、姿勢、体の動き、及び骨格情報(ボーンデータとも呼ぶ)の少なくとも一つを含む特徴量を抽出する。センシング情報が映像データを含んでいる場合、特徴量抽出部12は、映像データを解析して、スタジアム3で観戦している各観客を識別し、識別された各観客の笑顔の度合い、手首の移動量、頭の移動量、目の開き度合い、視線方向などにより、特徴量を抽出することができる。また、特徴量抽出部12は、音響データ、物体認識、及び周波数解析情報の少なくとも一つに基づいて、特徴量を抽出してもよい。 The feature amount extraction unit 12 extracts the feature amount based on the sensing information sensed by the user. For example, the feature amount extraction unit 12 extracts a feature amount including at least one of the face, posture, body movement, and skeletal information (also referred to as bone data) of a user (for example, an event participant). When the sensing information includes video data, the feature amount extraction unit 12 analyzes the video data to identify each spectator spectating at the stadium 3, the degree of smile of each identified spectator, and the wrist. The feature amount can be extracted according to the amount of movement, the amount of movement of the head, the degree of opening of the eyes, the direction of the line of sight, and the like. Further, the feature amount extraction unit 12 may extract the feature amount based on at least one of the acoustic data, the object recognition, and the frequency analysis information.
 第1推定部13は、特徴量抽出部12で抽出された特徴量に基づいて、ユーザの属性及び行動の少なくとも一方を推定する。第1推定部13の具体的な推定内容は後述するが、例えばスポーツイベントの場合、各観客の試合への盛り上がり度合いや集中度などを推定する。 The first estimation unit 13 estimates at least one of the user's attributes and actions based on the feature amount extracted by the feature amount extraction unit 12. The specific estimation contents of the first estimation unit 13 will be described later, but in the case of a sporting event, for example, the degree of excitement and concentration of each spectator in the game are estimated.
 クラスタリング部14は、第1推定部13による推定に基づいて、ユーザ又は複数のユーザからなるグループを単位として、複数のクラスタに分類する。クラスタの種類は任意である。例えば、スポーツイベントの場合、応援の仕方に着目して、同じような応援の仕方をしている観客を同一のクラスタに割り振ってもよい。あるいは、試合への関心の高さに着目して、試合への関心度別に各観客を複数のクラスタのいずれかに分類してもよい。 The clustering unit 14 classifies the clusters into a plurality of clusters based on the estimation by the first estimation unit 13 with the user or a group consisting of a plurality of users as a unit. The type of cluster is arbitrary. For example, in the case of a sporting event, attention may be paid to the cheering method, and spectators who have the same cheering method may be assigned to the same cluster. Alternatively, each spectator may be classified into one of a plurality of clusters according to the degree of interest in the game, focusing on the degree of interest in the game.
 情報処理部15は、第1推定部13による推定とクラスタリング部14による分類との少なくとも一方に基づいて、所定の情報処理を行う。情報処理部15が行う具体的な情報処理の中身は任意である。後述するように、情報処理部15は、クラスタリングされた観客に対して、個々の観客に見合った情報を提供するための情報処理を行ってもよい。あるいは、情報処理部15は、リモート環境で観戦している観客に対して、観客の盛り上がり度合いに応じて、臨場感の度合いを制御するための情報提供を行ってもよい。 The information processing unit 15 performs predetermined information processing based on at least one of the estimation by the first estimation unit 13 and the classification by the clustering unit 14. The specific content of information processing performed by the information processing unit 15 is arbitrary. As will be described later, the information processing unit 15 may perform information processing for providing information suitable for each spectator to the clustered spectators. Alternatively, the information processing unit 15 may provide information to the spectators who are watching the game in the remote environment to control the degree of presence according to the degree of excitement of the spectators.
 図3に示すように、情報処理装置1は、イベント情報取得部16を備えていてもよい。イベント情報取得部16は、ユーザが参加するイベントの経過情報を取得する。例えば、スポーツイベントの場合、イベント情報取得部16は、試合が始まってから何分後にどのチームの誰が得点を上げたか等の経過情報を取得する。第1推定部13は、特徴量とイベントの経過情報に基づいて、イベントに参加するユーザの属性及び行動の少なくとも一方を推定してもよい。 As shown in FIG. 3, the information processing apparatus 1 may include an event information acquisition unit 16. The event information acquisition unit 16 acquires progress information of an event in which the user participates. For example, in the case of a sporting event, the event information acquisition unit 16 acquires progress information such as who scored a score on which team several minutes after the start of the match. The first estimation unit 13 may estimate at least one of the attributes and actions of the user who participates in the event based on the feature amount and the progress information of the event.
 図3に示すように、情報処理装置1は、タグ付け部17を備えていてもよい。タグ付け部17は、第1推定部13による推定に基づいて、イベントに参加するユーザ又はグループを単位として、タグ情報を付与する。タグ情報とは、例えばスポーツイベントの場合には、ホームチームを応援するホームファン、アウェイチームを応援するアウェイファン、及び試合観戦の初心者を識別する情報であってもよい。タグ付け部17は、観客の映像データに基づいて、ホームチームが得点を上げたときに盛り上がっている観客をホームファンとしてタグ付け、アウェイチームが得点を上げたときに盛り上がっている観客をアウェイファンとしてタグ付け、どちらのチームが得点を上げても盛り上がらない観客を初心者としてタグづけてもよい。 As shown in FIG. 3, the information processing apparatus 1 may include a tagging unit 17. The tagging unit 17 adds tag information in units of users or groups participating in the event based on the estimation by the first estimation unit 13. For example, in the case of a sporting event, the tag information may be information that identifies a home fan who supports the home team, an away fan who supports the away team, and a beginner who watches a game. Based on the video data of the spectators, the tagging unit 17 tags the spectators who are excited when the home team scores as home fans, and the spectators who are excited when the away team scores points are away fans. Audiences who do not get excited no matter which team scores the score may be tagged as a beginner.
 情報処理部15は、同一のタグ情報が付与されたユーザ又はグループに対して、タグ情報に基づく情報を提供してもよい。また、情報処理部15は、ユーザの属性及び行動の少なくとも一方に応じた情報提供及び情報交換の少なくとも一方を行ってもよい。情報提供とは、例えば、イベントに関連し、かつユーザが関心を持つと思われる情報の提供である。情報交換とは、例えば、イベントに参加している他の観客との会話などである。情報処理部15の機能は、例えば図1の配信サーバ9に内蔵することができる。配信サーバ9は、処理サーバ6からの指示に基づいて、タグ情報に関連した種々の情報を、タグ付けされた観客の携帯端末8等に送信する。一例としては、初心者にタグ付けされた観客に、試合のルールに関する情報を提供したり、得点を上げたチームを応援している観客に、得点を上げた選手のスタッツ情報を提供してもよい。 The information processing unit 15 may provide information based on the tag information to the user or group to which the same tag information is given. Further, the information processing unit 15 may provide at least one of information provision and information exchange according to at least one of the user's attributes and actions. Information provision is, for example, the provision of information related to an event and which the user may be interested in. Information exchange is, for example, conversation with other spectators participating in the event. The function of the information processing unit 15 can be built in, for example, the distribution server 9 of FIG. The distribution server 9 transmits various information related to the tag information to the tagged spectator's mobile terminal 8 or the like based on the instruction from the processing server 6. As an example, spectators tagged as beginners may be provided with information about the rules of the match, and spectators who are rooting for the team that scored may be provided with stats information for the players who scored. ..
 図3に示すように、情報処理装置1は、状況画像生成部18を備えていてもよい。状況画像生成部18は、イベントの会場を撮影した映像データ中のユーザの画像に、ユーザの内的状態を示す識別子を付加した状況画像を生成する。イベントにおいては、状況画像は、運営者がイベント状況を確認するための運営者用画像として使用される。より詳細には、状況画像生成部18は、イベントの進行情報と、ユーザの盛り上がり度合い及び集中度の少なくとも一方に関する情報とを含む状況画像(運営者用画像)を生成してもよい。状況画像(運営者用画像)の具体例については後述する。イベントの演出を行う運営者は、状況画像(運営者用画像)により、演出が意図された通りの効果を生じさせたか否かを確認することができる。ユーザの内的状態とは、例えば、ユーザの盛り上がり度合いや集中度などである。内的状態を示す識別子とは、例えば、ユーザの盛り上がり度合いに応じた径サイズの円であり、この円はユーザの顔画像に重畳されてもよい。 As shown in FIG. 3, the information processing apparatus 1 may include a situation image generation unit 18. The situation image generation unit 18 generates a situation image in which an identifier indicating the internal state of the user is added to the image of the user in the video data obtained by shooting the venue of the event. In the event, the situation image is used as an image for the operator for the operator to confirm the event situation. More specifically, the situation image generation unit 18 may generate a situation image (operator image) including information on the progress of the event and information on at least one of the degree of excitement and the degree of concentration of the user. A specific example of the situation image (image for the operator) will be described later. The operator who directs the event can confirm whether or not the effect produced the intended effect by the situation image (image for the operator). The internal state of the user is, for example, the degree of excitement and the degree of concentration of the user. The identifier indicating the internal state is, for example, a circle having a diameter size according to the degree of excitement of the user, and this circle may be superimposed on the face image of the user.
 図3に示すように、情報処理装置1は、第2推定部19と表示制御部20を備えていてもよい。第2推定部19は、センシング情報に基づいて、イベントにリモート環境から参加するユーザの盛り上がり度合い及び集中度の少なくとも一つを含む内的状態を推定する。 As shown in FIG. 3, the information processing apparatus 1 may include a second estimation unit 19 and a display control unit 20. The second estimation unit 19 estimates the internal state including at least one of the degree of excitement and the degree of concentration of the users who participate in the event from the remote environment based on the sensing information.
 表示制御部20は、第2推定部19で推定された内的状態に基づいて、イベントに関する情報を表示する表示領域のサイズを調整する。また、表示制御部20は、イベントにリモート環境から参加するユーザが見る表示部に、イベントの実際の会場との類似度を内的状態に応じて調整した映像を表示させてもよい。また、表示制御部20は、イベントにリモート環境から参加するユーザが所定の条件を満たすときに、ユーザが視認可能な範囲内に、所定の条件に応じた情報提供画像及び視覚効果画像の少なくとも一方を表示させてもよい。視覚効果画像は、イベントの会場に参加しており、内的状態が所定の条件を満たす別のユーザの仮想人物画像を含んでいてもよい。 The display control unit 20 adjusts the size of the display area for displaying information about the event based on the internal state estimated by the second estimation unit 19. Further, the display control unit 20 may display an image in which the degree of similarity with the actual venue of the event is adjusted according to the internal state on the display unit viewed by the user who participates in the event from the remote environment. Further, when the user who participates in the event from the remote environment satisfies the predetermined condition, the display control unit 20 has at least one of the information providing image and the visual effect image according to the predetermined condition within the range visible to the user. May be displayed. The visual effect image may include a virtual person image of another user who is participating in the venue of the event and whose internal state satisfies a predetermined condition.
 図3に示すように、情報処理装置1は、情報交換部21を備えていてもよい。情報交換部21は、イベントにリモート環境から参加するユーザが所定の条件を満たすときに、アバタを介して、アバタに対応する別の人物と情報交換を行う。 As shown in FIG. 3, the information processing apparatus 1 may include an information exchange unit 21. When a user who participates in an event from a remote environment satisfies a predetermined condition, the information exchange unit 21 exchanges information with another person corresponding to the avatar via the avatar.
 図4は本実施形態による情報処理装置1のソフトウェア構成を示すブロック図である。図4のソフトウェア構成は、処理サーバ6、DBサーバ7及び配信サーバ9により実行されるものである。 FIG. 4 is a block diagram showing a software configuration of the information processing apparatus 1 according to the present embodiment. The software configuration of FIG. 4 is executed by the processing server 6, the DB server 7, and the distribution server 9.
 図4に示すように、情報処理装置1のソフトウェア構成は、リアルタイム処理31と、予測処理32と、データ配信処理33とを備えている。リアルタイム処理31と予測処理32は主に処理サーバ6とDBサーバ7により実行される。データ配信処理33は主に配信サーバ9により実行される。 As shown in FIG. 4, the software configuration of the information processing apparatus 1 includes a real-time processing 31, a prediction processing 32, and a data distribution processing 33. The real-time processing 31 and the prediction processing 32 are mainly executed by the processing server 6 and the DB server 7. The data distribution process 33 is mainly executed by the distribution server 9.
 リアルタイム処理31は、センシング情報(例えば映像データ)に基づいてリアルタイムに特徴量を抽出する処理を行う。リアルタイム処理31は、顔認識実行プログラム31aと、ボーンデータ抽出プログラム31bと、音響データ抽出プログラム31cと、試合実況データ抽出プログラム31dと、抽出結果格納処理31eとを有する。 The real-time processing 31 performs processing for extracting features in real time based on sensing information (for example, video data). The real-time processing 31 includes a face recognition execution program 31a, a bone data extraction program 31b, an acoustic data extraction program 31c, a match live data extraction program 31d, and an extraction result storage process 31e.
 顔認識実行プログラム31aは、映像データに基づいて顔認識処理を行って、各観客の笑顔の度合い、手首の移動量、頭の移動量、目の開き度合い、視線方向などを推測する。ボーンデータ抽出プログラム31bは、映像データに基づいて、各顧客のボーンデータを抽出する。音響データ抽出プログラム31cは、イベント会場の音響データを抽出する。抽出される音響データは、例えば、音の方向と音量に関する情報を含んでいる。試合実況データ抽出プログラム31dは、例えば主催チームのオフィシャルサイトで試合の実況データを配信している場合、この実況データを取得する。 The face recognition execution program 31a performs face recognition processing based on the video data, and estimates the degree of smile of each spectator, the amount of movement of the wrist, the amount of movement of the head, the degree of opening of the eyes, the direction of the line of sight, and the like. The bone data extraction program 31b extracts the bone data of each customer based on the video data. The acoustic data extraction program 31c extracts the acoustic data of the event venue. The extracted acoustic data contains, for example, information about the direction and volume of sound. The game live data extraction program 31d acquires the game live data, for example, when the game live data is distributed on the official site of the host team.
 抽出結果格納処理31eは、顔認識実行プログラム31aの出力データと、ボーンデータ抽出プログラム31bで抽出されたデータと、音響データ抽出プログラム31cで抽出されたデータと、試合実況データ抽出プログラム31dで抽出されたデータとを、特徴量としてDBサーバ7に格納する制御を行う。 The extraction result storage process 31e is extracted by the output data of the face recognition execution program 31a, the data extracted by the bone data extraction program 31b, the data extracted by the acoustic data extraction program 31c, and the match live data extraction program 31d. Control is performed to store the collected data in the DB server 7 as a feature amount.
 予測処理32は、各観客にタグ情報を付与する処理と、各観客を複数のクラスタに分類する処理と、各観客の行動判定を行う処理とを行う。より具体的には、予測処理32は、予測データ作成プログラム32aと、ファンタグ付けプログラム32bと、ファンクラスタリングプログラム32cと、ファン行動判定プログラム32dと、予測結果格納処理32eとを有する。 The prediction process 32 performs a process of assigning tag information to each spectator, a process of classifying each spectator into a plurality of clusters, and a process of determining the behavior of each spectator. More specifically, the prediction process 32 includes a prediction data creation program 32a, a fan tagging program 32b, a fan clustering program 32c, a fan behavior determination program 32d, and a prediction result storage process 32e.
 予測データ作成プログラム32aは、ファンタグ付けプログラム32b、ファンクラスタリングプログラム32c、及びファン行動判定プログラム32dを作成するためのプログラムである。 The prediction data creation program 32a is a program for creating a fan tagging program 32b, a fan clustering program 32c, and a fan behavior determination program 32d.
 ファンタグ付けプログラム32bは、リアルタイム処理31で得られた特徴量に基づいて、イベントに参加するユーザ(スポーツイベントの場合はファン)にタグ情報を付与する。タグ情報は、上述したように、例えばホームファン、アウェイファン、初心者などである。 The fan tagging program 32b adds tag information to users who participate in the event (fans in the case of sports events) based on the feature amount obtained by the real-time processing 31. As described above, the tag information is, for example, a home fan, an away fan, a beginner, and the like.
 ファンクラスタリングプログラム32cは、リアルタイム処理31で得られた特徴量に基づいて、イベントに参加するユーザ(ファンなど)を複数のクラスタに分類する。複数のクラスタは、上述したように、例えば試合の盛り上がり度合いで複数のクラスタに分類してもよいし、試合への集中度で複数のクラスタに分類してもよいし、いくつかの条件を組み合わせて、複数のクラスタに分類してもよい。 The fan clustering program 32c classifies users (fans, etc.) who participate in the event into a plurality of clusters based on the feature amount obtained by the real-time processing 31. As described above, a plurality of clusters may be classified into a plurality of clusters according to the degree of excitement of the game, or may be classified into a plurality of clusters according to the degree of concentration in the game, or a combination of several conditions. It may be classified into a plurality of clusters.
 ファン行動判定プログラム32dは、イベントに参加するユーザ(ファンなど)の行動を判定する。例えば、試合を凝視している熱心なファン、試合を観ずに飲食や会話等をしているファン、スマートフォンばかり見て退屈していそうなファン等を判定する。 The fan behavior determination program 32d determines the behavior of users (fans, etc.) who participate in the event. For example, it is determined that an enthusiastic fan who is staring at the game, a fan who eats and drinks or has a conversation without watching the game, a fan who seems to be bored by looking only at a smartphone, and the like.
 予測結果格納処理32eは、ファンタグ付けプログラム32b、ファンクラスタリングプログラム32c、及びファン行動判定プログラム32dの実行により得られた予測結果を、所定のデータベースに格納する処理を行う。 The prediction result storage process 32e performs a process of storing the prediction result obtained by executing the fan tagging program 32b, the fan clustering program 32c, and the fan behavior determination program 32d in a predetermined database.
 データ配信処理33は、配信サーバ9と運営者サーバ(ツール選択用サーバ)30に対して、予測処理32により得られた予測結果のデータに基づいて、観客と運営者への配信用の情報を生成する。観客向けの配信用の情報は、配信サーバ9から、対応する観客の携帯端末8に送信される。運営者向けの配信用の情報は、運営者サーバ30から、運営者のPC等に送信される。運営者サーバ30には、クラウドストレージ22に格納された観客席の映像データと試合の映像データが入力される。 The data distribution process 33 distributes information for distribution to the audience and the operator to the distribution server 9 and the operator server (tool selection server) 30 based on the prediction result data obtained by the prediction process 32. Generate. Information for distribution for spectators is transmitted from the distribution server 9 to the corresponding spectator's mobile terminal 8. Information for distribution for the operator is transmitted from the operator server 30 to the operator's PC or the like. The video data of the spectators' seats and the video data of the match stored in the cloud storage 22 are input to the operator server 30.
 図5は本実施形態による情報処理装置1の処理動作を示すフローチャートである。このフローチャートは、イベントの開催期間中に、定期的、不定期的、又は継続的に実施される。まず、イベントに参加するユーザの氏名等のユーザ情報を取得する(ステップS1)。イベントに参加するユーザがイベントへの参加を申し込むにあたって、氏名や連絡先などのユーザ情報を登録している場合には、この登録情報を取得してもよい。例えば、イベントの観客席が予め指定されている場合、観客席の座席情報を含めてユーザ情報を取得してもよい。ユーザの座席情報を取得できれば、どの席に誰が座っているかを把握でき、情報配信の信頼性を向上できる。 FIG. 5 is a flowchart showing the processing operation of the information processing apparatus 1 according to the present embodiment. This flow chart is performed regularly, irregularly, or continuously during the duration of the event. First, user information such as the names of users participating in the event is acquired (step S1). When a user who participates in an event applies for participation in an event and has registered user information such as a name and contact information, this registration information may be acquired. For example, when the spectator seats of the event are designated in advance, the user information may be acquired including the seat information of the spectator seats. If the user's seat information can be acquired, it is possible to grasp who is sitting in which seat and improve the reliability of information distribution.
 次に、センシング情報を取得して、センシング情報に基づいてユーザの特徴量を抽出する(ステップS2)。センシング情報に観客席の映像データが含まれている場合、上述した顔認識実行プログラム31aとボーンデータ抽出プログラム31bを実行して、ユーザの顔の表情、姿勢、体の動き、骨格情報などを含む特徴量を抽出できる。また、映像データに基づいて、オプティカルフローによりユーザの動き量を含む特徴量を抽出できる。さらに、センシング情報に音響データが含まれている場合、ユーザの音声を含む特徴量を抽出できる。 Next, the sensing information is acquired, and the feature amount of the user is extracted based on the sensing information (step S2). When the sensing information includes video data of the audience seat, the face recognition execution program 31a and the bone data extraction program 31b described above are executed to include the facial expression, posture, body movement, skeletal information, etc. of the user. Feature quantity can be extracted. Further, based on the video data, the feature amount including the movement amount of the user can be extracted by the optical flow. Further, when the sensing information includes acoustic data, the feature amount including the user's voice can be extracted.
 次に、イベントの環境情報及びコンテンツ情報を取得する(ステップS3)。例えば、スポーツイベントの場合、図4に示した試合実況データ抽出プログラム31dにて、試合の経過情報を取得してもよい。あるいは、試合の出場選手や所属チームの選手のスタッツ情報をDBサーバ7等から取得してもよい。あるいは、イベントの参加人数や、開催場所、天気、温度、季節、開催時刻等の環境情報を取得してもよい。 Next, acquire the event environment information and content information (step S3). For example, in the case of a sporting event, the game progress information may be acquired by the game live data extraction program 31d shown in FIG. Alternatively, the stats information of the players participating in the match and the players of the team to which they belong may be acquired from the DB server 7 or the like. Alternatively, environmental information such as the number of participants in the event, the venue, the weather, the temperature, the season, and the time of the event may be acquired.
 次に、イベントに参加するユーザ又はグループ単位での属性判定と行動判定を行い、その判定結果に基づいてクラスタリング処理を正常に行えたか否かを判断する(ステップS4)。ここでは、例えば、映像データに基づいて抽出された特徴量に対して、既存のデータベースを用いて教師有りの学習を行った結果を利用して、各ユーザの属性判定と行動判定を行う。また、属性が似ている複数の参加人をグループ化して、複数のクラスタに分類するクラスタリング処理を行う。 Next, attribute determination and action determination are performed for each user or group participating in the event, and it is determined whether or not the clustering process has been normally performed based on the determination result (step S4). Here, for example, the attribute determination and the behavior determination of each user are performed using the result of learning with a teacher using an existing database for the feature amount extracted based on the video data. In addition, a clustering process is performed in which a plurality of participants having similar attributes are grouped and classified into a plurality of clusters.
 学習が正しく行われていない場合など、クラスタリング処理を正常に行えない場合、あるいは行えても信頼性が低い場合は、複数のクラスタへの分類を行わずに、解析結果をデータ保持DBに登録する(ステップS5)。 If the clustering process cannot be performed normally, such as when learning is not performed correctly, or if the reliability is low even if it can be performed, the analysis results are registered in the data retention DB without classifying into multiple clusters. (Step S5).
 ステップS4で、属性判定と行動判定を行って、クラスタリング処理を正常に行えた場合は、複数のクラスタへの分類結果を結果DBに登録する(ステップS6)。 If the attribute determination and the action determination are performed in step S4 and the clustering process can be performed normally, the classification results for a plurality of clusters are registered in the result DB (step S6).
 次に、ステップS5とS6の処理結果に基づいて、出力処理を行う(ステップS7)。クラスタリング処理を正常に行えた場合は、各クラスタに分類されたユーザに対して、各クラスタに対応する情報提供を行ったり、運営者に対して複数のクラスタの分類結果を提供したりする。あるいは、イベントの会場全体または一部に、イベントを盛り上げるための演出等を行ってもよい。このように、出力処理の具体的な内容は任意である。 Next, output processing is performed based on the processing results of steps S5 and S6 (step S7). If the clustering process can be performed normally, the information corresponding to each cluster is provided to the users classified into each cluster, and the classification result of a plurality of clusters is provided to the operator. Alternatively, the entire or part of the venue of the event may be directed to liven up the event. As described above, the specific content of the output processing is arbitrary.
 図6Aはボーンデータ抽出プログラム31bにて抽出されたボーンデータ34の一例を示す図である。図6Aはスタジアム3の観客席を撮影した映像データにボーンデータ34を重畳したものである。ボーンデータ34は、図示のように、円と折れ線で構成されている。円は観客の顔の位置を示しており、円に接続された折れ線は体の動きを表している。立ち上がって応援している観客は、折れ線が長くなる。また、拍手をしているかどうかも、ボーンデータ34で検出できる。例えば、得点を上げたタイミングで折れ線が長くなった観客は、その得点を上げたチームを応援しているファンであると推定できる。 FIG. 6A is a diagram showing an example of bone data 34 extracted by the bone data extraction program 31b. FIG. 6A shows the bone data 34 superimposed on the video data obtained by photographing the audience seats of the stadium 3. As shown in the figure, the bone data 34 is composed of a circle and a polygonal line. The circle indicates the position of the spectator's face, and the polygonal line connected to the circle indicates the movement of the body. Audiences who stand up and cheer for them have a long line. Further, whether or not the person is applauding can also be detected by the bone data 34. For example, an spectator whose polygonal line becomes longer at the timing of increasing the score can be presumed to be a fan supporting the team that increased the score.
 図6Bは属性判定を行った結果の一例を示す図である。上述したように、得点を上げたときに体の動きが大きかった否かで、ホームファンか、アウェイファンか、初心者かを判別することができる。図6Bは、観客席を撮影した映像データを解析することにより、ホームエリア35a内のホームファンが集まっているエリア35bと、アウェイファンが集まっているエリア35cと、初心者が集まっているエリア35dとを自動判別した例を示している。 FIG. 6B is a diagram showing an example of the result of attribute determination. As mentioned above, it is possible to determine whether the player is a home fan, an away fan, or a beginner based on whether or not the body moves a lot when the score is increased. FIG. 6B shows an area 35b in the home area 35a where home fans are gathered, an area 35c where away fans are gathered, and an area 35d where beginners are gathered by analyzing video data obtained by photographing the audience seats. Is shown as an example of automatic determination.
 本実施形態による情報処理装置1は、イベントに参加するユーザの内的状態の推定と属性判定を行うために、イベントの状況に合わせて、「意味のある属性」を予め用意している。情報処理装置1は、映像データを解析した結果に基づいて、予め用意した「意味のある属性」のどれかに当てはめる処理を行う。これにより、ユーザの属性判定を簡易に行うことができる。 The information processing apparatus 1 according to the present embodiment prepares "meaningful attributes" in advance according to the situation of the event in order to estimate the internal state of the user participating in the event and determine the attributes. The information processing apparatus 1 performs a process of applying to any of the "meaningful attributes" prepared in advance based on the result of analyzing the video data. This makes it possible to easily determine the attributes of the user.
 図7A、図7B及び図7Cは、スポーツイベントの観客の内的状態の推定と属性判定を行うための「意味のある属性」の種別の一例を示す図である。図7Aは試合が行われている最中(ピリオド中とも呼ぶ)の「意味のある属性」の種別を示し、図7Bは試合が中断している最中(ピリオド以外とも呼ぶ)の「意味のある属性」の種別を示し、図7Cは試合中の全時間帯の「意味のある属性」の種別を示している。 7A, 7B and 7C are diagrams showing an example of a type of "meaningful attribute" for estimating the internal state of the spectator of a sporting event and determining the attribute. FIG. 7A shows the types of "meaningful attributes" during a match (also called during a period), and FIG. 7B shows the "meaning" during a match being interrupted (also called a non-period). The type of "certain attribute" is shown, and FIG. 7C shows the type of "meaningful attribute" for all time zones during the match.
 図7Aに示すように、試合の継続中には、試合を盛り上げるための属性として、試合展開への反応と、試合を見ているかどうかがある。試合展開への反応は、笑顔の度合いと、手首の移動量と、目の開き度合いで判断できる。試合を見ているかどうかは、笑顔の度合いと、コート(スタジアム)3を見ているか否かで判断できる。 As shown in FIG. 7A, during the continuation of the match, the attributes for enlivening the match include the reaction to the match development and whether or not the match is being watched. The reaction to the game development can be judged by the degree of smile, the amount of wrist movement, and the degree of eye opening. Whether or not you are watching the game can be judged by the degree of smile and whether or not you are watching the court (stadium) 3.
 また、試合の継続中には、他の候補の属性として、反応の有無(意図のない反応)と、意図のある反応と、得点前後の反応と、イベント会場全体の音声量と、解説がある。意図のない反応は、コート(スタジアム)3を見ているか否かでプレー中の集中度を判断でき、コートを見ているかと目の開き度合いで、コートを凝視しているかを判断できる。意図のある反応は、手首の移動量により意図して手首を動かしていることを判断でき、また、顔の下の領域のオプティカルフローの移動量により、拍手していると判断できる。また、笑顔の度合いで、得点を上げて喜んだか、悲しんだかを判断できる。音響データの大きさにより、会場全体の音声量を判断できる。また、口の開き方により、試合の解説をしているのか、解説を聞いているかを判断できる。 In addition, during the continuation of the match, the attributes of other candidates include the presence or absence of a reaction (unintentional reaction), the intentional reaction, the reaction before and after the score, the audio volume of the entire event venue, and commentary. .. For unintentional reactions, the degree of concentration during play can be determined by whether or not the player is looking at the court (stadium) 3, and the degree of eye opening can be used to determine whether or not the player is staring at the court. The intentional reaction can be determined to be intentionally moving the wrist based on the amount of movement of the wrist, and can be determined to be applauding based on the amount of movement of the optical flow in the area under the face. Also, depending on the degree of smile, you can judge whether you are happy or sad by increasing the score. The amount of sound in the entire venue can be judged from the size of the acoustic data. Also, depending on how you open your mouth, you can judge whether you are explaining the game or listening to the explanation.
 図7Bに示すように、試合の中断中には、パフォーマンスへの反応度のための属性として、観客巻き込み型のパフォーマンスへの参加度と、非観客巻き込み型のパフォーマンスへの関心度がある。観客巻き込み型のパフォーマンスへの参加度は、パフォーマンスが行われているコート(スタジアム)3を見ているかでギブアウェイ参加度を判断でき、オプティカルフローにて観客巻き込み型チアダンス参加度を判断できる。非観客巻き込み型パフォーマンス参加度は、コート(スタジアム)3を見ていることと目の開き度合いの平均により、チアダンス関心度、スタッフダンス関心度、フリースロー関心度、記念撮影関心度、コント関心度を判断できる。また、顔の移動量、骨格の移動量、及びオプティカルフローの移動量により、DJの関心度を判断できる。 As shown in FIG. 7B, during the interruption of the game, the degree of participation in the spectator involvement type performance and the degree of interest in the non-spectator involvement type performance are the attributes for the responsiveness to the performance. As for the degree of participation in the audience involvement type performance, the degree of participation in the giveaway can be determined by looking at the court (stadium) 3 where the performance is performed, and the degree of participation in the audience involvement type cheerleader can be determined by the optical flow. The degree of non-audience involvement type performance participation is the degree of cheerleader interest, staff dance interest, free throw interest, commemorative photo interest, and control interest, depending on the average degree of eye opening and watching court (stadium) 3. Can be judged. In addition, the degree of interest of DJ can be determined from the amount of movement of the face, the amount of movement of the skeleton, and the amount of movement of the optical flow.
 図7Cに示すように、試合の全時間帯において、他の人とのシンクロ率に関する属性と、試合に直接関係しない行動に関する属性とがある。他の人とのシンクロ率については、ボーンデータ、目の開き度合い、笑顔の度合い、顔の移動量に基づいて教師有り学習を行うことで、ファン属性(ホームファン、アウェイファン、初心者)と応援スタイル属性(じっと観戦する、声を上げて応援する、身振り手振りで応援する、撮影重視)を判断できる。また、顔の周りの色を検出することで、チームのグッズの属性と、チームカラーの服の属性を判断できる。その他、他の人とのシンクロ率については、私服か否かの属性と、年代の属性と、グループで応援に来たか否かの属性と、小さい子供と一緒に来たか否かの属性とがある。 As shown in FIG. 7C, there are attributes related to the synchronization rate with other people and attributes related to actions that are not directly related to the game in the entire time zone of the game. Regarding the synchronization rate with other people, we support with fan attributes (home fans, away fans, beginners) by conducting supervised learning based on bone data, degree of eye opening, degree of smile, and amount of face movement. You can judge the style attributes (watching the game, cheering with a loud voice, cheering with gestures, focusing on shooting). Also, by detecting the color around the face, the attributes of the team goods and the attributes of the team-colored clothes can be determined. In addition, regarding the synchronization rate with other people, the attribute of whether or not it is plain clothes, the attribute of age, the attribute of whether or not they came to support in a group, and the attribute of whether or not they came with a small child be.
 また、試合に直接関係しない行動については、バスケットボールに無関係な属性として、飲んでいるか否かの属性と、食べているか否かの属性とがある。また、観客席にいるか否かに関して、データの有無により離席しているか否かの属性を判断できる。また、手首の位置が顔より上か否かで、両手を上げているか否かの属性を判断でき、手首のX方向がクロスしているか否かで、腕を組んでいるか否かの属性を判断でき、顔と手首の距離により顔に手を付けているか否かの属性を判断できる。 For behaviors that are not directly related to the game, there are attributes that are not related to basketball, such as whether or not they are drinking and whether or not they are eating. In addition, regarding whether or not the person is in the audience seat, the attribute of whether or not the person is away from the seat can be determined depending on the presence or absence of data. In addition, the attribute of whether or not both hands are raised can be determined by whether or not the wrist is above the face, and the attribute of whether or not the arms are crossed is determined by whether or not the X direction of the wrist is crossed. It can be judged, and the attribute of whether or not the face is touched can be judged from the distance between the face and the wrist.
 また、試合に直接関係しない行動については、バスケットボールに関係するかもしれない行動の属性として、スマートフォンを見ているか否かの属性がある。また、試合展開に関係なく体が動いているか否かで、落ち着いているか否の属性を判断できる。また、笑顔の度合いで、表情(心配、変化頻度)の属性を判断できる。この他、パンフレットを見ているか否かの属性と、会話が多いか少ないかの属性と、写真撮影(試合撮影、自撮り、演出撮影)をしているか否かの属性と、コート(スタジアム)3のスクリーンを見ているか否かの属性がある。 For actions that are not directly related to the game, there is an attribute of actions that may be related to basketball, such as whether or not you are watching a smartphone. In addition, the attribute of calmness can be determined by whether or not the body is moving regardless of the game development. In addition, the attributes of facial expressions (anxiety, frequency of change) can be determined by the degree of smile. In addition, the attribute of whether or not you are looking at the pamphlet, the attribute of whether or not you have many conversations, the attribute of whether or not you are taking pictures (match photography, selfie, directing photography), and the court (stadium) There is an attribute of whether or not you are looking at the screen of 3.
 スポーツイベントを含めて種々のイベントでは、参加するユーザの盛り上がり度合いや集中度をより高めるために、イベントの運営者が種々の演出を行うことが多い。本実施形態では、運営者が行った演出の効果を運営者自身で客観的に検証するための運営者用ツール画面を用意している。この運営者用ツール画面には、上述した状況画像(運営者用画像)が表示される。 In various events including sporting events, the event operator often performs various productions in order to increase the degree of excitement and concentration of the participating users. In this embodiment, an operator tool screen is prepared for the operator to objectively verify the effect of the effect performed by the operator. The above-mentioned situation image (image for the operator) is displayed on the tool screen for the operator.
 図8は運営者用ツール画面40の一例を示す図である。図8の運営者用ツール画面40は、運営者が所持するPC等に表示させることができる。運営者は、運営者用ツール画面40に表示された状況画像(運営者用画像)により、イベント中に行った演出に対するイベント参加者の反応等を詳細に検証できる。 FIG. 8 is a diagram showing an example of the operator tool screen 40. The operator tool screen 40 of FIG. 8 can be displayed on a PC or the like owned by the operator. The operator can verify in detail the reaction of the event participants to the production performed during the event by the situation image (image for the operator) displayed on the operator tool screen 40.
 図8の運営者用ツール画面40は、バスケットボールの試合中にパフォーマンス等の演出を行った例を示している。図8の運営者用ツール画面40は、試合の映像データを表示する第1エリア40aと、観客の反応を表示する第2エリア40bと、試合の得点変化と観客の盛り上がり度合い、集中度を表示する第3エリア40cとを有する。本明細書では、第2エリア40bと第3エリア40cに表示される画像を総称して、状況画像(運営者用画像)と呼ぶ。 The operator tool screen 40 in FIG. 8 shows an example in which a performance or the like is produced during a basketball game. The operator tool screen 40 of FIG. 8 displays the first area 40a for displaying the video data of the game, the second area 40b for displaying the reaction of the spectators, the score change of the game, the degree of excitement of the spectators, and the degree of concentration. It has a third area 40c and the like. In the present specification, the images displayed in the second area 40b and the third area 40c are collectively referred to as a situation image (operator image).
 図9Aは図8の第2エリア40bに表示される画像を拡大した図である。第2エリア40bの画像には、観客の顔に、盛り上がり度合いを示す円(識別子)41が重畳されている。盛り上がり度合いが大きいほど、円41の半径を大きくしている。第2エリア40bの上側には、第2エリア40bに表示する観客席の範囲を選択する複数のボタン42が並んでいる。図9Aの例では、Tokyo benchの範囲が選択されている。運営者は、ボタンを任意に選択することで、スタジアム3全体の観客席中の任意の場所を表示させて、その場所での観客の盛り上がり度合いを視覚的に把握することができる。 FIG. 9A is an enlarged view of the image displayed in the second area 40b of FIG. In the image of the second area 40b, a circle (identifier) 41 indicating the degree of excitement is superimposed on the face of the audience. The greater the degree of excitement, the larger the radius of the circle 41. Above the second area 40b, a plurality of buttons 42 for selecting the range of spectator seats to be displayed in the second area 40b are arranged. In the example of FIG. 9A, the range of Tokyo bench is selected. By arbitrarily selecting a button, the operator can display an arbitrary place in the spectator seats of the entire stadium 3 and visually grasp the degree of excitement of the spectators at that place.
 図9Bは図8の第3エリア40cに表示される画像の一部を拡大した図である。図9Bのグレイの範囲43は試合継続期間であり、グレイの間の黒の範囲44は試合中断期間である。横軸は時間であり、縦実線は対戦チーム同士の得点を表している。図9Bには、2本の折れ線が図示されているが、そのうちの一方は観客の盛り上がり度合いを示し、他方は観客の集中度を示している。また、矩形枠45内の数字は、その時点での各チームの得点と得点差を示している。 FIG. 9B is an enlarged view of a part of the image displayed in the third area 40c of FIG. The gray range 43 of FIG. 9B is the match duration, and the black range 44 between the grays is the match interruption period. The horizontal axis is time, and the vertical solid line shows the scores of the opposing teams. In FIG. 9B, two polygonal lines are shown, one of which shows the degree of excitement of the audience and the other of which shows the degree of concentration of the audience. Further, the numbers in the rectangular frame 45 indicate the score and the score difference of each team at that time.
 運営者用ツール画面40内の第2エリア40bと第3エリア40cの画像により、運営者は、試合の経過とともに、観客の盛り上がり度合いと集中度がどのように変化したかを詳細に検証することができる。スポーツイベントでは、ホームファンとアウェイファンでは、得点状況により異なった反応を示すが、図8の運営者用ツール画面40では、ホームファンとアウェイファンが座っている範囲を別々に詳細に検証できる。 With the images of the second area 40b and the third area 40c in the tool screen 40 for the operator, the operator should examine in detail how the degree of excitement and concentration of the spectators changed with the progress of the game. Can be done. At a sporting event, the home fan and the away fan show different reactions depending on the scoring situation, but on the operator tool screen 40 of FIG. 8, the range in which the home fan and the away fan are sitting can be examined in detail separately.
 図10は運営者用ツール画面40を生成する処理動作の一例を示す図である。図10の処理動作は、処理サーバ6が行ってもよいし、運営者サーバ30が行ってもよい。本明細書では、運営者用ツール画面40を生成する処理を盛り上がり可視化ツールと呼ぶ。 FIG. 10 is a diagram showing an example of a processing operation for generating the operator tool screen 40. The processing operation of FIG. 10 may be performed by the processing server 6 or the operator server 30. In the present specification, the process of generating the operator tool screen 40 is referred to as an exciting visualization tool.
 盛り上がり可視化ツールは、試合解析処理46aと、盛り上がりの円描画プログラム46bと、解析データ処理プログラム46cと、ウェブ表示プログラム46dと、CSS(Cascading Style Sheets)46eとを有する。 The excitement visualization tool has a match analysis process 46a, an excitement circle drawing program 46b, an analysis data processing program 46c, a web display program 46d, and CSS (Cascading Style Sheets) 46e.
 試合解析処理46aは、後述するように、映像データを解析した結果に基づいて、各観客の分析項目として、各観客の集中度、盛り上がり度合い、ファン属性の情報を出力する。また、試合解析処理46aは、試合の分析項目として、得点、選手交代、試合停止期間、試合中断中に行われるパフォーマンス内容、パフォーマンス時間の情報を出力する。また、試合解析処理46aは、映像データのフレームごと、及びファンごとの集中度と盛り上がり度合いの情報を出力する。 As will be described later, the match analysis process 46a outputs information on the concentration, excitement, and fan attributes of each spectator as analysis items for each spectator based on the result of analyzing the video data. In addition, the match analysis process 46a outputs information on the score, player substitution, match suspension period, performance content performed during the match interruption, and performance time as match analysis items. Further, the game analysis process 46a outputs information on the degree of concentration and the degree of excitement for each frame of the video data and for each fan.
 盛り上がりの円描画プログラム46bは、観客席の映像データと試合解析処理46aの出力データとに基づいて、個々の観客の盛り上がり度合いに比例した円を観客の顔画像に重畳させる。 The exciting circle drawing program 46b superimposes a circle proportional to the degree of excitement of each spectator on the face image of the spectator based on the video data of the spectator seat and the output data of the game analysis process 46a.
 解析データ処理プログラム46cは、解析後のCSVデータから、ツールで必要な情報のみを抽出する。解析後にpickleデータをjson形式に変換する。 The analysis data processing program 46c extracts only the information required by the tool from the CSV data after analysis. Convert pickle data to json format after analysis.
 ウェブ表示プログラム46dは、試合のビデオ映像と、円描画プログラム46bで生成された画像と、ツールのレイアウト定義を記述したCSS46eとに基づいて、運営者用ツール画面40の第1エリア40aに試合のビデオ映像を再生する処理を行うとともに、第2エリア40b及び第3エリア40cの画像を生成する。ウェブ表示プログラム46dとCSS46eの処理は、WebAPサーバ47が行ってもよい。 The web display program 46d is based on the video image of the game, the image generated by the circle drawing program 46b, and the CSS46e that describes the layout definition of the tool, and the game is displayed in the first area 40a of the tool screen 40 for the operator. The process of reproducing the video image is performed, and the images of the second area 40b and the third area 40c are generated. The WebAP server 47 may perform the processing of the web display program 46d and the CSS46e.
 図11は図10の試合解析処理46aの詳細を示す図である。試合解析処理46aは、観客映像データ処理プログラム48aと、試合データ処理プログラム48bと、グラフ表示プログラム48cと、指標評価プログラム48dと、デザイナ用簡易CSV作成プログラム48eと、設定変更プログラム48fとを有する。 FIG. 11 is a diagram showing details of the game analysis process 46a of FIG. The game analysis process 46a includes a spectator video data processing program 48a, a game data processing program 48b, a graph display program 48c, an index evaluation program 48d, a simple CSV creation program 48e for designers, and a setting change program 48f.
 観客映像データ処理プログラム48aには、特徴量化処理48gで生成された特徴量に関するデータが入力される。特徴量化処理48gでは、観客席の映像と試合の映像に対して顔認識実行プログラム31a、ボーンデータ抽出プログラム31b、オプティカルフロー処理等を実行することにより、特徴量を抽出する。特徴量化処理48gにより、特徴量を表す検出データとファン属性情報とが出力される。特徴量化処理48gでは、観客の振るまいと試合展開からファンの属性を決定する。 Data related to the feature amount generated by the feature quantification process 48g is input to the audience video data processing program 48a. In the feature quantity processing 48g, the feature quantity is extracted by executing the face recognition execution program 31a, the bone data extraction program 31b, the optical flow processing, and the like on the video of the spectator seat and the video of the game. By the feature quantification process 48g, the detection data representing the feature quantity and the fan attribute information are output. In the feature quantification process 48g, the attributes of the fans are determined from the behavior of the spectators and the development of the game.
 観客映像データ処理プログラム48aは、観客席の映像データに基づいて、個人識別、分析項目の追加、複数カメラ4の映像データ統合、カメラ4間の補正等を行い、各観客の集中度、盛り上がり度合い、及びファン属性などを出力する。 The spectator video data processing program 48a performs personal identification, addition of analysis items, video data integration of a plurality of cameras 4, correction between cameras 4, etc. based on the video data of the spectator seats, and the degree of concentration and excitement of each spectator. , And fan attributes, etc. are output.
 試合データ処理プログラム48bには、試合経過データが例えばインターネットを介して入力される。試合経過データは、試合経過情報の他、試合中に行われるパフォーマンス情報や試合の映像時間などの情報を含んでいる。試合データ処理プログラム48bは、観客映像データ処理プログラム48aの出力データと試合経過データとに基づいて、試合データ処理、個人項目統合、パフォーマンス情報処理、手入力数値の矛盾検出等を行って、各観客の集中度、盛り上がり度合い及びファン属性を出力するとともに、試合についての分析項目として、得点、選手交代、試合停止期間、パフォーマンス内容と期間の情報を出力する。 Match progress data is input to the match data processing program 48b via, for example, the Internet. The match progress data includes information such as performance information performed during the match and video time of the match, in addition to the match progress information. The match data processing program 48b performs match data processing, personal item integration, performance information processing, manual inconsistency detection, etc. based on the output data of the spectator video data processing program 48a and the match progress data, and each spectator. In addition to outputting the degree of concentration, the degree of excitement, and the fan attributes, information on points, player substitution, game suspension period, performance content and period is output as analysis items for the game.
 グラフ表示プログラム48cは、試合データ処理プログラム48bの実行で得られたデータに基づいて、集中度と盛り上がり度の時間変化を示すグラフ49を生成する。 The graph display program 48c generates a graph 49 showing the time change of the degree of concentration and the degree of excitement based on the data obtained by executing the game data processing program 48b.
 図12はグラフ表示プログラム48cが生成するグラフ49の一例を示す図であり、図11に図示されたものを拡大した図である。横軸は試合の経過時間、縦軸は観客の盛り上がり度合いと集中度を示している。図12には、観客の盛り上がり度合いのグラフと集中度のグラフが図示されている。図12のグレーの期間49aは試合継続期間であり、濃いグレーの期間49bは、何らかの演出が行われた期間を示している。このグラフを運営者に提示することで、演出により観客が盛り上がったか否かを一目で把握できる。 FIG. 12 is a diagram showing an example of the graph 49 generated by the graph display program 48c, and is an enlarged view of the graph 49 shown in FIG. The horizontal axis shows the elapsed time of the game, and the vertical axis shows the degree of excitement and concentration of the spectators. FIG. 12 shows a graph of the degree of excitement and a graph of the degree of concentration of the audience. The gray period 49a in FIG. 12 is the game continuation period, and the dark gray period 49b indicates the period in which some kind of production is performed. By presenting this graph to the operator, it is possible to grasp at a glance whether or not the audience was excited by the production.
 図11の指標評価プログラム48dは、集中度と盛り上がり度合いの評価のための指標を生成する。図13は指標評価プログラム48dで生成された指標50の一例を示す図であり、図11に図示されたものを拡大した図である。図13の指標50の例では、ホームチームが1~3ポイントの得点を上げた場合と、ホームチームの得点との相関と、アウェイチームが1~3ポイントの得点を上げた場合と、アウェイチームの得点との相関とのそれぞれについて、盛り上がり度合いが増加した観客の総数についての指標と、盛り上がり度合いが減少した観客の総数についての指標と、増加率についての指標と、所定期間内の盛り上がり度合いの平均値についての指標とが図示されている。 The index evaluation program 48d in FIG. 11 generates an index for evaluating the degree of concentration and the degree of excitement. FIG. 13 is a diagram showing an example of the index 50 generated by the index evaluation program 48d, and is an enlarged view of the one shown in FIG. In the example of the index 50 in FIG. 13, the correlation between the home team scoring 1 to 3 points, the correlation with the home team score, the away team scoring 1 to 3 points, and the away team For each of the correlations with the score of, the index of the total number of spectators whose degree of excitement increased, the index of the total number of spectators whose degree of excitement decreased, the index of the rate of increase, and the degree of excitement within a predetermined period. An index for the average value is shown.
 図11のデザイナ用簡易CSV作成プログラム48eは、試合データ処理プログラム48bの実行結果に基づいて、映像のフレームごと、及びファンごとの集中度と盛り上がり度合いの情報を含むCSVファイルを作成する。 The simple CSV creation program 48e for designers in FIG. 11 creates a CSV file containing information on the degree of concentration and excitement for each frame of video and for each fan based on the execution result of the game data processing program 48b.
 図11の設定変更プログラム48fは、試合データ処理プログラム48b、グラフ表示プログラム48c、指標評価プログラム48d、及びデザイナ用簡易CSV作成プログラム48eの実行により得られたファイルを格納する入出力フォルダを切り替える。また、設定変更プログラム48fは、グラフ表示プログラム48cが生成するグラフの色や表示位置の変更などを行う。また、設定変更プログラム48fは、指標評価プログラム48dが生成する指標を計算するための映像フレームを切り替える。 The setting change program 48f in FIG. 11 switches the input / output folder for storing the files obtained by executing the game data processing program 48b, the graph display program 48c, the index evaluation program 48d, and the simple CSV creation program 48e for the designer. Further, the setting change program 48f changes the color and display position of the graph generated by the graph display program 48c. Further, the setting change program 48f switches the video frame for calculating the index generated by the index evaluation program 48d.
 図14は図4のファンタグ付けプログラム32bの実行により生成された各観客のタグ情報を自動検証する処理動作を示す図である。本明細書では、タグ情報を自動検証する処理をタグ情報自動検証ツールと呼ぶ。タグ情報自動検証ツールの処理は、図1の処理サーバ6が行ってもよいし、運営者サーバ30等が行ってもよい。 FIG. 14 is a diagram showing a processing operation for automatically verifying the tag information of each spectator generated by executing the fan tagging program 32b of FIG. In this specification, the process of automatically verifying tag information is referred to as a tag information automatic verification tool. The processing of the tag information automatic verification tool may be performed by the processing server 6 of FIG. 1, the operator server 30, or the like.
 タグ情報自動検証ツールは、得点情報ファイル生成プログラム51aと、ボーンデータクリーニングプログラム51bと、ファンタグ付け/評価プログラム51cとを有する。 The tag information automatic verification tool has a score information file generation program 51a, a bone data cleaning program 51b, and a fan tagging / evaluation program 51c.
 得点情報ファイル生成プログラム51aは、特徴量化処理48gで得られた特徴量に関するデータに基づいて、試合映像と観客映像の時間のずれ補正と、得点チームと観客映像中の時間ファイル(以下、得点情報ファイル)を生成する。 The score information file generation program 51a corrects the time lag between the match video and the spectator video based on the data related to the feature obtained by the feature quantification process 48g, and the time file in the scoring team and the spectator video (hereinafter, score information). File) is generated.
 ボーンデータクリーニングプログラム51bは、得点情報ファイルに基づいて、ホームチームとアウェイチームが得点を上げたときに観客のボーン座標を抽出し、ボーン座標に基づいて得点時の観客の移動量を出力する。 The bone data cleaning program 51b extracts the bone coordinates of the spectator when the home team and the away team raise the score based on the score information file, and outputs the movement amount of the spectator at the time of scoring based on the bone coordinates.
 ファンタグ付け/評価プログラム51cは、ホームチームが得点を上げたときとアウェイチームが得点を上げたときで、ボーンデータの変化と得点時のボーンデータの平均値とを比較して、観客のタグ付けの判定を行う。判定は、例えばルールベース又は深層学習(DNN:Deep Neural Network)により行う。 The fan tagging / evaluation program 51c compares the changes in bone data with the average value of bone data at the time of scoring when the home team scores and when the away team scores, and tags the audience. Judgment of attachment is made. The determination is made, for example, by rule-based or deep learning (DNN: Deep Neural Network).
 図15Aはファンタグ付け/評価プログラム51cで生成されたタグ情報(推定タグとも呼ぶ)を含む映像データ52aを示し、図15Bは正しいタグ情報(正解タグとも呼ぶ)を含む映像データ52bを示している。図15A及び図15Bは、図14に図示された画像を拡大したものである。図15A及び図15Bでは、異なるタグ情報52a~52cをそれぞれ異なる色で表示している。 FIG. 15A shows video data 52a including tag information (also referred to as estimated tag) generated by the fan tagging / evaluation program 51c, and FIG. 15B shows video data 52b including correct tag information (also referred to as correct tag). There is. 15A and 15B are enlargements of the image illustrated in FIG. In FIGS. 15A and 15B, different tag information 52a to 52c are displayed in different colors.
 ファンタグ付け/評価プログラム51cでは、正解タグと推定タグを比較した結果を、図14に示したような表形式で提示するとともに、タグ情報の正解率を数値で提示する。 In the fan tagging / evaluation program 51c, the result of comparing the correct answer tag and the estimated tag is presented in a table format as shown in FIG. 14, and the correct answer rate of the tag information is presented numerically.
 本実施形態による情報処理装置1は、種々の目的に使用可能である。図16は本実施形態による情報処理装置1の主要な応用例を例示した図である。なお、本実施形態による情報処理装置1は、図16に示した応用例以外にも適用可能である。 The information processing device 1 according to the present embodiment can be used for various purposes. FIG. 16 is a diagram illustrating a main application example of the information processing apparatus 1 according to the present embodiment. The information processing apparatus 1 according to the present embodiment can be applied to applications other than those shown in FIG.
 図16に示すように、本実施形態による情報処理装置1は、イベントが盛り上がるタイミングをアーカイブ自動撮影する機能を有していてもよい。これにより、イベントが盛り上がるタイミングの映像を容易に取得できる。 As shown in FIG. 16, the information processing apparatus 1 according to the present embodiment may have a function of automatically archiving the timing at which an event rises. As a result, it is possible to easily acquire a video of the timing when the event is exciting.
 また、本実施形態による情報処理装置1は、コート(スタジアム)3上の注目されていないエリアで演出を行ったり、清掃作業を行う機能を備えていてもよい。注目されていないエリアで演出を行うことで、注目されていないエリアに観客の視線を向けることができる。また、注目されていないエリアで清掃作業を行うことで、観客の目を煩わすことなく、コートを清掃することができる。 Further, the information processing apparatus 1 according to the present embodiment may have a function of performing an effect or performing a cleaning work in an area on the court (stadium) 3 that is not attracting attention. By directing in an area that is not attracting attention, the audience's line of sight can be directed to the area that is not attracting attention. In addition, by performing the cleaning work in an area that is not attracting attention, the coat can be cleaned without bothering the eyes of the spectators.
 また、本実施形態による情報処理装置1は、観客の応援スタイルに応じて演出を変えてもよい。本実施形態によれば、観客の応援スタイルを自動抽出できるため、観客の盛り上がり度合いや集中度に応じて、演出の種類を切り替えることができる。例えば、声を出している観客には、観客席を振動させるなどして、盛り上がり度合いをより高めてもよい。 Further, the information processing device 1 according to the present embodiment may change the production according to the cheering style of the audience. According to this embodiment, since the cheering style of the audience can be automatically extracted, the type of production can be switched according to the degree of excitement and concentration of the audience. For example, for a spectator who is speaking out, the degree of excitement may be further increased by vibrating the spectator seats.
 また、本実施形態による情報処理装置1は、観客の試合への集中度が低下したときに、広告情報を提供したり、飲食の情報を提供してもよい。例えば、試合が一時的に中断した期間を見計らって、広告情報や飲食の情報を提供することで、広告効果を高めたり、スタジアム3内のレストラン等の売上を向上できる。 Further, the information processing device 1 according to the present embodiment may provide advertising information or food and drink information when the concentration of the spectators to the game is reduced. For example, by providing advertising information and food and drink information in anticipation of a period during which the game is temporarily interrupted, it is possible to enhance the advertising effect and improve the sales of restaurants and the like in the stadium 3.
 また、本実施形態による情報処理装置1は、注目選手のグッズに関する情報を観客のスマートフォン等に配信する機能を備えていてもよい。 Further, the information processing device 1 according to the present embodiment may have a function of distributing information on the goods of the attention player to the smartphone of the spectator or the like.
 また、本実施形態による情報処理装置1は、観客の応援スタイルに応じて、仮想通貨やポイント等の特典を観客に付与する機能を備えていてもよい。また、イベントの参加回数の情報を観客ごとに取得して、参加回数の多い観客に対して、ポイント等の特典を付与してもよい。 Further, the information processing device 1 according to the present embodiment may have a function of giving benefits such as virtual currency and points to the spectator according to the support style of the spectator. In addition, information on the number of times the event has been attended may be acquired for each spectator, and benefits such as points may be given to the spectators who have participated frequently.
 また、本実施形態による情報処理装置1は、イベント会場全体やチーム間の盛り上がりを可視化して、主催者チームにフィードバックする機能を備えていてもよい。これにより、主催者チーム内の各選手の気迫を高めることができる。 Further, the information processing device 1 according to the present embodiment may have a function of visualizing the excitement of the entire event venue or between teams and feeding back to the organizer team. As a result, it is possible to increase the spirit of each player in the organizer team.
 また、本実施形態による情報処理装置1は、盛り上がったタイミングでの観客自身と友人が写った撮影画像を、対応する観客のスマートフォンに配信する機能を備えていてもよい。 Further, the information processing device 1 according to the present embodiment may have a function of delivering a photographed image of the spectator himself and a friend at the time of excitement to the corresponding spectator's smartphone.
 また、本実施形態による情報処理装置1は、盛り上がりの少ない属性やエリア内の観客だけに所定の音源を届ける機能を備えていてもよい。 Further, the information processing apparatus 1 according to the present embodiment may have an attribute with less excitement and a function of delivering a predetermined sound source only to the audience in the area.
 また、本実施形態による情報処理装置1は、属性ごとの特徴を捉えたフェイククラウドによる仮想スタジアム3の映像を作成する機能を備えていてもよい。例えば、観客の盛り上がり度合いが高まったときに、仮想人物(アバタ)を表示させて、観客がアバタと一緒に応援することができるようにしたり、観客の盛り上がり度合いがさらに高まると、アバタと会話ができるようにしたり、ハイタッチができるようにするなどの機能を備えてもよい。この機能の具体例は、後述する図17Dや図17E等で説明する。 Further, the information processing device 1 according to the present embodiment may have a function of creating an image of a virtual stadium 3 by a fake cloud that captures the characteristics of each attribute. For example, when the degree of excitement of the audience increases, a virtual person (avatar) is displayed so that the audience can cheer with the avatar, or when the degree of excitement of the audience increases further, a conversation with the avatar will occur. It may be provided with a function such as enabling it or enabling high five. Specific examples of this function will be described with reference to FIGS. 17D and 17E described later.
 また、本実施形態による情報処理装置1は、オンライン観戦(リモート環境での観戦)をしている観客の盛り上がり度合いをイベントの開催場所に伝送するとともに、イベントの開催場所での盛り上がり度合いをオンライン観戦場所に伝送する機能を備えていてもよい。例えば、オンライン観戦場所での盛り上がり総量が基準量に達した場合に、基準量に達した旨の情報をネットワーク経由でイベントの開催場所に伝送し、イベントの開催場所で、エフェクトサウンドや視覚効果画像等による演出を行ってもよい。また、逆に、イベントの開催場所での音声量等が基準量に達した場合には、オンライン観戦場所で何らかの視覚効果の演出等を行ってもよい。 Further, the information processing device 1 according to the present embodiment transmits the degree of excitement of the spectators who are watching the game online (watching the game in a remote environment) to the venue of the event, and the degree of excitement at the venue of the event is watched online. It may have a function of transmitting to a place. For example, when the total amount of excitement at the online watching place reaches the standard amount, the information that the standard amount has been reached is transmitted to the event venue via the network, and the effect sound and visual effect image are transmitted at the event venue. You may perform the production by such as. On the contrary, when the amount of sound or the like at the venue of the event reaches the reference amount, some visual effect may be produced at the online watching place.
 また、本実施形態による情報処理装置1は、シンクロ率の高い、同じタイミングで反応している観客同士のマッチングを行う機能を備えていてもよい。観客同士は、互いに離れた場所にいることを想定しており、例えばイベントの開催場所にいる観客と、オンライン観戦をしている観客とのマッチングである。例えば、オンライン観戦している観客の応援スタイルに合致する、他の場所で応援しているアバタを表示させて、会話やハイタッチなどができるようにしてもよい。この機能の具体例も後述する図17Dや図17E等で説明する。 Further, the information processing apparatus 1 according to the present embodiment may have a function of matching spectators who are reacting at the same timing with a high synchronization rate. The spectators are assumed to be far from each other, for example, matching a spectator at the event venue with a spectator watching online. For example, an avatar cheering at another place that matches the cheering style of the spectator watching online may be displayed so that conversation and high five can be performed. Specific examples of this function will also be described with reference to FIGS. 17D and 17E described later.
 図17A~図17Hはリモート環境でのイベント参加、すなわちオンライン観戦の一例を示す図である。図17A~図17Hでは、オンライン観戦を行う場所に観客を案内して、観戦の仕方を説明した後に、実際に観戦を行う様子を示している。この場所は、観客の自宅ではなく、壁面に試合の映像を表示できる機能と、AR(Augmented Reality)技術を利用したアバタ等の3D映像を表示できる機能とを備えていることを想定している。なお、ARを利用した3D映像の代わりに、壁面に表示される映像中の一部に、イベントの開催場所等の他の場所で応援している観客や、仮想的な人物等を表示させる機能を備えていてもよい。図17A~図17Fは、サッカーの試合をオンライン観戦する例を示しているが、オンライン観戦するスポーツイベントの種類は問わない。 FIGS. 17A to 17H are diagrams showing an example of participating in an event in a remote environment, that is, watching an online game. FIGS. 17A to 17H show a state in which the spectators are guided to the place where the online spectator is to be watched, the spectators are explained how to watch the spectators, and then the spectators are actually watched. It is assumed that this place has a function to display the video of the game on the wall instead of the spectator's home, and a function to display 3D video such as avatar using AR (Augmented Reality) technology. .. In addition, instead of the 3D image using AR, a function to display the audience cheering at other places such as the event venue, virtual people, etc. in a part of the image displayed on the wall surface. May be provided. 17A to 17F show an example of watching a soccer game online, but the type of sporting event to watch online does not matter.
 まず、図17Aは、案内者が観客を観戦場所に案内する様子を示している。観客は、多面プロジェクタシステムを備えた空間(以下、ワープスクエアと呼ぶ)と呼ばれる観戦場所に案内される。ワープスクエアには、ソファがあり、ソファの前方の壁面にスタジアム3の映像が表示されるようになっている。壁面への表示は、例えばプロジェクタにて行う。また、ソファに座った観客の画像を撮影する不図示のカメラ4が設けられており、このカメラ4で撮影された映像データに基づいて、観客の盛り上がり度合いと集中度などが解析される。案内者は、観客に対してメガホンを渡して、メガホンを使って大声で応援できることを説明する。これにより、ワープスクエア内の観客は、スタジアム3の観客と同じ応援スタイルで応援することができる。 First, FIG. 17A shows how the guide guides the spectators to the spectator place. The spectators are guided to a spectator place called a space equipped with a multi-faceted projector system (hereinafter referred to as "warp square"). There is a sofa in Warp Square, and the image of Stadium 3 is displayed on the wall in front of the sofa. The display on the wall surface is performed by, for example, a projector. Further, a camera 4 (not shown) for capturing an image of the audience sitting on the sofa is provided, and the degree of excitement and concentration of the audience are analyzed based on the video data captured by the camera 4. The guide gives the audience a megaphone and explains that they can use the megaphone to cheer loudly. As a result, the spectators in Warp Square can cheer in the same cheering style as the spectators in Stadium 3.
 次に、図17Bに示すように、観客はワープスクエア内のソファに座って、リモートコントローラの実行ボタンを押すか、あるいはユーザの動作(例えば、ソファに座るとか、壁面に視線を向けるなど)を自動認識する行動センシングにより、前方の壁面にスタジアム3の映像が映し出される。 Next, as shown in FIG. 17B, the spectator sits on the couch in Warp Square and presses the execute button on the remote controller, or the user's actions (eg, sitting on the couch, looking at the wall, etc.). The image of the stadium 3 is projected on the front wall surface by the automatic recognition behavior sensing.
 次に、図17Cに示すように、試合が始まって観客が応援すると、観客の盛り上がり度合いと集中度が高くなるほど、又は/及び、イベントで提供されるコンテンツの盛り上がりが観客の行動から推定される場合、壁面に表示される映像のサイズが大きくなる。観客は、声を上げて応援をするだけでなく、メガホンを大きく振り回して応援するほど、映像のサイズが大きくなる。図17Cの右側の図に示すように、盛り上がり度合いが最高潮に達すると、ソファーの前方の壁面全体にスタジアム3の試合の映像が映し出される。このように、観客の盛り上がり度合いや集中度で映像サイズを可変させることで、より大きな映像サイズでの観戦を楽しみたい観客の盛り上がり度合いをより高めることができる。 Next, as shown in FIG. 17C, when the match starts and the spectators cheer, the higher the degree of excitement and concentration of the spectators, and / or the excitement of the content provided at the event is estimated from the behavior of the spectators. In that case, the size of the image displayed on the wall surface becomes large. The larger the audience, the larger the size of the image, the more they cheer by swinging the megaphone as well as raising their voices. As shown in the figure on the right side of FIG. 17C, when the degree of excitement reaches the climax, the image of the game of the stadium 3 is projected on the entire wall surface in front of the sofa. In this way, by varying the image size according to the degree of excitement and concentration of the spectators, it is possible to further increase the degree of excitement of the spectators who want to enjoy watching the game with a larger image size.
 次に、図17Dの左側の図に示すように、贔屓のチームが得点を上げた場合には、AR技術を利用して、仮想人物(アバタ)を3D画像で表示させて、アバタとの間で会話ができるようにしてもよい。また、図17Dの右側の図に示すように、贔屓のチームが得点を上げた場合には、メガホンを激しく振ることで、紙吹雪の画像を表示させるとともに、「ゴーーーーーール」の音声を流すような機能を設けてもよい。アバタは、スタジアム3で同じような応援スタイルをしている観客の姿を反映させた仮想人物でもよい。これにより、スタジアム3にいる観客との一体感や連帯感を高めることができる。 Next, as shown in the figure on the left side of FIG. 17D, when the favored team raises the score, the virtual person (avatar) is displayed as a 3D image using AR technology, and the conversation with the avatar is performed. You may be able to have a conversation with. In addition, as shown in the figure on the right side of FIG. 17D, when the favored team raises the score, the image of confetti is displayed by shaking the megaphone violently, and the voice of "gooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooo A function may be provided. The avatar may be a virtual person who reflects the appearance of an spectator who has a similar cheering style at Stadium 3. As a result, it is possible to enhance the sense of unity and solidarity with the spectators in the stadium 3.
 贔屓チームの選手が得点を上げた場合には、図17Eに示すように、アバタも大喜びするような3D画像を表示させてもよい。図17Eの右側の図に示すように、アバタがメガホンを介してのハイタッチを求める仕草をする例を示している。アバタの動作に従ってメガホンを介してのハイタッチをすると、図17Fの左側の図に示すように、ハイタッチをした場所でエフェクトサウンドが響くような演出や、視覚効果画像の表示を行ってもよい。 If a player on the favored team scores a high score, as shown in FIG. 17E, a 3D image that makes the avatar happy may be displayed. As shown in the figure on the right side of FIG. 17E, an example is shown in which an avatar makes a gesture for a high five via a megaphone. When a high five is performed via a megaphone according to the operation of the avatar, as shown in the figure on the left side of FIG. 17F, an effect sound may be produced or a visual effect image may be displayed at the place where the high five is performed.
 その後、試合の中断期間になると、観客はソファに座って応援を止めるため、それに応じて、壁面に表示される映像のサイズも、図17Fの右側の図のように小さくなる。 After that, when the game is interrupted, the spectators sit on the sofa and stop cheering, so the size of the image displayed on the wall becomes smaller as shown in the figure on the right side of Fig. 17F.
 その後、試合が再開されると、上述した図17C~図17Fと同様の映像を表示させる。映像の表示のさせ方は、図17C~図17Fに図示したものに限定されない。例えば、図17Gの左側の図のように、試合が盛り上がったときに、画面を縦方向に二分割して、上側には試合の映像を表示し、下側にはスタジアム3の観客席で応援しているかのようなパノラマ映像を表示してもよい。この状態でメガホンを振って応援すると、図17Gの右側の図のように、スタジアム3の観客のアバタのシルエットが表示されて、アバタの応援の音声が聞こえるようにしてもよい。 After that, when the match is resumed, the same images as those in FIGS. 17C to 17F described above are displayed. The method of displaying the image is not limited to that shown in FIGS. 17C to 17F. For example, as shown in the figure on the left side of Fig. 17G, when the game is lively, the screen is divided into two in the vertical direction, the video of the game is displayed on the upper side, and the audience seats of the stadium 3 are on the lower side. You may display a panoramic image as if you were doing it. When cheering by shaking the megaphone in this state, the silhouette of the avatar of the spectator in the stadium 3 may be displayed and the voice of the avatar cheering may be heard as shown in the figure on the right side of FIG. 17G.
 ゴール直前の盛り上がり度合いが高まった状態では、図17Hの左側の図のように、隣のアバタから話しかけられて、それに対して音声で答えると、「いいね」のアイコンを表示させてもよい。ゴールすると、図17Hの右側の図に示すように、「ゴーーーーーール」の音声が聞こえ、メガホンの動きに同期して、紙吹雪の映像が表示されてもよい。 In a state where the degree of excitement just before the goal is high, as shown in the figure on the left side of Fig. 17H, if the next avatar speaks to it and answers it by voice, the "Like" icon may be displayed. When the goal is reached, as shown in the figure on the right side of FIG. 17H, the voice of "gooooooo" may be heard, and the image of confetti may be displayed in synchronization with the movement of the megaphone.
 上述した実施形態では、主にバスケットボールのスポーツイベントを観戦する例を主に説明したが、バスケットボールとは映像データの撮影条件が異なっていたり、観客の行動が異なるスポーツがある。この場合、図5、図10、図11及び図14等に示した処理を必要に応じて変更する必要がある。 In the above-described embodiment, an example of mainly watching a basketball sporting event was mainly described, but there are sports in which the shooting conditions of video data are different from those of basketball and the behavior of the spectators is different. In this case, it is necessary to change the processes shown in FIGS. 5, 10, 11, 14, and the like as necessary.
 図18は種々のスポーツとバスケットボールとの類似性を示した図である。図18の横軸は、バスケットボールと撮影条件が異なる度合いを示しており、右側ほど撮影条件がより異なることを示している。撮影条件とは、屋内/屋外と、暗い/明るい、広い/狭い、人数が多い/少ないなどである。図18の縦軸は、バスケットボールと観客の行動が異なる度合いを示しており、下側ほど観客の行動がより異なることを示している。 FIG. 18 is a diagram showing the similarity between various sports and basketball. The horizontal axis of FIG. 18 indicates the degree to which the shooting conditions are different from those of basketball, and the right side indicates that the shooting conditions are more different. The shooting conditions are indoor / outdoor, dark / bright, wide / narrow, and large / small number of people. The vertical axis of FIG. 18 shows the degree to which the behaviors of the basketball and the spectators are different, and the lower side shows that the behaviors of the spectators are more different.
 図18からわかるように、撮影条件と観客の行動を総合的に勘案すると、アイスホッケが最もバスケットボールに類似しており、ゴルフが最もバスケットボールと相違している。よって、ゴルフについては、上述した図5、図10、図11及び図14等に示した処理を大幅に変更する必要がありうる。また、ゴルフ以外のスポーツについても、図5、図10、図11及び図14等の処理を必要に応じて変更する必要性が生じうる。 As can be seen from FIG. 18, when the shooting conditions and the behavior of the spectators are comprehensively considered, ice hockey is most similar to basketball, and golf is most different from basketball. Therefore, for golf, it may be necessary to drastically change the processes shown in FIGS. 5, 10, 11 and 14 described above. Further, for sports other than golf, it may be necessary to change the processing of FIGS. 5, 10, 11 and 14 as necessary.
 本実施形態による情報処理装置1及び情報処理システム2は、スポーツイベント以外の種々のイベントにも適用可能である。図19はスポーツ以外の種々のイベントとバスケットボールとの類似性を示した図である。図18と同様に、図19の横軸はバスケットボールと撮影条件が異なる度合いを示し、縦軸はバスケットボールと観客の行動が異なる度合いを示している。 The information processing device 1 and the information processing system 2 according to the present embodiment can be applied to various events other than sporting events. FIG. 19 is a diagram showing the similarity between various non-sports events and basketball. Similar to FIG. 18, the horizontal axis of FIG. 19 indicates the degree to which the shooting conditions differ from that of basketball, and the vertical axis indicates the degree to which the behavior of the basketball and the spectator differ.
 図19からわかるように、撮影条件と観客の行動を総合的に勘案すると、音楽(ライブ)イベントが最もバスケットボールに類似しており、映画館での映画鑑賞が最もバスケットボールと相違している。よって、映画館については、上述した図5、図10、図11及び図14等に示した処理を大幅に変更する必要がありうる。また、映画館以外のイベントについても、図5、図10、図11及び図14等の処理を必要に応じて変更する必要性が生じうる。 As can be seen from FIG. 19, when the shooting conditions and the behavior of the audience are comprehensively considered, the music (live) event is most similar to basketball, and watching a movie in a movie theater is the most different from basketball. Therefore, for movie theaters, it may be necessary to drastically change the processes shown in FIGS. 5, 10, 11, 14, and the like described above. Further, for events other than movie theaters, it may be necessary to change the processing of FIGS. 5, 10, 11 and 14 as necessary.
 このように、本実施形態では、センシング情報に基づいて、イベントに参加するユーザの内的状態を推測するため、ユーザの属性及び行動の少なくとも一方を推定できるとともに、ユーザを複数のクラスタに分類することができる。また、本実施形態によれば、イベントに参加するユーザの属性や行動の推定結果と複数のクラスタの分類結果に基づいて、ユーザやユーザのグループをタグ付けすることができ、ユーザやユーザのグループに適した情報を提供できる。これにより、例えば、ホームチームのファン、アウェイチームのファン、初心者のそれぞれに見合った情報を提供でき、スポーツ観戦の魅力をより向上できる。 As described above, in the present embodiment, since the internal state of the user participating in the event is estimated based on the sensing information, at least one of the user's attributes and behavior can be estimated, and the user is classified into a plurality of clusters. be able to. Further, according to the present embodiment, it is possible to tag a user or a group of users based on the estimation result of the attributes and behaviors of the users participating in the event and the classification result of a plurality of clusters, and the user or the group of users can be tagged. Can provide information suitable for. As a result, for example, it is possible to provide information suitable for each of the fans of the home team, the fans of the away team, and the beginners, and it is possible to further improve the attractiveness of watching sports.
 さらに、ユーザの盛り上がり度合いと集中度に基づいて、盛り上がり度合いを高めるための演出を行うことができる。また、イベントの経過情報と、ユーザの盛り上がり度合いや集中度の時間変化を示す情報とを表示する運営者用ツール画面40を設けることで、従来は主観的にしか判断できなかった演出の効果を詳細かつ客観的に分析及び評価することができ、より魅力的な演出の制作に役立てることができる。 Furthermore, it is possible to perform an effect to increase the degree of excitement based on the degree of excitement and concentration of the user. In addition, by providing the operator tool screen 40 that displays the progress information of the event and the information indicating the time change of the degree of excitement and the degree of concentration of the user, the effect of the effect that could be judged only subjectively in the past can be achieved. It can be analyzed and evaluated in detail and objectively, and can be useful for producing more attractive productions.
 さらに、本実施形態では、リモート環境でイベントに参加するユーザの内的状態を推測して、イベント会場にいるかのような演出を提供することができる。例えば、リモート環境のユーザの盛り上がり度合いに応じて演出方法を変化させて、盛り上がり度合いをより高めるように仕向けたり、イベント会場にいるかのような演出を行ったり、イベント会場にいるユーザをアバタとして表示させて、アバタとの会話やハイタッチなどができるようにすることで、離れた場所で応援しているファン同士での一体感や連帯感を高めて、イベントへの参加意欲を高めることができる。 Further, in the present embodiment, it is possible to infer the internal state of the user who participates in the event in the remote environment and provide an effect as if he / she is at the event venue. For example, the production method can be changed according to the degree of excitement of the user in the remote environment to make the degree of excitement higher, or the effect as if at the event venue can be performed, or the user at the event venue can be displayed as an avatar. By making it possible to have conversations and high fives with Avata, it is possible to increase the sense of unity and solidarity among fans who are cheering at a remote location, and to increase the motivation to participate in the event.
 上述した実施形態で説明した情報処理装置の少なくとも一部は、ハードウェアで構成してもよいし、ソフトウェアで構成してもよい。ソフトウェアで構成する場合には、情報処理装置の少なくとも一部の機能を実現するプログラムをフレキシブルディスクやCD-ROM等の記録媒体に収納し、コンピュータに読み込ませて実行させてもよい。記録媒体は、磁気ディスクや光ディスク等の着脱可能なものに限定されず、ハードディスク装置やメモリなどの固定型の記録媒体でもよい。 At least a part of the information processing apparatus described in the above-described embodiment may be configured by hardware or software. When configured by software, a program that realizes at least a part of the functions of the information processing apparatus may be stored in a recording medium such as a flexible disk or a CD-ROM, read by a computer, and executed. The recording medium is not limited to a removable one such as a magnetic disk or an optical disk, and may be a fixed recording medium such as a hard disk device or a memory.
 また、情報処理装置の少なくとも一部の機能を実現するプログラムを、インターネット等の通信回線(無線通信も含む)を介して頒布してもよい。さらに、同プログラムを暗号化したり、変調をかけたり、圧縮した状態で、インターネット等の有線回線や無線回線を介して、あるいは記録媒体に収納して頒布してもよい。 Further, a program that realizes at least a part of the functions of the information processing device may be distributed via a communication line (including wireless communication) such as the Internet. Further, the program may be encrypted, modulated, compressed, and distributed via a wired line or a wireless line such as the Internet, or stored in a recording medium.
 なお、本技術は以下のような構成を取ることができる。
 (1)ユーザのセンシング情報に基づいて特徴量を抽出する特徴量抽出部と、
 前記特徴量に基づいて、前記ユーザの属性及び行動の少なくとも一方を推定する第1推定部と、
 前記第1推定部による推定に基づいて、前記ユーザ又は複数のユーザからなるグループを単位として、複数のクラスタに分類するクラスタリング部と、
 前記第1推定部による推定と前記クラスタリング部による分類との少なくとも一方に基づいて、所定の情報処理を行う情報処理部と、を備える、情報処理装置。
 (2)前記センシング情報は、撮像装置の撮像画像を含んでおり、
 前記クラスタリング部は、前記撮像画像の解析結果に基づいて前記複数のクラスタに分類する、(1)に記載の情報処理装置。
 (3)前記撮像画像は、前記ユーザの画像を含んでおり、
 前記特徴量抽出部は、前記ユーザの顔、姿勢、体の動き、及び骨格情報の少なくとも一つを含む前記特徴量を抽出する、(2)に記載の情報処理装置。
 (4)前記特徴量抽出部は、音響データ、物体認識、及び周波数解析情報の少なくとも一つに基づいて前記特徴量を抽出する、(1)乃至(3)のいずれか一項に記載の情報処理装置。
 (5)前記ユーザが参加するイベントの経過情報を取得するイベント情報取得部をさらに備え、
 前記第1推定部は、前記特徴量と前記イベントの経過情報とに基づいて、前記イベントに参加するユーザの属性及び行動の少なくとも一方を推定する、(1)乃至(4)のいずれか一項に記載の情報処理装置。
 (6)前記第1推定部による推定に基づいて、前記ユーザ又は前記グループを単位として、タグ情報を付与するタグ付け部を備える、(1)乃至(5)のいずれか一項に記載の情報処理装置。
 (7)前記情報処理部は、同一の前記タグ情報が付与された前記ユーザ又は前記グループに対して、前記タグ情報に基づいた情報を提供する、(6)に記載の情報処理装置。
 (8)前記情報処理部は、前記ユーザの属性及び行動の少なくとも一方に応じた情報提供及び情報交換の少なくとも一方を行う、(7)に記載の情報処理装置。
 (9)前記ユーザを撮影した画像に、前記センシング情報に基づいて決定される前記ユーザの内的状態を示す識別子を付加した状況画像を生成する状況画像生成部を備える、(7)又は(8)に記載の情報処理装置。
 (10)前記状況画像生成部は、前記ユーザが参加するイベントの経過情報と、前記ユーザの盛り上がり度合い及び集中度の少なくとも一方に関する情報とを含む前記状況画像を生成する、(9)に記載の情報処理装置。
 (11)前記第1推定部は、前記センシング情報に基づいて、前記ユーザの盛り上がり度合い及び集中度の少なくとも一方を含む内的状態を推定し、
 前記クラスタリング部は、前記特徴量及び前記内的状態に基づいて、前記複数のクラスタに分類する、(1)乃至(10)のいずれか一項に記載の情報処理装置。
 (12)前記クラスタリング部は、イベントの経過情報に応じた前記内的状態の変化に基づいて、前記複数のクラスタに分類する、(11)に記載の情報処理装置。
 (13)イベントの会場にいるユーザと、リモート環境から前記イベントに参加するユーザとの少なくとも一方に関する前記センシング情報を取得するセンシング情報取得部をさらに備える、(1)乃至(12)のいずれか一項に記載の情報処理装置。
 (14)前記クラスタリング部は、前記イベントにリモート環境から参加するユーザ又は複数のユーザからなるグループを単位として、前記複数のクラスタに分類する、
13)に記載の情報処理装置。
 (15)前記センシング情報に基づいて、前記イベントにリモート環境から参加するユーザの盛り上がり度合い及び集中度の少なくとも一つを含む内的状態を推定する第2推定部と、
 前記第2推定部で推定された前記内的状態に基づいて、前記イベントに関する情報を表示する表示領域のサイズを調整する表示制御部と、備える、(13)又は(14)に記載の情報処理装置。
 (16)前記表示制御部は、前記イベントにリモート環境から参加するユーザが見る表示部に、前記ユーザの盛り上がり度合い及び集中度の少なくとも一方が高まることに応じて、前記イベントの開催場所の観客席との一体感を高める映像を表示させる、(15)に記載の情報処理装置。
 (17)前記表示制御部は、前記イベントにリモート環境から参加するユーザの内的状態が所定の条件を満たすときに、前記ユーザが視認可能な範囲内に、前記所定の条件に応じた情報提供画像及び視覚効果画像の少なくとも一方を表示させる、(15)乃至(16)のいずれか一項に記載の情報処理装置。
 (18)前記視覚効果画像は、前記イベントの会場に参加しており、内的状態が前記所定の条件を満たす別のユーザの仮想人物画像である、(17)に記載の情報処理装置。
 (19)前記イベントにリモート環境から参加するユーザは、前記所定の条件を満たすときに、前記仮想人物画像を介して、前記仮想人物画像に対応する前記別の人物と情報交換を行う情報交換部を備える、(18)に記載の情報処理装置。
 (20)ユーザのセンシング情報に基づいて特徴量を抽出するステップと、
 前記特徴量に基づいて、前記ユーザの属性及び行動の少なくとも一方を推定するステップと、
 前記推定に基づいて、前記ユーザ又は複数のユーザからなるグループを単位として、複数のクラスタに分類するステップと、
 前記推定と前記複数のクラスタとの少なくとも一方に基づいて、所定の情報処理を行うステップと、を備える、情報処理方法。
The present technology can have the following configurations.
(1) A feature amount extraction unit that extracts a feature amount based on the user's sensing information,
A first estimation unit that estimates at least one of the user's attributes and behavior based on the feature amount,
A clustering unit that classifies the user or a group consisting of a plurality of users into a plurality of clusters based on the estimation by the first estimation unit.
An information processing apparatus including an information processing unit that performs predetermined information processing based on at least one of estimation by the first estimation unit and classification by the clustering unit.
(2) The sensing information includes an image captured by the image pickup device.
The information processing apparatus according to (1), wherein the clustering unit is classified into the plurality of clusters based on the analysis result of the captured image.
(3) The captured image includes the image of the user.
The information processing apparatus according to (2), wherein the feature amount extraction unit extracts the feature amount including at least one of the user's face, posture, body movement, and skeletal information.
(4) The information according to any one of (1) to (3), wherein the feature amount extraction unit extracts the feature amount based on at least one of acoustic data, object recognition, and frequency analysis information. Processing device.
(5) Further provided with an event information acquisition unit for acquiring progress information of the event in which the user participates.
The first estimation unit estimates at least one of the attributes and behaviors of the user participating in the event based on the feature amount and the progress information of the event, any one of (1) to (4). The information processing device described in.
(6) The information according to any one of (1) to (5), comprising a tagging unit for adding tag information in units of the user or the group based on the estimation by the first estimation unit. Processing equipment.
(7) The information processing apparatus according to (6), wherein the information processing unit provides information based on the tag information to the user or the group to which the same tag information is attached.
(8) The information processing apparatus according to (7), wherein the information processing unit provides at least one of information provision and information exchange according to at least one of the user's attributes and actions.
(9) (7) or (8) or (8) or (8) or (8), which comprises a situation image generation unit for generating a situation image in which an identifier indicating the internal state of the user, which is determined based on the sensing information, is added to the image taken by the user. ). The information processing device.
(10) The situation image generation unit, according to (9), which generates the situation image including the progress information of the event in which the user participates and the information about at least one of the degree of excitement and the degree of concentration of the user. Information processing device.
(11) The first estimation unit estimates an internal state including at least one of the degree of excitement and the degree of concentration of the user based on the sensing information.
The information processing apparatus according to any one of (1) to (10), wherein the clustering unit is classified into the plurality of clusters based on the feature amount and the internal state.
(12) The information processing apparatus according to (11), wherein the clustering unit is classified into the plurality of clusters based on the change in the internal state according to the progress information of the event.
(13) Any one of (1) to (12) further comprising a sensing information acquisition unit that acquires the sensing information regarding at least one of the user at the event venue and the user participating in the event from the remote environment. The information processing device described in the section.
(14) The clustering unit classifies the event into the plurality of clusters in units of users who participate in the event from a remote environment or a group consisting of a plurality of users.
The information processing apparatus according to 13).
(15) A second estimation unit that estimates an internal state including at least one of the degree of excitement and the degree of concentration of users who participate in the event from a remote environment based on the sensing information.
The information processing according to (13) or (14), comprising a display control unit that adjusts the size of a display area for displaying information about the event based on the internal state estimated by the second estimation unit. Device.
(16) The display control unit is a seat of the audience at the venue of the event in response to an increase in at least one of the degree of excitement and the degree of concentration of the user on the display unit viewed by a user who participates in the event from a remote environment. The information processing apparatus according to (15), which displays an image that enhances a sense of unity with.
(17) When the internal state of a user who participates in the event from a remote environment satisfies a predetermined condition, the display control unit provides information according to the predetermined condition within a range visible to the user. The information processing apparatus according to any one of (15) to (16), which displays at least one of an image and a visual effect image.
(18) The information processing apparatus according to (17), wherein the visual effect image is a virtual person image of another user who participates in the venue of the event and whose internal state satisfies the predetermined condition.
(19) An information exchange unit in which a user who participates in the event from a remote environment exchanges information with the other person corresponding to the virtual person image via the virtual person image when the predetermined condition is satisfied. The information processing apparatus according to (18).
(20) A step of extracting a feature amount based on the user's sensing information,
A step of estimating at least one of the user's attributes and behavior based on the feature amount, and
Based on the estimation, the step of classifying into a plurality of clusters in units of the user or a group consisting of a plurality of users, and
An information processing method comprising a step of performing predetermined information processing based on at least one of the estimation and the plurality of clusters.
 本開示の態様は、上述した個々の実施形態に限定されるものではなく、当業者が想到しうる種々の変形も含むものであり、本開示の効果も上述した内容に限定されない。すなわち、特許請求の範囲に規定された内容およびその均等物から導き出される本開示の概念的な思想と趣旨を逸脱しない範囲で種々の追加、変更および部分的削除が可能である。 The aspects of the present disclosure are not limited to the individual embodiments described above, but also include various modifications that can be conceived by those skilled in the art, and the effects of the present disclosure are not limited to the above-mentioned contents. That is, various additions, changes and partial deletions are possible without departing from the conceptual idea and purpose of the present disclosure derived from the contents specified in the claims and their equivalents.
 1 情報処理装置、2 情報処理システム、3 コート(スタジアム)、4 カメラ、5 ネットワーク機器、6 処理サーバ、7 DBサーバ、8 携帯端末、9 配信サーバ、11 センシング情報取得部、12 特徴量抽出部、13 第1推定部、14 クラスタリング部、15 情報処理部、16 イベント情報取得部、17 タグ付け部、18 状況画像生成部、19 第2推定部、20 表示制御部、21 情報交換部、22 クラウドストレージ、30 運営者サーバ(ツール選択用サーバ) 1 Information processing device, 2 Information processing system, 3 Court (stadium), 4 Camera, 5 Network equipment, 6 Processing server, 7 DB server, 8 Mobile terminal, 9 Distribution server, 11 Sensing information acquisition unit, 12 Feature quantity extraction unit , 13 1st estimation unit, 14 clustering unit, 15 information processing unit, 16 event information acquisition unit, 17 tagging unit, 18 situation image generation unit, 19 second estimation unit, 20 display control unit, 21 information exchange unit, 22 Cloud storage, 30 operator server (server for tool selection)

Claims (20)

  1. ユーザのセンシング情報に基づいて特徴量を抽出する特徴量抽出部と、
     前記特徴量に基づいて、前記ユーザの属性及び行動の少なくとも一方を推定する第1推定部と、
     前記第1推定部による推定に基づいて、前記ユーザ又は複数のユーザからなるグループを単位として、複数のクラスタに分類するクラスタリング部と、
     前記第1推定部による推定と前記クラスタリング部による分類との少なくとも一方に基づいて、所定の情報処理を行う情報処理部と、を備える、情報処理装置。
    A feature amount extraction unit that extracts feature amounts based on the user's sensing information,
    A first estimation unit that estimates at least one of the user's attributes and behavior based on the feature amount,
    A clustering unit that classifies the user or a group consisting of a plurality of users into a plurality of clusters based on the estimation by the first estimation unit.
    An information processing apparatus including an information processing unit that performs predetermined information processing based on at least one of estimation by the first estimation unit and classification by the clustering unit.
  2.  前記センシング情報は、撮像装置の撮像画像を含んでおり、
     前記クラスタリング部は、前記撮像画像の解析結果に基づいて前記複数のクラスタに分類する、請求項1に記載の情報処理装置。
    The sensing information includes an image captured by the image pickup device.
    The information processing apparatus according to claim 1, wherein the clustering unit is classified into the plurality of clusters based on the analysis result of the captured image.
  3.  前記撮像画像は、前記ユーザの画像を含んでおり、
     前記特徴量抽出部は、前記ユーザの顔、姿勢、体の動き、骨格情報、音声の少なくとも一つを含む前記特徴量を抽出する、請求項2に記載の情報処理装置。
    The captured image includes the image of the user, and the captured image includes the image of the user.
    The information processing device according to claim 2, wherein the feature amount extraction unit extracts the feature amount including at least one of the user's face, posture, body movement, skeletal information, and voice.
  4.  前記特徴量抽出部は、音響データ、物体認識、及び周波数解析情報の少なくとも一つに基づいて前記特徴量を抽出する、請求項1に記載の情報処理装置。 The information processing device according to claim 1, wherein the feature amount extraction unit extracts the feature amount based on at least one of acoustic data, object recognition, and frequency analysis information.
  5.  前記ユーザが参加するイベントの経過情報を取得するイベント情報取得部をさらに備え、
     前記第1推定部は、前記特徴量と前記イベントの経過情報とに基づいて、前記イベントに参加する前記ユーザの属性及び行動の少なくとも一方を推定する、請求項1に記載の情報処理装置。
    Further, an event information acquisition unit for acquiring progress information of an event in which the user participates is provided.
    The information processing apparatus according to claim 1, wherein the first estimation unit estimates at least one of the attributes and actions of the user who participates in the event based on the feature amount and the progress information of the event.
  6.  前記第1推定部による推定に基づいて、前記ユーザ又は前記グループを単位として、タグ情報を付与するタグ付け部を備える、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, further comprising a tagging unit for adding tag information in units of the user or the group based on the estimation by the first estimation unit.
  7.  前記情報処理部は、同一の前記タグ情報が付与された前記ユーザ又は前記グループに対して、前記タグ情報に基づいた情報を提供する、請求項6に記載の情報処理装置。 The information processing device according to claim 6, wherein the information processing unit provides information based on the tag information to the user or the group to which the same tag information is attached.
  8.  前記情報処理部は、ユーザの属性及び行動の少なくとも一方に応じた情報提供及び情報交換の少なくとも一方を行う、請求項7に記載の情報処理装置。 The information processing device according to claim 7, wherein the information processing unit provides at least one of information provision and information exchange according to at least one of a user's attributes and actions.
  9.  前記ユーザを撮影した画像に、前記センシング情報に基づいて決定される前記ユーザの内的状態を示す識別子を付加した状況画像を生成する状況画像生成部を備える、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, further comprising a situation image generation unit that generates a situation image in which an identifier indicating the internal state of the user, which is determined based on the sensing information, is added to an image taken by the user. ..
  10.  前記状況画像生成部は、前記ユーザが参加するイベントの経過情報と、前記ユーザの盛り上がり度合い及び集中度の少なくとも一方に関する情報とを含む前記状況画像を生成する、請求項9に記載の情報処理装置。 The information processing apparatus according to claim 9, wherein the situation image generation unit generates the situation image including progress information of an event in which the user participates and information on at least one of the degree of excitement and the degree of concentration of the user. ..
  11.  前記第1推定部は、前記センシング情報に基づいて、前記ユーザの盛り上がり度合い及び集中度の少なくとも一方を含む内的状態を推定し、
     前記クラスタリング部は、前記特徴量及び前記内的状態に基づいて、前記複数のクラスタに分類する、請求項1に記載の情報処理装置。
    The first estimation unit estimates an internal state including at least one of the degree of excitement and the degree of concentration of the user based on the sensing information.
    The information processing apparatus according to claim 1, wherein the clustering unit is classified into the plurality of clusters based on the feature amount and the internal state.
  12.  前記クラスタリング部は、イベントの経過情報に応じた前記内的状態の変化に基づいて、前記複数のクラスタに分類する、請求項11に記載の情報処理装置。 The information processing device according to claim 11, wherein the clustering unit is classified into the plurality of clusters based on the change in the internal state according to the progress information of the event.
  13.  イベントの会場にいるユーザと、リモート環境から前記イベントに参加するユーザとの少なくとも一方に関する前記センシング情報を取得するセンシング情報取得部をさらに備える、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, further comprising a sensing information acquisition unit that acquires the sensing information regarding at least one of a user at an event venue and a user participating in the event from a remote environment.
  14.  前記クラスタリング部は、前記イベントにリモート環境から参加するユーザ又は複数のユーザからなるグループを単位として、前記複数のクラスタに分類する、請求項13に記載の情報処理装置。 The information processing device according to claim 13, wherein the clustering unit is classified into the plurality of clusters in units of a user who participates in the event from a remote environment or a group consisting of a plurality of users.
  15.  前記センシング情報に基づいて、前記イベントにリモート環境から参加するユーザの盛り上がり度合い及び集中度の少なくとも一つを含む内的状態を推定する第2推定部と、
     前記第2推定部で推定された前記内的状態に基づいて、前記イベントに関する情報を表示する表示領域のサイズを調整する表示制御部と、備える、請求項13に記載の情報処理装置。
    A second estimation unit that estimates an internal state including at least one of the degree of excitement and concentration of users who participate in the event from a remote environment based on the sensing information.
    The information processing apparatus according to claim 13, further comprising a display control unit that adjusts the size of a display area for displaying information about the event based on the internal state estimated by the second estimation unit.
  16.  前記表示制御部は、前記イベントにリモート環境から参加するユーザが見る表示部に、前記ユーザの盛り上がり度合い及び集中度の少なくとも一方が高まることに応じて、前記イベントの開催場所の観客席との一体感を高める映像を表示させる、請求項15に記載の情報処理装置。 The display control unit is one with the audience seats of the venue of the event according to the increase in at least one of the degree of excitement and the degree of concentration of the user on the display unit viewed by the user who participates in the event from the remote environment. The information processing apparatus according to claim 15, which displays an image that enhances the experience.
  17.  前記表示制御部は、前記イベントにリモート環境から参加するユーザの内的状態が所定の条件を満たすときに、前記ユーザが視認可能な範囲内に、前記所定の条件に応じた情報提供画像及び視覚効果画像の少なくとも一方を表示させる、請求項15に記載の情報処理装置。 When the internal state of a user who participates in the event from a remote environment satisfies a predetermined condition, the display control unit can provide an information providing image and a visual sense according to the predetermined condition within a range visible to the user. The information processing apparatus according to claim 15, wherein at least one of the effect images is displayed.
  18.  前記視覚効果画像は、前記イベントの会場に参加しており、内的状態が前記所定の条件を満たす別のユーザの仮想人物画像である、請求項17に記載の情報処理装置。 The information processing device according to claim 17, wherein the visual effect image is a virtual person image of another user who participates in the venue of the event and whose internal state satisfies the predetermined condition.
  19.  前記イベントにリモート環境から参加するユーザは、前記所定の条件を満たすときに、前記仮想人物画像を介して、前記仮想人物画像に対応する前記別の人物と情報交換を行う情報交換部を備える、請求項18に記載の情報処理装置。 A user who participates in the event from a remote environment includes an information exchange unit that exchanges information with the other person corresponding to the virtual person image via the virtual person image when the predetermined condition is satisfied. The information processing apparatus according to claim 18.
  20.  ユーザのセンシング情報に基づいて特徴量を抽出するステップと、
     前記特徴量に基づいて、前記ユーザの属性及び行動の少なくとも一方を推定するステップと、
     前記推定に基づいて、前記ユーザ又は複数のユーザからなるグループを単位として、複数のクラスタに分類するステップと、
     前記推定と前記複数のクラスタとの少なくとも一方に基づいて、所定の情報処理を行うステップと、を備える、情報処理方法。
    Steps to extract features based on user sensing information,
    A step of estimating at least one of the user's attributes and behavior based on the feature amount, and
    Based on the estimation, the step of classifying into a plurality of clusters in units of the user or a group consisting of a plurality of users, and
    An information processing method comprising a step of performing predetermined information processing based on at least one of the estimation and the plurality of clusters.
PCT/JP2021/040879 2020-11-16 2021-11-05 Information processing device and information processing method WO2022102550A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2020-190541 2020-11-16
JP2020190541A JP2023181568A (en) 2020-11-16 2020-11-16 Information processing apparatus and information processing method

Publications (1)

Publication Number Publication Date
WO2022102550A1 true WO2022102550A1 (en) 2022-05-19

Family

ID=81602276

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2021/040879 WO2022102550A1 (en) 2020-11-16 2021-11-05 Information processing device and information processing method

Country Status (2)

Country Link
JP (1) JP2023181568A (en)
WO (1) WO2022102550A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019012974A1 (en) * 2017-07-14 2019-01-17 シャープ株式会社 Information processing apparatus, terminal apparatus, information providing system, program for causing computer to function as information processing apparatus, program for causing computer to function as terminal apparatus, and method for controlling information processing apparatus
JP2020042557A (en) * 2018-09-11 2020-03-19 トヨタ紡織株式会社 Excitement promotion system
JP2020144622A (en) * 2019-03-06 2020-09-10 株式会社Creator’s NEXT Information processing device, information processing method, and program

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019012974A1 (en) * 2017-07-14 2019-01-17 シャープ株式会社 Information processing apparatus, terminal apparatus, information providing system, program for causing computer to function as information processing apparatus, program for causing computer to function as terminal apparatus, and method for controlling information processing apparatus
JP2020042557A (en) * 2018-09-11 2020-03-19 トヨタ紡織株式会社 Excitement promotion system
JP2020144622A (en) * 2019-03-06 2020-09-10 株式会社Creator’s NEXT Information processing device, information processing method, and program

Also Published As

Publication number Publication date
JP2023181568A (en) 2023-12-25

Similar Documents

Publication Publication Date Title
US11660531B2 (en) Scaled VR engagement and views in an e-sports event
US11538213B2 (en) Creating and distributing interactive addressable virtual content
JP6965896B2 (en) Display control system and display control method
US11218783B2 (en) Virtual interactive audience interface
KR101711488B1 (en) Method and System for Motion Based Interactive Service
CN102656542B (en) Camera navigation for presentations
US8177611B2 (en) Scheme for inserting a mimicked performance into a scene and providing an evaluation of same
CN112543669A (en) Discovery and detection of events in interactive content
CN111971097A (en) Online championship integration
JP2020039029A (en) Video distribution system, video distribution method, and video distribution program
JP2017504457A (en) Method and system for displaying a portal site containing user selectable icons on a large display system
US9025832B2 (en) Automated sensor driven friending
JP7202935B2 (en) Attention level calculation device, attention level calculation method, and attention level calculation program
US20210342886A1 (en) Virtual Advertising Database with Command Input Data
US11941177B2 (en) Information processing device and information processing terminal
CN110507996A (en) Keep the user experience in gaming network personalized
WO2022102550A1 (en) Information processing device and information processing method
US20230033892A1 (en) Information processing device and information processing terminal
US20230009322A1 (en) Information processing device, information processing terminal, and program
WO2021070733A1 (en) Information processing device, information processing method, and program
US20230038998A1 (en) Information processing device, information processing terminal, and program
JP7027300B2 (en) Information processing equipment, information processing methods and information processing programs
US20230381673A1 (en) eSPORTS SPECTATOR ONBOARDING
JP2015219423A (en) Try-on pseudo experience system, control method of try-on pseudo experience, and computer program
WO2023091564A1 (en) System and method for provision of personalized multimedia avatars that provide studying companionship

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21891796

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21891796

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP