WO2015159487A1 - Image delivery method, image reception method, server, terminal apparatus, and image delivery system - Google Patents

Image delivery method, image reception method, server, terminal apparatus, and image delivery system Download PDF

Info

Publication number
WO2015159487A1
WO2015159487A1 PCT/JP2015/001655 JP2015001655W WO2015159487A1 WO 2015159487 A1 WO2015159487 A1 WO 2015159487A1 JP 2015001655 W JP2015001655 W JP 2015001655W WO 2015159487 A1 WO2015159487 A1 WO 2015159487A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
terminal device
camera
videos
server
Prior art date
Application number
PCT/JP2015/001655
Other languages
French (fr)
Japanese (ja)
Inventor
敏康 杉尾
陽司 柴原
悠樹 丸山
徹 松延
陽一 杉野
幹博 大内
寿郎 笹井
邦昭 磯貝
竜二 牟田
貴子 堀
伊藤 智祥
Original Assignee
パナソニックIpマネジメント株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP2014082774A external-priority patent/JP2015204512A/en
Priority claimed from JP2015045352A external-priority patent/JP6607433B2/en
Application filed by パナソニックIpマネジメント株式会社 filed Critical パナソニックIpマネジメント株式会社
Priority to EP15779927.1A priority Critical patent/EP3133819A1/en
Publication of WO2015159487A1 publication Critical patent/WO2015159487A1/en
Priority to US15/285,736 priority patent/US10271082B2/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/262Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/21805Source of audio or video content, e.g. local disk arrays enabling multiple viewpoints, e.g. using a plurality of cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/23418Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/266Channel or content management, e.g. generation and management of keys and entitlement messages in a conditional access system, merging a VOD unicast channel into a multicast channel
    • H04N21/2665Gathering content from different sources, e.g. Internet and satellite
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/181Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources

Definitions

  • the present invention relates to a video distribution method for distributing video shot from a plurality of viewpoints.
  • Patent Document 1 As a video distribution method, for example, a technique described in Patent Document 1 is known.
  • a video distribution method for distributing video shot from a plurality of viewpoints is known (see, for example, Patent Document 2).
  • a user can designate and view an arbitrary video from a plurality of videos obtained by shooting a specific scene from different viewpoints.
  • an object of the present invention is to provide a video distribution method or a video reception method capable of smoothly switching video.
  • a video distribution method is a video distribution method by a server that distributes to a terminal device any one of a plurality of videos taken from different viewpoints by a plurality of users.
  • a video reception method is a video reception method by a terminal device that receives any of a plurality of videos taken from a plurality of viewpoints from a server and displays the videos. Selecting the first video from the video, a requesting step for requesting the server to transmit the first video, a first receiving step for receiving the first video from the server, and the first video A display step of displaying, and a second reception step of starting reception of a second video that is one of the plurality of videos and is likely to be selected next during reception of the first video.
  • the present invention can provide a video distribution method or video reception method capable of smoothly switching video.
  • FIG. 2 is a block diagram of a server according to Embodiment 1.
  • FIG. 2 is a block diagram of a terminal device according to Embodiment 1.
  • FIG. 4 is a diagram showing processing of the video distribution system according to Embodiment 1.
  • FIG. 6 is a diagram illustrating an example of an initial screen according to Embodiment 1.
  • FIG. 6 is a diagram showing an example of related video selection processing according to Embodiment 1.
  • FIG. 6 is a diagram showing an example of related video selection processing according to Embodiment 1.
  • FIG. 6 is a diagram showing an example of related video selection processing according to Embodiment 1.
  • FIG. 6 is a diagram showing an example of related video selection processing according to Embodiment 1.
  • FIG. 6 is a diagram showing an example of related video selection processing according to Embodiment 1.
  • FIG. 6 is a diagram showing an example of related video selection processing according to Embodiment 1.
  • FIG. 6 is a diagram showing an example of related video selection processing according to Embodiment 1.
  • FIG. 6 is a diagram showing an example of a display screen according to Embodiment 1.
  • 1 is an overall configuration diagram of a content supply system that realizes a content distribution service.
  • 1 is an overall configuration diagram of a digital broadcasting system. It is a figure which shows an example of a smart phone. It is a block diagram which shows the structural example of a smart phone.
  • Patent Document 1 discloses a method for sending a large image including the periphery of a viewing image.
  • Patent Document 2 discloses a method of distributing a viewpoint video around a display viewpoint video among a plurality of videos with different viewpoints as a group video.
  • a video distribution method is a video distribution method by a server that distributes to a terminal device one of a plurality of videos taken from different viewpoints by a plurality of users, and one of the plurality of videos.
  • a distribution step of distributing the first video requested from the terminal device to the terminal device, and a second video that is one of the plurality of videos and is likely to be requested next from the terminal device And a transmission step of starting transmission of the second video to the terminal device while delivering the first video to the terminal device.
  • the second video is sent in advance to the terminal device during the display of the first video.
  • the terminal device can smoothly switch from the first video to the second video.
  • a video having a high degree of association with the first video is selected as the second video from the plurality of videos.
  • the terminal device can smoothly switch the video.
  • the selection step it is determined that the degree of association is higher as the position of the shooting scene is closer to the position of the shooting scene of the first video.
  • the selection step it is further determined that the degree of association is higher as the width of the shooting scene is closer to the width of the shooting scene of the first video.
  • the degree of association between the subject included in the first video and the video in which the same subject is shot is set high.
  • the second video is selected based on a frame rate, resolution, or bit rate of the plurality of videos.
  • an image that is frequently selected by another user among the plurality of images is selected as the second image.
  • the second video is selected based on the user's viewing history or pre-registered preference information.
  • a video reception method is a video reception method by a terminal device that receives any of a plurality of videos taken from a plurality of viewpoints from a server and displays the videos. Selecting the first video from the video, a requesting step for requesting the server to transmit the first video, a first receiving step for receiving the first video from the server, and the first video A display step of displaying, and a second reception step of starting reception of a second video that is one of the plurality of videos and is likely to be selected next during reception of the first video.
  • the terminal device receives the second video in advance during the display of the first video. Thereby, the terminal device can smoothly switch from the first video to the second video.
  • the video receiving method further includes a step of storing the received second video, and the second video stored when the second video is selected during display of the first video. Displaying two images.
  • the video distribution method further includes a step of receiving the third video from the server when a third video different from the first video and the second video is selected during display of the first video. And displaying the stored second video until the third video is received.
  • the terminal device can display the second video during the waiting time for switching from the first video to another video.
  • an image overlooking a place where the plurality of videos are taken is displayed, and an image including a plurality of icons indicating the positions of the plurality of viewpoints is displayed.
  • an icon indicating the position of the viewpoint of the second video among the plurality of icons is highlighted.
  • the server is a server that distributes to a terminal device any one of a plurality of videos shot from different viewpoints by a plurality of users, and is one of the plurality of videos, A distribution unit that distributes the first video designated by the terminal device to the terminal device, and a selection for selecting a second video that is one of the plurality of videos and is likely to be requested next from the terminal device And a transmission unit that starts transmission of the second video to the terminal device while delivering the first video to the terminal device.
  • the second video is sent in advance to the terminal device during the display of the first video.
  • the terminal device can smoothly switch from the first video to the second video.
  • a terminal device is a terminal device that receives any of a plurality of videos taken from a plurality of viewpoints from a server and displays the videos.
  • a selection unit that selects a video, a request unit that requests the server to transmit the first video, a first reception unit that receives the first video from the server, and a display unit that displays the first video
  • a second receiving unit that starts receiving a second video that is one of the plurality of videos and is likely to be selected next during reception of the first video.
  • the terminal device receives the second video in advance during the display of the first video. Thereby, the terminal device can smoothly switch from the first video to the second video.
  • the video distribution system includes a server and a terminal device.
  • the second video is sent in advance to the terminal device during the display of the first video.
  • the terminal device can smoothly switch from the first video to the second video.
  • FIG. 1 is a block diagram showing a configuration of a video distribution system 100 according to the present embodiment.
  • the video distribution system 100 includes a plurality of cameras 101, a terminal device 102, and a server 103 that can communicate with each other via a network 104.
  • the plurality of cameras 101 generate a plurality of video signals by photographing the same scene from different viewpoints in the same time zone.
  • Each camera 101 is carried by each of a plurality of users.
  • the plurality of cameras 101 are owned by a plurality of spectators in a place such as a sports stadium.
  • a plurality of video signals photographed by the plurality of cameras 101 are transmitted to the server 103 via the network 104.
  • the video signal includes information indicating a photographing viewpoint (camera position), a camera direction, a magnification, and the like.
  • the camera 101 only needs to be a device having at least a photographing function, such as a digital still camera, a digital video camera, a smartphone, or a mobile terminal.
  • the terminal device 102 is a terminal used by the user, and has at least a function of displaying an image.
  • the terminal device 102 is a smartphone, a portable terminal, a personal computer, or the like.
  • the terminal device 102 has the same function as the camera 101, and the user may be included in the audience, or the user may view a video from a place other than the stadium.
  • the server 103 holds a plurality of video signals transmitted from the plurality of cameras 101. Further, the server 103 transmits a part of the plurality of video signals to be held to the terminal device 102 in accordance with a request from the terminal device 102. In addition, the server 103 analyzes the contents of the plurality of video signals and calculates the relevance of the plurality of video signals based on the obtained video characteristics. Further, the server 103 transmits a related video signal having a high degree of relevance to the selected video signal to the terminal device 102 in addition to the selected video signal designated by the terminal device 102.
  • FIG. 2 is a block diagram showing the configuration of the server 103.
  • the server 103 includes a reception unit 111, a video storage unit 112, a control unit 113, and a transmission unit 114.
  • the receiving unit 111 receives a plurality of video signals 151 in which the same scene is captured from different viewpoints by the plurality of cameras 101.
  • the reception unit 111 receives the viewpoint designation signal 152 transmitted from the terminal device 102.
  • This viewpoint designation signal 152 designates one of the plurality of video signals 151.
  • the video storage unit 112 stores a plurality of video signals 151 received by the reception unit 111.
  • the control unit 113 selects the video signal 151 designated by the viewpoint designation signal 152 from the plurality of video signals 151 stored in the video storage unit 112 as the selected video signal 153, and selects the selected video via the transmission unit 114.
  • the signal 153 is transmitted to the terminal device 102.
  • the control unit 113 selects a related video signal 154 related to the selected video signal 153 from the plurality of video signals 151 stored in the video storage unit 112, and the related video signal is transmitted via the transmission unit 114.
  • 154 is transmitted to the terminal device 102.
  • FIG. 3 is a block diagram of the terminal device 102.
  • the terminal device 102 includes a receiving unit 121, a storage unit 122, a decoding unit 123, an output unit 124, a transmission unit 125, a control unit 126, and an input unit 127.
  • the receiving unit 121 receives the selected video signal 153 and the related video signal 154 transmitted from the server 103.
  • the storage unit 122 temporarily holds the selected video signal 153 and the related video signal 154 received by the receiving unit 121.
  • the decoding unit 123 generates a decoded video by decoding the selected video signal 153.
  • the output unit 124 generates an output video 155 including the decoded video, and displays the output video 155 on a display device such as a display provided in the terminal device 102, for example.
  • the input unit 127 receives a user operation. For example, the input unit 127 receives a user operation on the touch panel provided in the terminal device 102. When the input unit 127 receives an operation of changing the viewpoint by the user, the control unit 126 transmits a viewpoint designation signal 152 indicating the viewpoint to be changed to the server 103 via the transmission unit 125.
  • FIG. 4 is a sequence diagram of video distribution processing in the video distribution system 100.
  • a plurality of video signals 151 are already held in the server 103.
  • the plurality of video signals 151 may be videos updated in real time from the plurality of cameras 101 as in the case of the stadium, in which the user is a stadium spectator. It may be a past video.
  • the terminal device 102 starts an application program (application) in accordance with a user operation (S101).
  • the terminal device 102 displays an initial screen (S102).
  • the terminal apparatus 102 receives information indicating the positions (viewpoint positions) of the plurality of cameras 101 when the plurality of video signals 151 are captured as initial information from the server 103. Information indicating the position is displayed as an initial screen.
  • FIG. 5 is a diagram showing an example of this initial screen.
  • the background image 201 an image overlooking a place where a plurality of videos are taken is used.
  • a camera icon 202 indicating the position of the viewable video and the position of the camera 101 that shot the video is displayed on the background image 201.
  • thumbnails may be displayed instead of the camera icon 202 or in addition to the camera icon 202. Furthermore, a thumbnail may be displayed instead of the camera icon 202 when the initial screen is enlarged.
  • thumbnails when the number of videos is large, only the camera icon 202 or the thumbnail of the video having a high recommendation level for the user may be displayed based on the degree of relevance described later. When thumbnails are displayed, the thumbnails may be displayed larger than the camera icon 202.
  • the camera icon 202 may be displayed for each group or a representative video of each group.
  • the representative video is determined based on, for example, video characteristics (resolution, frame rate, bit rate, or the like). For example, a video having the highest resolution, a video having the highest frame rate, or a video having the highest bit rate is determined as the representative video.
  • information indicating the related contents of each group may be displayed together with the camera icon 202. Further, instead of the camera icon 202, thumbnails of representative videos of each group or reduced videos may be displayed.
  • the terminal apparatus 102 may receive the representative video from the server 103 in advance. That is, the terminal device 102 may receive all the representative videos of each group when the initial screen is displayed. Alternatively, when a certain group or representative video is selected, the terminal device 102 may receive a part or all of the video included in the group from the server 103.
  • the terminal device 102 may set only the camera icon 202 corresponding to the video that has been stored for a while in the storage unit 122 after a while from the start of reception.
  • the terminal device 102 may select the viewpoint to be displayed so that the number of camera icons 202 to be displayed is constant even when the screen is enlarged or reduced according to a user operation.
  • the background image 201 of the initial screen may be switched depending on the position where the user is currently present. For example, when the user is on the infield side stand of the stadium, the landscape image seen from the infield side stand is set as the background image 201, and when the user is on the outfield side stand, the landscape image seen from the outfield stand is the background image 201. May be set.
  • the camera icon 202 displayed on the initial screen may be switched according to the background image 201.
  • the camera icon 202 may be switched according to the position of the user. For example, when the user is on the infield side stand, an image of a landscape seen from the infield side stand may be set as the background image 201, and a camera icon 202 indicating a shooting viewpoint existing in the landscape may be displayed.
  • the video to be received in advance may be switched according to the position of the user.
  • the terminal device 102 may receive in advance an image shot from the outfield side stand.
  • the initial screen or the video received in advance may be switched depending on the viewing status of all users or some users. For example, a video with a large number of viewing users or a large number of users viewed in the past may be preferentially received.
  • the terminal apparatus 102 transmits a viewpoint designation signal 152 indicating the selected viewpoint to the server 103 (S104).
  • the server 103 that has received the viewpoint designation signal 152 starts transmission of the selected video signal 153 designated by the viewpoint designation signal 152 to the terminal device 102 (S105).
  • the terminal device 102 that has received the selected video signal 153 decodes the selected video signal 153 and starts displaying the obtained video (S106).
  • the server 103 that has received the viewpoint designation signal 152 selects the related video signal 154 related to the selected video signal 153 (S107), and starts transmitting the related video signal 154 to the terminal device 102 (S108).
  • the selection of the related video signal 154 (S107) is performed after the transmission of the selected video signal 153 is started (S105).
  • these processing orders may be arbitrary, and some of them are performed in parallel. Also good.
  • the related video selection process (S108) will be described.
  • the server 103 uses at least one of the following methods as related video selection processing. In each of the following methods, the relevance level of each video is set, and the video with the highest final relevance level is selected as the related video. A plurality of videos may be selected as related videos in order from the higher priority side.
  • the server 103 calculates the position of the shooting scene of the selected video (area shown in the video) (S151), and determines the relevance of the video whose shooting scene is close to the position of the shooting scene of the selected video. Increase (S152). Specifically, the server 103 calculates the position of the shooting scene of each video using information included in the video signal 151 transmitted to the camera 101. More specifically, the video signal 151 includes information such as the viewpoint position where the video is shot, the direction of the camera 101, and the zoom magnification. The server 103 uses these pieces of information to calculate the position of the shooting scene that the camera 101 is shooting.
  • the server 103 may calculate the position of the shooting scene of each video in advance, or may be performed at an arbitrary timing as long as the video signal 151 is received.
  • the server 103 may increase the relevance of the video whose shooting scene is close to the shooting scene of the selected video.
  • the server 103 does not need to increase the degree of relevance for a video whose position of the imaging scene is very close (almost the same) to the position of the shooting scene of the selected video.
  • the server 103 identifies a subject (for example, a player) in the selected video (S161), and increases the relevance of a video in which the same subject as the subject in the selected video is captured (S162).
  • the camera 101 identifies a subject in the video by image analysis (face authentication or the like), and transmits a video signal 151 including information indicating the subject to the server 103.
  • the server 103 determines a subject in each video using the information.
  • the image analysis may be performed by the server 103.
  • the subject is not limited to a specific person, but may be a specific team.
  • the server 103 calculates the degree of association using the information generated by the camera 101 or the server 103 using at least one of the video captured by the camera 101 and the information acquired by the sensor attached to the camera. To do.
  • the server 103 acquires the popularity of a plurality of videos (S171), and increases the relevance of videos with high popularity (S172).
  • the degree of popularity indicates, for example, the number of times that a video has been viewed within a certain time in the present or the past, or the number of users who have viewed the video. Note that the degree of popularity is sequentially calculated in the server 103 based on the viewing status of a plurality of users, for example.
  • the server 103 acquires user preference information (S181), and increases the degree of relevance of the video that matches the user preference (S182).
  • the preference information is a user's viewing history or registered information indicating a user's preference or hobby registered in advance. For example, when the user has watched many videos of a specific player or team in the past, the server 103 increases the degree of relevance of the videos of the player or team. Moreover, when the player or team which a user supports is shown by registration information, the server 103 increases the relevance degree of the image
  • the server 103 acquires communication band information indicating a communication band that can be used by the terminal device 102 (S191), and changes the degree of association according to the communication band (S192). Specifically, the server 103 increases the degree of relevance of the video having the bit rate, the frame rate, or the resolution suitable for the communication band that can be used by the terminal device 102. For example, when the communication band that can be used by the terminal device 102 is sufficiently wide, the server 103 increases the relevance of the video having a high bit rate, frame rate, or resolution.
  • the server 103 may generate a plurality of bit rate video signals by converting the resolution or the frame rate of the video signal 151 transmitted from the camera 101, and store the plurality of video signals.
  • the selected video or the related video may be switched according to the available bandwidth.
  • the server 103 acquires communication band information indicating a communication band that can be used by the terminal apparatus 102 (S191), and determines the number of related videos according to the communication band (S193). Specifically, the server 103 increases the number of related videos as the communication band is wider.
  • the server 103 selects, as a related video (second video), a video having a high degree of association with the selected video (first video) from among a plurality of videos. Specifically, the server 103 determines that the degree of association is higher as the position of the shooting scene is closer to the position of the shooting scene of the selected video. Further, the server 103 determines that the degree of association is higher as the width of the shooting scene is closer to the width of the shooting scene of the selected video. Further, the server 103 sets a high degree of association between the subject included in the selected video and the video in which the same subject is captured.
  • the server 103 selects related videos based on the frame rate, resolution, or bit rate of a plurality of videos. In addition, the server 103 selects, as a related video, a video that has been frequently selected by another user from a plurality of videos. Further, the server 103 selects a related video based on the viewing history of the user or the preference information registered in advance.
  • FIG. 12 is a diagram showing an example of the display screen after the video is selected. As shown in FIG. 12, a selection video 211 that is a selected video, an overhead image 212, a top image 213, and operation buttons 214 to 216 are displayed on the display screen.
  • the bird's-eye view image 212 is an image for bird's-eye view of the shooting scene, and includes a camera icon 202.
  • This overhead image 212 is the same as the image displayed on the initial screen.
  • the top image 213 is a view of the entire shooting scene viewed from above, and includes a camera icon 202.
  • Operation buttons 214 to 216 are buttons for the user to operate.
  • the operation button 214 is selected, the display returns to the initial screen.
  • the operation button 215 or 216 is operated, the display video is switched to another viewpoint video. At this time, a video having a high degree of association with the selected video is preferentially selected.
  • the display image is switched to the image where the position of the shooting scene is closest to the position of the shooting scene of the selected image.
  • the display image is switched to the image with the highest recommendation.
  • the user can easily switch the display video to the video that can enjoy the game at that time and view the displayed video. it can.
  • the display image is switched to the image corresponding to the selected camera icon 202.
  • positioning of each image and operation button shown in FIG. 12 is an example, and is not limited to this example. Further, it is not necessary to display all of the plurality of images and the plurality of operation buttons, and only a part of them may be displayed.
  • the display of the camera icon 202 is changed according to the degree of association with the selected video. For example, a camera icon 202 corresponding to a video having a high degree of association with the selected video is highlighted. Note that only the camera icon 202 corresponding to a video having a high degree of association with the selected video among the plurality of videos may be displayed. Further, the display method of the camera icon 202 may be changed continuously or stepwise according to the degree of association. Information indicating the degree of association may be displayed near the camera icon 202.
  • a sensor may be incorporated in the ball, and it may be determined how the ball flew based on information detected by the sensor. Then, the trajectory of the ball may be superimposed on the overhead image 212 or the top image 213.
  • the terminal device 102 may receive the video signal of the viewpoint position close to the position of the ball in advance by the server 103.
  • the system obtains the flow of the game from some means (such as a ball sensor) and preliminarily estimates the camera icon 202 that the user wants to see based on the information, and the terminal device 102 is estimated.
  • the video may be received in advance.
  • the server 103 may set priorities for a plurality of videos based on the current situation such as the flow of the game or the position of the user.
  • a viewpoint switching operation is performed (S109).
  • the terminal device 102 since the terminal device 102 has received the related video signal 154 in advance, the terminal device 102 decodes the related video signal 154 and displays the related video (S110). In this way, the terminal device 102 can seamlessly switch the video by receiving in advance the related video that is likely to be selected next.
  • the terminal device 102 transmits a viewpoint designation signal 152 indicating the selected viewpoint to the server 103 (S111).
  • the server 103 that has received the viewpoint designation signal 152 transmits a selection video signal 153 designated by the viewpoint designation signal 152 to the terminal device 102. That is, the server 103 continues transmission of the related video signal 154 as transmission of the selected video signal 153 (S112). Further, the server 103 selects the related video signal 154 related to the new selected video signal 153 (S113), and starts transmitting the related video signal 154 to the terminal device 102 (S114).
  • the order of the video display (S110) and the transmission of the viewpoint designation signal 152 (S111) may be arbitrary, and some of them may be performed in parallel.
  • FIG. 13 is a flowchart illustrating the operation flow of the terminal apparatus 102.
  • FIG. 13 shows processing of the terminal device 102 in a state where a certain viewpoint video is displayed.
  • the terminal apparatus 102 determines whether or not the viewpoint switching is instructed by the user's operation (S121). When the viewpoint switching is instructed (Yes in S121), the terminal device 102 transmits the viewpoint designation signal 152 to the server 103 (S122).
  • the terminal device 102 determines whether the viewpoint switching destination selected video is a related video (S123). If the selected video is not related video (No in S123), the terminal apparatus 102 waits until receiving the selected video transmitted by the server 103 in response to the viewpoint designation signal 152 (S124), and if the selected video is received (S124). Yes), the selected video is displayed (S125).
  • the terminal device 102 displays the already stored related video as the selected video (S125).
  • switching of decoded video may be performed at the time when decoding of the random access frame is completed.
  • a waiting time is generated from the time when the user's viewpoint switching instruction is generated until the switching time, and the terminal device 102 may continue to play the video before switching during this waiting time, May be displayed.
  • the terminal device 102 searches for a random access point at a time closest to the playback time of the video before switching, and decodes and displays the video from there. Also good.
  • the terminal device 102 when receiving the related video related to the new selected video (Yes in S126), the terminal device 102 sequentially stores the received related video in the storage unit 122 (S127). Note that the data of the selected video after being displayed and the data of the related video that has not been used for a certain period after reception are sequentially deleted from the storage unit 122.
  • the terminal apparatus 102 displays the newly received related video information (S128). Specifically, the terminal device 102 highlights the camera icon 202 of the related video. For example, the related video camera icon 202 is displayed larger than the other camera icons 202. In addition, the contour line of the camera icon 202 of the related video is displayed thicker than the contour lines of the other camera icons 202. Alternatively, the color of the camera icon 202 of the related video is changed to a conspicuous color such as red. The highlighting method is not limited to this.
  • terminal apparatus 102 may perform the processing shown in FIG. 14 and 15 are flowcharts illustrating a flow of a modification example of the operation of the terminal device 102.
  • step S129 is added to the process shown in FIG. That is, when the selected video is not related video (No in S123), the terminal apparatus 102 displays the related video in a period until the selected video is received (S129). In addition, when the terminal device 102 stores a plurality of related videos, the terminal device 102 displays a related video having the highest degree of association with the newly selected video among the plurality of stored related videos. May be.
  • step S130 is added to the process shown in FIG. That is, when the selected video is not a related video (No in S123), the terminal device 102 displays the three-dimensional configuration data in a period until the selected video is received (S130).
  • the three-dimensional configuration data is the three-dimensional configuration data of a place where a plurality of videos are taken, and in the example shown in FIG. 5, is the three-dimensional configuration data of a baseball field.
  • the three-dimensional configuration data is generated by the server 103 using a plurality of video signals 151 and transmitted to the terminal device 102 in advance.
  • the terminal device 102 may generate an image to be displayed during this period using the three-dimensional configuration data.
  • the terminal apparatus 102 generates a video in which the viewpoint position is continuously changed from the viewpoint position of the immediately preceding display video in the three-dimensional configuration data to the viewpoint position of the selected video, and displays the video during the above period.
  • Such visual effects may also be used when video data is stored in the storage unit 122.
  • whether or not this visual effect is used may be switched according to the distance between the viewpoint position of the immediately preceding display image and the viewpoint position of the selected image. For example, when the distance is short, the visual effect is not used, and when the distance is long, the visual effect is used.
  • the terminal device 102 uses another communication method such as near field communication.
  • the video signal may be received directly from the 101.
  • the terminal apparatus 102 receives any of a plurality of videos taken from a plurality of viewpoints from the server 103 and displays the videos.
  • the terminal device 102 selects a selection video (first video) from a plurality of videos (S121).
  • the terminal device 102 requests the server 103 to transmit the selected video (S122).
  • the terminal device 102 receives the selected video from the server 103 (S124), and displays the selected video (S125).
  • the terminal device 102 starts receiving a related video that is one of a plurality of videos and is different from the selected video and is likely to be selected next. (S126).
  • the terminal device 102 accumulates the received related video (S127).
  • the terminal device 102 displays the stored related video (S125).
  • the terminal device 102 receives the third video from the server 103 (S124).
  • the terminal apparatus 102 displays the related video that has been accumulated until the third video is received (S129).
  • FIG. 16 is a flowchart showing the operation flow of the server 103.
  • the server 103 determines whether the viewpoint designation signal 152 is received from the terminal device 102 (S141). When the server 103 receives the viewpoint designation signal 152 (Yes in S141), the server 103 selects the video signal indicated by the viewpoint designation signal 152 as the selected video signal 153 from the accumulated video signals, and the selected video signal 153 is transmitted to the terminal device 102 (S142).
  • the server 103 selects the related video signal 154 having a high degree of relevance to the selected video based on the priority from the plurality of stored video signals 151 (S143), and uses the related video signal 154 as a terminal.
  • the data is transmitted to the device 102 (S144).
  • the server 103 delivers any of a plurality of videos taken from different viewpoints by a plurality of users to the terminal device 102.
  • the server 103 distributes the selected video (first video) requested by the terminal device 102 to the terminal device 102, which is one of a plurality of videos (S142).
  • the server 103 selects, from a plurality of videos, a related video (second video) that is different from the selected video and is highly likely to be requested next from the terminal device 102 (S143).
  • the related video is a video that is not requested from the terminal device 102.
  • the server 103 starts transmitting the related video to the terminal device 102 while delivering the selected video to the terminal device 102 (S144).
  • each processing unit included in each device included in the video distribution system according to the above embodiment is typically realized as an LSI which is an integrated circuit. These may be individually made into one chip, or may be made into one chip so as to include a part or all of them.
  • circuits are not limited to LSI, and may be realized by a dedicated circuit or a general-purpose processor.
  • An FPGA Field Programmable Gate Array
  • reconfigurable processor that can reconfigure the connection and setting of circuit cells inside the LSI may be used.
  • each component may be configured by dedicated hardware or may be realized by executing a software program suitable for each component.
  • Each component may be realized by a program execution unit such as a CPU or a processor reading and executing a software program recorded on a recording medium such as a hard disk or a semiconductor memory.
  • each device included in the video distribution system includes a processing circuit and a storage device (storage) electrically connected to the processing circuit (accessible from the processing circuit).
  • the processing circuit includes at least one of dedicated hardware and a program execution unit.
  • the storage device stores a software program executed by the program execution unit. The processing circuit uses the storage device to execute the video distribution method or the wave video reception method according to the above embodiment.
  • the present invention may be the software program or a non-transitory computer-readable recording medium on which the program is recorded.
  • the program can be distributed via a transmission medium such as the Internet.
  • the order in which the steps included in the above video distribution method or video reception method are executed is for illustrating the present invention specifically, and may be in an order other than the above. Also, some of the above steps may be executed simultaneously (in parallel) with other steps.
  • the video distribution method, the video reception method, the video distribution system, the server, and the terminal device have been described based on the embodiment. It is not limited. Unless it deviates from the gist of the present invention, the embodiment in which various modifications conceived by those skilled in the art have been made in the present embodiment, and forms constructed by combining components in different embodiments are also applicable to one or more of the present invention. It may be included within the scope of the embodiments.
  • the system can be applied to a video system in which intelligentization and widening of the target space are progressing.
  • a surveillance system implemented in a security camera in a store or a factory, an in-vehicle camera in a police, or Traffic information system using own camera or each on-vehicle camera or camera provided on road, (3) Environmental survey or delivery system using remote control or automatic control device such as drone, and (4) Entertainment
  • the present invention can be applied to a content transmission / reception system such as a video using an installation camera in a facility or a stadium, a mobile camera such as a drone, or a personally owned camera.
  • FIG. 17 is a diagram showing a configuration of the video information processing system ex100 in the present embodiment. In this embodiment, an example of preventing the generation of blind spots and an example of prohibiting photographing in a specific area will be described.
  • 17 includes a video information processing apparatus ex101, a plurality of cameras ex102, and a video reception apparatus ex103. Note that the video receiving device ex103 is not necessarily included in the video information processing system ex100.
  • the video information processing apparatus ex101 includes a storage unit ex111 and an analysis unit ex112.
  • Each of the N cameras ex102 has a function of capturing video and a function of transmitting captured video data to the video information processing apparatus ex101.
  • the camera ex102 may have a function of displaying an image being shot.
  • the camera ex102 converts the captured video signal into HEVC or H.264.
  • the encoded information may be encoded using an encoding method such as H.264 and transmitted to the video information processing apparatus ex101, or unencoded video data may be transmitted to the video information processing apparatus ex101.
  • each camera ex102 is a fixed camera such as a surveillance camera, a moving camera mounted on an unmanned flight type radio control or a car, or a user camera possessed by the user.
  • the moving camera receives the instruction signal transmitted from the video information processing apparatus ex101, and changes the position or shooting direction of the moving camera itself according to the received instruction signal.
  • the time of the plurality of cameras ex102 is calibrated using the time information of the server or the reference camera, etc. before the disclosure of photographing. Further, the spatial positions of the plurality of cameras ex102 are calibrated based on how the objects in the space to be imaged are captured or relative positions from the reference camera.
  • the storage unit ex111 included in the information processing apparatus ex101 stores video data transmitted from the N cameras ex102.
  • the analysis unit ex112 detects a blind spot from the video data stored in the storage unit ex111, and transmits an instruction signal indicating an instruction to the mobile camera for preventing the generation of the blind spot to the mobile camera.
  • the moving camera moves in accordance with the instruction signal and continues shooting.
  • the analysis unit ex112 performs blind spot detection using, for example, SfM (Structure from Motion).
  • SfM is a technique for restoring the three-dimensional shape of a subject from a plurality of videos taken from different positions, and is widely known as a shape restoration technique for simultaneously estimating the subject shape and the camera position.
  • the analysis unit ex112 restores the three-dimensional shape in the facility or the stadium from the video data saved in the saving unit ex111 using SfM, and detects an area that cannot be restored as a blind spot.
  • the analysis unit ex112 may perform SfM using these known information. Further, when the position and shooting direction of the moving camera can be acquired by a GPS and an angle sensor provided in the moving camera, the moving camera transmits information on the position and shooting direction of the moving camera to the analysis unit ex112, and the analysis unit The ex 112 may perform SfM using the transmitted position and shooting direction information.
  • the method of detecting the blind spot is not limited to the method using SfM described above.
  • the analysis unit ex112 may grasp the spatial distance of the object to be imaged by using information of a depth sensor such as a laser range finder.
  • the analysis unit ex112 detects information such as a camera position, a shooting direction, and a zoom magnification from an image that includes a preset marker or a specific object in the space, or the size of the marker or the like. Also good.
  • the analysis unit ex112 performs blind spot detection using an arbitrary method capable of detecting the imaging region of each camera.
  • the analysis unit ex112 acquires information such as a mutual positional relationship for a plurality of imaging targets from video data or a proximity distance sensor, and identifies an area where a blind spot is likely to occur based on the acquired positional relationship. May be.
  • the blind spot includes not only a portion where an image does not exist in a region to be photographed, but also a portion having a poor image quality compared to other portions and a portion where a predetermined image quality is not obtained.
  • This detection target portion may be set as appropriate according to the configuration or purpose of the system. For example, the required image quality may be set high for a specific subject in the space where the image is taken. Conversely, for a specific area in the shooting space, the required image quality may be set low, or it may be set not to be determined as a blind spot even if no video is shot.
  • the above-mentioned image quality includes various information related to the video such as the area occupied by the subject to be photographed in the video (for example, the number of pixels) or whether the subject to be photographed is in focus. Whether or not it is a blind spot may be determined based on the information or the combination thereof.
  • a region that needs to be detected in order to prevent the generation of a blind spot is not limited to a region that is actually a blind spot.
  • the analysis unit ex112 detects movements of a plurality of shooting targets from, for example, shot video data and the like, and based on the detected movements of the plurality of shooting targets and position information of the camera ex102, a new blind spot and It is also possible to estimate a possible region.
  • the video information processing apparatus ex101 may transmit an instruction signal to the moving camera so as to capture an area that may become a blind spot, and prevent the generation of a blind spot.
  • the video information processing apparatus ex101 needs to select a moving camera that transmits an instruction signal in order to capture a blind spot or an area that may become a blind spot.
  • the video information processing apparatus ex101 determines which dead spots or areas that may become blind spots for each of the plurality of moving cameras. It is necessary to decide whether to shoot. For example, the video information processing apparatus ex101 selects a moving camera that is closest to the blind spot or the area that is the blind spot based on the blind spot or the area that may be the blind spot and the position of the area that each moving camera is capturing. To do. Further, the video information processing apparatus ex101 determines, for each moving camera, whether or not a blind spot is newly generated when the moving camera cannot obtain the video data currently being shot. If it is not obtained, a moving camera determined not to generate a blind spot may be selected.
  • the video information processing apparatus ex101 can prevent the generation of a blind spot by detecting a blind spot and transmitting an instruction signal to the moving camera so as to prevent the blind spot.
  • the instruction signal may be a signal for instructing the user of the user camera to move.
  • the user camera displays an instruction image that instructs the user to change the direction of the camera based on the instruction signal.
  • the user camera may display an instruction image indicating a movement route on a map as an instruction to move the user.
  • the user camera may display detailed shooting instructions such as shooting direction, angle, angle of view, image quality, and movement of the shooting area in order to improve the quality of the acquired image. If control is possible on the ex101 side, the video information processing apparatus ex101 may automatically control the feature amount of the camera ex102 regarding such shooting.
  • the user camera is, for example, a smartphone, a tablet terminal, a wearable terminal, or an HMD (Head Mounted Display) held by a spectator in the stadium or a guard in the facility.
  • HMD Head Mounted Display
  • the display terminal that displays the instruction image need not be the same as the user camera that captures the video data.
  • the user camera may transmit an instruction signal or an instruction image to a display terminal associated with the user camera in advance, and the display terminal may display the instruction image.
  • information on the display terminal corresponding to the user camera may be registered in advance in the video information processing apparatus ex101.
  • the video information processing apparatus ex101 may display the instruction image on the display terminal by directly transmitting the instruction signal to the display terminal corresponding to the user camera.
  • the analysis unit ex112 may generate a free viewpoint video (three-dimensional reconstruction data) by restoring the three-dimensional shape in the facility or the stadium from the video data stored in the storage unit ex111 using, for example, SfM. Good.
  • This free viewpoint video is stored in the storage unit ex111.
  • the video information processing apparatus ex101 reads video data corresponding to the visual field information (and / or viewpoint information) transmitted from the video reception apparatus ex103 from the storage unit ex111 and transmits the video data to the video reception apparatus ex103.
  • the video reception device ex103 may be one of the plurality of cameras 111.
  • the video information processing apparatus ex101 may detect a shooting prohibited area.
  • the analysis unit ex112 analyzes the photographed image, and transmits a photographing prohibition signal to the moving camera when the mobile camera is photographing the photographing prohibition region.
  • the mobile camera stops shooting while receiving the shooting prohibition signal.
  • the analysis unit ex112 matches the three-dimensional virtual space restored using SfM with the captured image, thereby determining whether the mobile camera set in advance in the space is capturing the prohibited image area. judge.
  • the analysis unit ex112 determines whether the moving camera is shooting the shooting prohibited area using a marker or a characteristic object arranged in the space as a trigger.
  • the photographing prohibited area is, for example, a toilet in a facility or a stadium.
  • the user camera when the user camera is shooting a shooting prohibited area, the user camera displays a message on a display or the like connected wirelessly or by wire, or outputs a sound or sound from a speaker or an earphone.
  • the user may be informed that the current location is a shooting prohibited location.
  • the shooting prohibited area and the current shooting area are shown on the displayed map.
  • the resumption of photographing is automatically performed when, for example, the photographing prohibition signal is not output.
  • photographing may be resumed when the photographing prohibition signal is not output and the user performs an operation to resume photographing.
  • calibration may be performed again.
  • notification for confirming the current position or prompting the user to move may be performed.
  • a passcode or fingerprint authentication that turns off such a function for recording may be used.
  • image processing such as mosaicing may be automatically performed when a video in the photographing prohibited area is displayed or stored outside.
  • the video information processing apparatus ex101 can determine that shooting is prohibited and notify the user to stop shooting, thereby setting a certain region to shooting prohibited.
  • the video information processing system ex100 sets an incentive for the user who transferred the shot video.
  • the video information processing apparatus ex101 delivers a video value to a user who has transferred video at a free or discounted rate, a monetary value that can be used in an online or offline store or game, a game, etc. Points that have non-monetary value such as social status in virtual space.
  • the video information processing apparatus ex101 gives a particularly high point to a user who has transferred a captured video of a valuable field of view (and / or viewpoint) such as many requests.
  • the video information processing apparatus ex101 may transmit additional information to the user camera based on the analysis result of the analysis unit ex112. In this case, the user camera superimposes additional information on the captured video and displays it on the screen.
  • the additional information is, for example, information on players such as a player name or height when a game in a stadium is being shot, and the name or face photo of the player is associated with each player in the video. Is displayed.
  • the video information processing apparatus ex101 may extract additional information by searching via the Internet based on part or all of the video data area.
  • the camera ex102 receives such additional information by short-range wireless communication including Bluetooth (registered trademark) or visible light communication from lighting such as a stadium, and maps the received additional information to video data. Also good.
  • the camera ex102 is a table in which this mapping is stored in a storage unit connected to the camera ex102 by wire or wirelessly, and shows a correspondence relationship between information obtained by visible light communication technology and additional information, etc. It may be performed based on a certain rule of the above, or may be performed using the most probable combination result by Internet search.
  • the monitoring system for example, information of a caution person is superimposed on a user camera held by a guard in the facility, so that the monitoring system can be highly accurate.
  • the analysis unit ex112 may determine which area in the facility or stadium the user camera is capturing by matching the free viewpoint image and the captured image of the user camera. Note that the imaging region determination method is not limited to this, and various imaging region determination methods or other imaging region determination methods described in the above-described embodiments may be used.
  • the video information processing apparatus ex101 transmits the past video to the user camera based on the analysis result of the analysis unit ex112.
  • the user camera displays the past video on the screen by superimposing the past video on the shot video or replacing the shot video with the past video.
  • the highlight scene of the first half is displayed as a past video. Accordingly, the user can enjoy the highlight scene of the first half as a video in the direction in which he / she is viewing during the halftime.
  • the past video is not limited to the highlight scene in the first half, but may be a highlight scene of a past game held at the stadium.
  • the timing at which the video information processing apparatus ex101 delivers the past video is not limited to half time, and may be, for example, after the match or during the match. Particularly during a game, based on the analysis result of the analysis unit ex112, the video information processing apparatus ex101 may deliver a scene that is considered important and missed by the user.
  • the video information processing apparatus ex101 may distribute the past video only when requested by the user, or may distribute a distribution permission message before the past video is distributed.
  • the video information processing apparatus ex101 may transmit advertisement information to the user camera based on the analysis result of the analysis unit ex112.
  • the user camera superimposes advertisement information on the captured video and displays it on the screen.
  • the advertisement information may be distributed immediately before the past video distribution during the half time or after the match, as shown in, for example, Modification 5. Accordingly, the distributor can obtain an advertisement fee from the advertiser, and can provide a video distribution service to the user at a low cost or free of charge.
  • the video information processing apparatus ex101 may distribute an advertisement distribution permission message immediately before distribution of the advertisement information, may provide a service for free only when the user views the advertisement, or views the advertisement. Service may be provided at a lower cost than when not.
  • the system or the staff who knows the location of the user based on some location information or the automatic delivery system of the venue will bring the ordered drink to the seat Will deliver.
  • the decision may be handed to the staff or may be made based on credit card information set in advance in the mobile terminal application or the like.
  • the advertisement may include a link to an e-commerce site, and online shopping such as normal home delivery may be possible.
  • the video receiving device ex103 may be one of the cameras ex102 (user camera).
  • the analysis unit ex112 determines which area in the facility or stadium the user camera is shooting by matching the free viewpoint video and the video shot by the user camera. Note that the method for determining the imaging region is not limited to this.
  • the user camera when the user performs a swipe operation in the direction of the arrow displayed on the screen, the user camera generates viewpoint information indicating that the viewpoint is moved in that direction.
  • the video information processing apparatus ex101 reads the video data obtained by shooting the area moved by the viewpoint information from the shooting area of the user camera determined by the analysis unit ex112 from the storage unit ex111, and transmits the video data to the user camera. Start.
  • the user camera displays the video distributed from the video information processing apparatus ex101 instead of the captured video.
  • the users in the facility or the stadium can view the video from a favorite viewpoint with a simple operation like a screen swipe.
  • a spectator watching on the third base side of a baseball field can view a video from the first base side viewpoint.
  • the security guards in the facility can watch the video that should be watched as an interrupt from the viewpoint or the center that they want to confirm by a simple operation like a screen swipe while changing the viewpoint appropriately. Therefore, it is possible to increase the accuracy of the monitoring system.
  • the user camera may switch and display the video of a part of the shooting area of the user camera including the obstacle from the shot video to the distribution video from the video information processing apparatus ex101.
  • the entire screen may be switched from the captured video to the distributed video and displayed.
  • the user camera may display an image in which the object to be viewed is seen through the obstacle by combining the captured image and the distribution image. According to this configuration, it is possible to view the video distributed from the video information processing apparatus ex101 even when the shooting target cannot be seen from the position of the user due to the influence of the obstacle, so that the influence of the obstacle can be reduced. it can.
  • the distribution video is displayed as a video of an area that cannot be seen due to an obstacle
  • display switching control different from the display switching control according to the input process by the user such as the screen swipe described above, may be performed.
  • the display from the shot video to the distribution video is performed. Switching may be performed automatically.
  • display switching from the shot video to the distribution video may be automatically performed.
  • the display switching to the distribution video may be automatically performed.
  • the display switching from the captured video to the distribution video and the display switching from the distribution video to the captured video may be performed in accordance with the user input processing.
  • Modification 8 The speed at which the video data is transferred to the video information processing apparatus ex101 may be instructed based on the importance of the video data captured by each camera ex102.
  • the analysis unit ex112 determines the importance of the video data stored in the storage unit ex111 or the camera ex102 that captured the video data.
  • the determination of the importance is performed based on, for example, information such as the number of people or moving objects included in the video, the image quality of the video data, or a combination thereof.
  • the determination of the importance of the video data may be based on the position of the camera ex102 where the video data is shot or the area where the video data is shot. For example, when there are a plurality of other cameras ex102 being shot near the target camera ex102, the importance of the video data shot by the target camera ex102 is reduced. In addition, even when the position of the target camera ex102 is far from the other camera ex102, when there are a plurality of other cameras ex102 shooting the same area, the importance of the video data shot by the target camera ex102 is set. make low.
  • the determination of the importance of the video data may be performed based on the number of requests in the video distribution service.
  • the importance determination method is not limited to the method described above or a combination thereof, and may be any method according to the configuration or purpose of the monitoring system or the video distribution system.
  • the determination of the importance may not be based on the captured video data.
  • the importance of the camera ex102 that transmits video data to a terminal other than the video information processing apparatus ex101 may be set high.
  • the importance of the camera ex102 that transmits video data to a terminal other than the video information processing apparatus ex101 may be set low.
  • the analysis unit ex112 may determine the importance of the video data using the free viewpoint video and the video shot by the camera ex102.
  • the video information processing apparatus ex101 transmits a communication speed instruction signal to the camera ex102 based on the importance determination result performed by the analysis unit ex112. For example, the video information processing apparatus ex101 instructs a high communication speed to the camera ex102 that captures a video with high importance. Further, the video information processing apparatus ex101 may transmit not only the speed control but also a signal instructing a method in which important information is transmitted a plurality of times in order to reduce a disadvantage caused by the lack. Thereby, communication within the facility or the entire stadium can be performed efficiently. Communication between the camera ex102 and the video information processing apparatus ex101 may be wired communication or wireless communication. The video information processing apparatus ex101 may control only one of wired communication and wireless communication.
  • the camera ex102 transmits the captured video data to the video information processing apparatus ex101 at a communication speed according to the communication speed instruction signal. Note that if the retransmission of the camera ex102 fails a predetermined number of times, the camera ex102 may stop the retransmission of the captured video data and start the transfer of the next captured video data. As a result, communication within the facility or the entire stadium can be efficiently performed, and high-speed processing in the analysis unit ex112 can be realized.
  • the video data of the bit rate capable of transmitting the captured video data at the assigned communication speed may be transmitted, or the video data transfer may be stopped.
  • the camera ex102 when video data is used to prevent the generation of blind spots, only a part of the shooting area included in the captured video data may be necessary to fill the blind spots. There is sex.
  • the camera ex102 generates the extracted video data by extracting at least the area necessary for preventing the generation of the blind spot from the video data, and the generated extracted video data is used as the video information processing apparatus. You may transmit to ex101. According to this configuration, the occurrence of blind spots can be suppressed with a smaller communication band.
  • the camera ex102 needs to transmit the position information of the camera ex102 and the shooting direction information to the video information processing apparatus ex101.
  • the camera ex102 to which only a bandwidth that is not sufficient for transferring the video data may be transmitted, only the position information detected by the camera ex102 and the information on the shooting direction.
  • the video information processing apparatus ex101 estimates position information and shooting direction information of the camera ex102
  • the camera ex102 converts the shot video data to a resolution necessary for estimating the position information and shooting direction information.
  • the converted video data may be transmitted to the video information processing apparatus ex101.
  • the video information processing apparatus ex101 can acquire shooting area information from a larger number of cameras ex102, for example, when the shooting area information is used for the purpose of detecting a focused area, for example. It is valid.
  • the switching of the video data transfer process according to the allocated communication band described above may be performed by the camera ex102 based on the notified communication band, or the video information processing apparatus ex101 performs the operation of each camera ex102.
  • the control signal indicating the determined operation may be notified to each camera ex102.
  • the processing can be appropriately shared according to the calculation amount necessary for determining the switching of the operation, the processing capability of the camera ex102, the necessary communication band, and the like.
  • the analysis unit ex112 may determine the importance of the video data based on the visual field information (and / or viewpoint information) transmitted from the video reception device ex103. For example, the analysis unit ex112 sets the importance of captured video data including many areas indicated by the visual field information (and / or viewpoint information) to be high. The analysis unit ex112 may determine the importance of the video data in consideration of the number of people included in the video or the number of moving objects. Note that the importance determination method is not limited to this.
  • the communication control method described in the present embodiment is not necessarily used in a system that reconstructs a three-dimensional shape from a plurality of video data.
  • the communication control method described in the present embodiment is It is valid.
  • the video information processing apparatus ex101 may transmit an overview video showing the entire shooting scene to the video receiving apparatus ex103.
  • the video information processing apparatus ex101 when the video information processing apparatus ex101 receives the distribution request transmitted from the video receiving apparatus ex103, the video information processing apparatus ex101 reads an overview video of the entire facility or stadium from the storage unit ex111, and the external video is received by the video receiving apparatus. send to ex103.
  • the overview video may have a long update interval (may be a low frame rate) or may have a low image quality.
  • the viewer touches a portion to be seen in the overview video displayed on the screen of the video receiving device ex103. Accordingly, the video reception device ex103 transmits visual field information (and / or viewpoint information) corresponding to the touched portion to the video information processing device ex101.
  • the video information processing apparatus ex101 reads video data corresponding to the visual field information (and / or viewpoint information) from the storage unit ex111, and transmits the video data to the video receiving apparatus ex103.
  • the analysis unit ex112 generates a free viewpoint video by preferentially restoring the three-dimensional shape (three-dimensional reconstruction) on the region indicated by the visual field information (and / or viewpoint information).
  • the analysis unit ex112 restores the three-dimensional shape of the entire facility or the stadium with an accuracy that shows an overview.
  • the video information processing apparatus ex101 can efficiently restore the three-dimensional shape. As a result, it is possible to realize a high frame rate and high image quality of a free viewpoint video in an area desired by the viewer.
  • the video information processing apparatus ex101 may store in advance, for example, three-dimensional shape restoration data of a facility or a stadium generated in advance from a design drawing or the like as a preliminary video.
  • the prior image is not limited to this, and may be virtual space data obtained by mapping, for each object, the unevenness of the space obtained from the depth sensor and the picture derived from the image or the image data at the past or during calibration.
  • the analysis unit ex112 when soccer is being performed in a stadium, the analysis unit ex112 performs reconstruction of a three-dimensional shape limited to only players and balls, and combines the obtained restoration data and a prior image to generate a free viewpoint video. May be generated.
  • the analysis unit ex112 may preferentially restore the three-dimensional shape with respect to the player and the ball.
  • the video information processing apparatus ex101 can efficiently restore the three-dimensional shape.
  • the analysis unit ex112 may perform the reconstruction of the three-dimensional shape by limiting to only the person and the moving object or giving priority to them.
  • the time of each device may be calibrated at the start of shooting based on the reference time of the server.
  • the analysis unit ex112 uses a plurality of video data captured at a time that falls within a preset time range according to the accuracy of time setting among a plurality of captured video data captured by the plurality of cameras ex102. 3D shape restoration. For the detection of this time, for example, the time when the captured video data is stored in the storage unit ex111 is used. The time detection method is not limited to this. As a result, the video information processing apparatus ex101 can efficiently restore the three-dimensional shape, thereby realizing a high frame rate and high image quality of the free viewpoint video.
  • the analysis unit ex112 may restore the three-dimensional shape using only the high-quality data or using the high-quality data preferentially among the plurality of video data stored in the storage unit ex111. .
  • the analysis unit ex112 may restore the three-dimensional shape using the camera attribute information.
  • the camera ex102 transmits the captured video data and camera attribute information to the video information processing apparatus ex101.
  • the camera attribute information is, for example, a shooting position, a shooting angle, a shooting time, or a zoom magnification.
  • the video information processing apparatus ex101 can efficiently restore the three-dimensional shape, it is possible to realize a high frame rate and high image quality of the free viewpoint video.
  • the camera ex102 defines three-dimensional coordinates in the facility or in the stadium, and information about which coordinates the camera ex102 took from which angle, how much zoom, and at what time, along with the video. It transmits to the video information processing apparatus ex101 as camera attribute information. Further, when the camera ex102 is activated, the clock on the communication network in the facility or stadium is synchronized with the clock in the camera, and time information is generated.
  • FIG. 18 is a diagram illustrating an example of a notification displayed on the screen of the camera ex102 when the camera ex102 is activated.
  • the camera ex102 is moved from the camera ex102.
  • the vector information up to the advertisement is acquired and the reference of the camera position and angle is specified.
  • the camera coordinates and angle at that time are specified from the motion information of the camera ex102.
  • the display is not limited to this, and a display that uses an arrow or the like to indicate coordinates, an angle, a moving speed of the imaging region, or the like during the imaging period may be used.
  • the coordinates of the camera ex102 may be specified using GPS, WiFi (registered trademark), 3G, LTE (Long Term Evolution), and 5G (wireless LAN) radio waves, or a beacon (Bluetooth (registered trademark)). , Ultrasonic), or the like. Further, information on which base station in the facility or stadium the captured video data has reached may be used.
  • the system may be provided as an application that operates on a mobile terminal such as a smartphone.
  • An account such as various SNSs may be used to log in to the system.
  • An application-dedicated account or a guest account with limited functions may be used.
  • By using the account in this way it is possible to evaluate a favorite video or a favorite account.
  • video data having a viewpoint similar to the viewpoint of the video data being shot or viewed the resolution of these video data Can be increased. Thereby, it is possible to restore the three-dimensional shape from these viewpoints with higher accuracy.
  • the user can select a preferred image video in the application and follow the other party, so that the selected image can be viewed with priority over other users, or can be used for text chat, etc., subject to the other party's approval. You can have a connection. In this way, a new community can be generated.
  • the user can edit an image or video taken by another person or create a new image or video by collaging the image of another person with his own image.
  • This makes it possible to share a new video work, such as sharing a new image or video only with people in the community.
  • a video work can be used for augmented reality games by inserting a CG character in this editing.
  • 3D model data can be sequentially output, so that a 3D printer or the like of a facility can output a 3D object based on 3D model data in a characteristic scene such as a goal scene. .
  • a 3D printer or the like of a facility can output a 3D object based on 3D model data in a characteristic scene such as a goal scene.
  • an object based on the scene during the game can be sold as a souvenir such as a key holder, or distributed to participating users.
  • the center identifies areas where there is a high possibility of crimes based on crime maps based on the results of analysis using past crime data, etc., or areas related to the crime occurrence probability identified in this way Holds data.
  • the frequency of image transmission / reception may be increased, or the image may be changed to a moving image.
  • a moving image or three-dimensional reconstruction data using SfM or the like may be used.
  • the center or each terminal simultaneously corrects an image or virtual space using information from other sensors such as a depth sensor or a thermo sensor, so that the police officer can grasp the situation more accurately.
  • the center can feed back the object information to a plurality of terminals by using the 3D reconstruction data. This allows individuals with each terminal to track the object.
  • an in-vehicle camera that takes pictures outside the vehicle is obligatory in some countries. Even in such an in-vehicle camera, by using three-dimensional data modeled from a plurality of images, it is possible to more accurately grasp the weather in the direction of the destination, the state of the road surface, the degree of traffic congestion, and the like.
  • the storage medium may be any medium that can record a program, such as a magnetic disk, an optical disk, a magneto-optical disk, an IC card, and a semiconductor memory.
  • the system includes an apparatus using an image processing method.
  • Other configurations in the system can be appropriately changed according to circumstances.
  • FIG. 19 is a diagram showing an overall configuration of a content supply system ex200 that realizes a content distribution service.
  • the communication service providing area is divided into desired sizes, and base stations ex206, ex207, ex208, ex209, and ex210, which are fixed wireless stations, are installed in each cell.
  • This content supply system ex200 includes a computer ex211, a PDA (Personal Digital Assistant) ex212, a camera ex213, a smartphone ex214, a game machine ex215, etc. via the Internet ex201, the Internet service provider ex202, the communication network ex204, and the base stations ex206 to ex210. Are connected.
  • a PDA Personal Digital Assistant
  • each device may be directly connected to a communication network ex204 such as a telephone line, cable television, or optical communication without going through the base stations ex206 to ex210 which are fixed wireless stations.
  • the devices may be directly connected to each other via short-range wireless or the like.
  • the camera ex213 is a device that can shoot a moving image such as a digital video camera
  • the camera ex216 is a device that can shoot a still image and a moving image such as a digital camera.
  • the smartphone ex214 is a GSM (registered trademark) (Global System for Mobile Communications) method, a CDMA (Code Division Multiple Access) method, a W-CDMA (Wideband-Code Division MultipleL method, or a Multiple Acceleration method).
  • GSM Global System for Mobile Communications
  • CDMA Code Division Multiple Access
  • W-CDMA Wideband-Code Division MultipleL method
  • Multiple Acceleration method a Multiple Acceleration method
  • HSPA High Speed Packet Access
  • a smartphone corresponding to a communication method using a high frequency band or a PHS (Personal Handyphone System)
  • PHS Personal Handyphone System
  • the camera ex213 and the like are connected to the streaming server ex203 through the base station ex209 and the communication network ex204, thereby enabling live distribution and the like.
  • live distribution content (for example, music live video) that the user captures using the camera ex213 is encoded and transmitted to the streaming server ex203.
  • the streaming server ex203 streams the content data transmitted to the requested client.
  • the client include a computer ex211, a PDA ex212, a camera ex213, a smartphone ex214, and a game machine ex215 that can decode the encoded data.
  • Each device that receives the distributed data decodes the received data and reproduces it.
  • the encoded processing of the captured data may be performed by the camera ex213, the streaming server ex203 that performs the data transmission processing, or may be performed in a shared manner.
  • the decryption processing of the distributed data may be performed by the client, the streaming server ex203, or may be performed in common with each other.
  • still images and / or moving image data captured by the camera ex216 may be transmitted to the streaming server ex203 via the computer ex211.
  • the encoding process in this case may be performed by any of the camera ex216, the computer ex211, and the streaming server ex203, or may be performed in a shared manner.
  • a plurality of devices connected to the system may be linked to display the same image, or the entire image is displayed on a device having a large display unit, and the smartphone ex214 or the like displays the image. A part of the area may be enlarged and displayed.
  • these encoding / decoding processes are generally performed in the computer ex211 and the LSI ex500 included in each device.
  • the LSI ex500 may be configured as a single chip or a plurality of chips.
  • moving image encoding / decoding software is incorporated into some recording media (CD-ROM, flexible disk, hard disk, etc.) that can be read by the computer ex211 etc., and encoding / decoding processing is performed using the software. May be.
  • moving image data acquired by the camera may be transmitted. The moving image data at this time is data encoded by the LSI ex500 included in the smartphone ex214.
  • the streaming server ex203 may be a plurality of servers or a plurality of computers, and may process, record, and distribute data in a distributed manner.
  • the client can receive and reproduce the encoded data.
  • the information transmitted by the user can be received, decrypted and reproduced by the client in real time, and even a user who does not have special rights or facilities can realize personal broadcasting.
  • multiplexed data obtained by multiplexing music data and the like on video data is transmitted to a communication or satellite ex302 via radio waves.
  • This video data is data encoded by the moving image encoding method described in the above embodiments.
  • the broadcasting satellite ex302 transmits a radio wave for broadcasting, and this radio wave is received by a home antenna ex304 capable of receiving satellite broadcasting.
  • the received multiplexed data is decoded and reproduced by a device such as the television (receiver) ex400 or the set top box (STB) ex317.
  • a recording medium ex315 such as DVD or BD, or a memory ex316 such as SD
  • encodes a video signal in the recording medium ex315 or memory ex316 and in some cases, a music signal
  • the moving picture decoding apparatus or moving picture encoding apparatus described in each of the above embodiments in the reader / recorder ex318 that writes in a multiplexed manner.
  • the reproduced video signal is displayed on the monitor ex319, and the video signal can be reproduced in another device or system by the recording medium ex315 in which the multiplexed data is recorded or the memory ex316.
  • a moving picture decoding apparatus may be mounted in a set-top box ex317 connected to a cable ex303 for cable television or an antenna ex304 for satellite / terrestrial broadcasting, and this may be displayed on a monitor ex319 of the television.
  • the moving picture decoding apparatus may be incorporated in the television instead of the set top box.
  • FIG. 21 is a diagram showing the smartphone ex214.
  • FIG. 22 is a diagram illustrating a configuration example of the smartphone ex214.
  • the smartphone ex214 includes an antenna ex450 for transmitting and receiving radio waves to and from the base station ex210, a camera unit ex465 that can take a video and a still image, a video captured by the camera unit ex465, a video received by the antenna ex450, and the like.
  • a display unit ex458 such as a liquid crystal display for displaying the decrypted data is provided.
  • the smartphone ex214 further includes an operation unit ex466 such as a touch panel, an audio output unit ex457 such as a speaker for outputting audio, an audio input unit ex456 such as a microphone for inputting audio, a captured video, a still image , A memory portion ex467 that can store the recorded audio, or the encoded data or the decoded data such as received video, still image, or mail, or the memory ex316 illustrated in FIG. And a slot part ex464 which is an interface part with the SIMex 468 for authenticating access to various data including the network.
  • an operation unit ex466 such as a touch panel
  • an audio output unit ex457 such as a speaker for outputting audio
  • an audio input unit ex456 such as a microphone for inputting audio
  • a memory portion ex467 that can store the recorded audio, or the encoded data or the decoded data such as received video, still image, or mail, or the memory ex316 illustrated in FIG.
  • a slot part ex464 which is an interface part with the SIMex 468
  • the smartphone ex214 controls the power supply circuit ex461, the operation input control unit ex462, the video signal processing unit ex455, the camera interface unit ex463, the LCD (for the main control unit ex460 that comprehensively controls the display unit ex458, the operation unit ex466, and the like.
  • a Liquid Crystal Display) control unit ex459, a modulation / demodulation unit ex452, a multiplexing / demultiplexing unit ex453, an audio signal processing unit ex454, a slot unit ex464, and a memory unit ex467 are connected to each other via a bus ex470.
  • the power supply circuit unit ex461 starts up the smartphone ex214 in an operable state by supplying power from the battery pack to each unit.
  • the smartphone ex214 converts the audio signal collected by the audio input unit ex456 in the audio call mode into a digital audio signal by the audio signal processing unit ex454 based on the control of the main control unit ex460 having a CPU, a ROM, a RAM, and the like. This is subjected to spectrum spread processing by the modulation / demodulation unit ex452, and is subjected to digital analog conversion processing and frequency conversion processing by the transmission / reception unit ex451, and then transmitted via the antenna ex450.
  • the smartphone ex214 amplifies reception data received via the antenna ex450 in the voice call mode, performs frequency conversion processing and analog-digital conversion processing, performs spectrum despreading processing in the modulation / demodulation unit ex452, and performs voice signal processing unit ex454. After being converted into an analog audio signal, the audio output unit ex457 outputs it.
  • the text data of the e-mail input by the operation of the operation unit ex466 of the main unit is sent to the main control unit ex460 via the operation input control unit ex462.
  • the main control unit ex460 performs spread spectrum processing on the text data in the modulation / demodulation unit ex452, performs digital analog conversion processing and frequency conversion processing in the transmission / reception unit ex451, and then transmits the text data to the base station ex210 via the antenna ex450.
  • almost the reverse process is performed on the received data and output to the display unit ex458.
  • the video signal processing unit ex455 compresses the video signal supplied from the camera unit ex465 by the moving image encoding method described in each of the above embodiments.
  • the encoded video data is sent to the multiplexing / demultiplexing unit ex453.
  • the audio signal processing unit ex454 encodes the audio signal picked up by the audio input unit ex456 while the camera unit ex465 captures video, still images, and the like, and sends the encoded audio data to the multiplexing / separating unit ex453. To do.
  • the multiplexing / demultiplexing unit ex453 multiplexes the encoded video data supplied from the video signal processing unit ex455 and the encoded audio data supplied from the audio signal processing unit ex454 by a predetermined method, and is obtained as a result.
  • the multiplexed data is subjected to spread spectrum processing by a modulation / demodulation unit (modulation / demodulation circuit unit) ex452, and subjected to digital analog conversion processing and frequency conversion processing by a transmission / reception unit ex451, and then transmitted through an antenna ex450.
  • the multiplexing / separating unit ex453 separates the multiplexed data into a video data bit stream and an audio data bit stream, and performs video signal processing on the video data encoded via the synchronization bus ex470.
  • the encoded audio data is supplied to the audio signal processing unit ex454 while being supplied to the unit ex455.
  • the video signal processing unit ex455 decodes the video signal by decoding using a video decoding method corresponding to the video encoding method shown in each of the above embodiments, and the display unit ex458 via the LCD control unit ex459. From, for example, video and still images included in a moving image file linked to a home page are displayed.
  • the audio signal processing unit ex454 decodes the audio signal, and the audio is output from the audio output unit ex457.
  • the terminal such as the smartphone ex214 is a transmission terminal having only an encoder and a receiving terminal having only a decoder, as well as the television ex400.
  • a possible implementation format is possible.
  • multiplexed data in which music data or the like is multiplexed with video data is received and transmitted.
  • data in which character data related to video is multiplexed in addition to audio data It may be video data itself instead of multiplexed data.
  • the present invention can be applied to a video distribution system that distributes video shot by a plurality of cameras.

Abstract

An image delivery method, which is executed by a server (103) that delivers to a terminal apparatus (102) a plurality of images captured by a plurality of users at different viewpoints, comprises: a delivery step (S142) of delivering to the terminal apparatus (102) a first image that is one of the plurality of images and that has been requested from the terminal apparatus (102); a selection step (S143) of selecting, from the plurality of images, a second image that will be most probably requested from the terminal apparatus (102) next; and a transmission step (S144) of starting transmission of the second image to the terminal apparatus (102) while delivering the first image to the terminal apparatus (102).

Description

映像配信方法、映像受信方法、サーバ、端末装置及び映像配信システムVideo distribution method, video reception method, server, terminal device, and video distribution system
 本発明は、複数の視点から撮影された映像を配信する映像配信方法等に関する。 The present invention relates to a video distribution method for distributing video shot from a plurality of viewpoints.
 映像配信方法として、例えば、特許文献1記載の技術が知られている。また、複数の視点から撮影された映像を配信する映像配信方法が知られている(例えば、特許文献2参照)。このような映像配信方法では、特定のシーンを異なる視点から撮影した複数の映像から、ユーザが任意の映像を指定し、視聴することができる。 As a video distribution method, for example, a technique described in Patent Document 1 is known. In addition, a video distribution method for distributing video shot from a plurality of viewpoints is known (see, for example, Patent Document 2). In such a video distribution method, a user can designate and view an arbitrary video from a plurality of videos obtained by shooting a specific scene from different viewpoints.
特開2009-206625号公報JP 2009-206625 A 特開2012-094990号公報JP 2012-094990 A
 このような、映像配信方法では、映像の切り替えをスムーズに行なえることが望まれている。 In such a video distribution method, it is desired that the video can be switched smoothly.
 そこで、本発明は、映像の切り替えをスムーズに行なえる映像配信方法又は映像受信方法を提供することを目的とする。 Therefore, an object of the present invention is to provide a video distribution method or a video reception method capable of smoothly switching video.
 上記目的を達成するために、本発明の一態様に係る映像配信方法は、複数のユーザにより異なる視点から撮影された複数の映像のいずれかを端末装置に配信するサーバによる映像配信方法であって、前記複数の映像の一つであり、前記端末装置から要求された第1映像を前記端末装置に配信する配信ステップと、前記複数の映像の一つであり、前記端末装置から次に要求される可能性が高い第2映像を選択する選択ステップと、前記第1映像を前記端末装置に配信している間に前記第2映像の前記端末装置への送信を開始する送信ステップとを含む。 In order to achieve the above object, a video distribution method according to an aspect of the present invention is a video distribution method by a server that distributes to a terminal device any one of a plurality of videos taken from different viewpoints by a plurality of users. A distribution step of distributing the first video requested by the terminal device to the terminal device; and one of the plurality of videos, which is next requested by the terminal device. A selection step of selecting a second video that is highly likely to be transmitted, and a transmission step of starting transmission of the second video to the terminal device while the first video is being distributed to the terminal device.
 また、本発明の一態様に係る映像受信方法は、複数の視点から撮影された複数の映像のいずれかをサーバから受信し、当該映像を表示する端末装置による映像受信方法であって、前記複数の映像から第1映像を選択する選択ステップと、前記第1映像の送信を前記サーバに要求する要求ステップと、前記第1映像を前記サーバから受信する第1受信ステップと、前記第1映像を表示する表示ステップと、前記第1映像の受信中に、前記複数の映像の一つであり、次に選択される可能性が高い第2映像の受信を開始する第2受信ステップとを含む。 A video reception method according to an aspect of the present invention is a video reception method by a terminal device that receives any of a plurality of videos taken from a plurality of viewpoints from a server and displays the videos. Selecting the first video from the video, a requesting step for requesting the server to transmit the first video, a first receiving step for receiving the first video from the server, and the first video A display step of displaying, and a second reception step of starting reception of a second video that is one of the plurality of videos and is likely to be selected next during reception of the first video.
 なお、これらの全般的または具体的な態様は、システム、方法、集積回路、コンピュータプログラムまたはコンピュータ読み取り可能なCD-ROMなどの記録媒体で実現されてもよく、システム、方法、集積回路、コンピュータプログラム及び記録媒体の任意な組み合わせで実現されてもよい。 These general or specific aspects may be realized by a system, a method, an integrated circuit, a computer program, or a recording medium such as a computer-readable CD-ROM. The system, method, integrated circuit, computer program Also, any combination of recording media may be realized.
 本発明は、映像の切り替えをスムーズに行なえる映像配信方法又は映像受信方法を提供できる。 The present invention can provide a video distribution method or video reception method capable of smoothly switching video.
実施の形態1に係る映像配信システムの構成を示す図である。It is a figure which shows the structure of the video delivery system which concerns on Embodiment 1. FIG. 実施の形態1に係るサーバのブロック図である。2 is a block diagram of a server according to Embodiment 1. FIG. 実施の形態1に係る端末装置のブロック図である。2 is a block diagram of a terminal device according to Embodiment 1. FIG. 実施の形態1に係る映像配信システムの処理を示す図である。4 is a diagram showing processing of the video distribution system according to Embodiment 1. FIG. 実施の形態1に係る初期画面の一例を示す図である。6 is a diagram illustrating an example of an initial screen according to Embodiment 1. FIG. 実施の形態1に係る関連映像選択処理の一例を示す図である。6 is a diagram showing an example of related video selection processing according to Embodiment 1. FIG. 実施の形態1に係る関連映像選択処理の一例を示す図である。6 is a diagram showing an example of related video selection processing according to Embodiment 1. FIG. 実施の形態1に係る関連映像選択処理の一例を示す図である。6 is a diagram showing an example of related video selection processing according to Embodiment 1. FIG. 実施の形態1に係る関連映像選択処理の一例を示す図である。6 is a diagram showing an example of related video selection processing according to Embodiment 1. FIG. 実施の形態1に係る関連映像選択処理の一例を示す図である。6 is a diagram showing an example of related video selection processing according to Embodiment 1. FIG. 実施の形態1に係る関連映像選択処理の一例を示す図である。6 is a diagram showing an example of related video selection processing according to Embodiment 1. FIG. 実施の形態1に係る表示画面の一例を示す図である。6 is a diagram showing an example of a display screen according to Embodiment 1. FIG. 実施の形態1に係る端末装置による処理のフローチャートである。3 is a flowchart of processing performed by a terminal device according to Embodiment 1; 実施の形態1に係る端末装置による処理の変形例のフローチャートである。10 is a flowchart of a modification of processing by the terminal device according to the first embodiment. 実施の形態1に係る端末装置による処理の変形例のフローチャートである。10 is a flowchart of a modification of processing by the terminal device according to the first embodiment. 実施の形態1に係るサーバによる処理のフローチャートである。3 is a flowchart of processing performed by a server according to the first embodiment. 映像情報処理システムの構成を示す図である。It is a figure which shows the structure of a video information processing system. カメラ起動時に表示される通知画面の一例を示す図である。It is a figure which shows an example of the notification screen displayed at the time of camera starting. コンテンツ配信サービスを実現するコンテンツ供給システムの全体構成図である。1 is an overall configuration diagram of a content supply system that realizes a content distribution service. デジタル放送用システムの全体構成図である。1 is an overall configuration diagram of a digital broadcasting system. スマートフォンの一例を示す図である。It is a figure which shows an example of a smart phone. スマートフォンの構成例を示すブロック図である。It is a block diagram which shows the structural example of a smart phone.
 (本発明の基礎となった知見)
 異なる視点の複数の映像を配信する場合には、ユーザが視聴対象の映像を選択し、サーバから端末装置に選択された映像が配信される。これにより、ユーザが映像を選択してから、映像が表示されるまでに待ち時間が発生する可能性がある。なお、通信帯域には制限があるため、全ての映像を予め端末装置に送信しておくことは困難である。
(Knowledge that became the basis of the present invention)
When distributing a plurality of videos from different viewpoints, the user selects a video to be viewed, and the selected video is distributed from the server to the terminal device. As a result, there is a possibility that a waiting time may occur after the user selects a video until the video is displayed. Since the communication band is limited, it is difficult to transmit all videos to the terminal device in advance.
 一方で、特許文献1には、視聴画像の周囲を含む大きな画像を送る方法が開示されている。また、特許文献2には、異なる視点の複数の映像のうち、表示視点映像の周辺の視点映像をグループ映像として配信する方法が開示されている。 On the other hand, Patent Document 1 discloses a method for sending a large image including the periphery of a viewing image. Patent Document 2 discloses a method of distributing a viewpoint video around a display viewpoint video among a plurality of videos with different viewpoints as a group video.
 しかしながら、複数のユーザにより任意の視点から撮影された映像を配信する場合には、上記技術では、シームレスな表示を行なうことが困難である。具体的には、上記技術では、撮影視点が予め定められているため、予めグループ映像等を判定することができる。一方、複数のユーザが任意に撮影した映像では、視点、画質及びズーム度などの撮影条件が任意に設定される。このような場合に、ユーザが好む視点からの映像をシームレスに表示することは困難である。 However, in the case of distributing video shot from an arbitrary viewpoint by a plurality of users, it is difficult to perform seamless display with the above technique. Specifically, in the above technique, since a shooting viewpoint is determined in advance, a group video or the like can be determined in advance. On the other hand, shooting conditions such as a viewpoint, image quality, and zoom degree are arbitrarily set for videos arbitrarily shot by a plurality of users. In such a case, it is difficult to seamlessly display the video from the viewpoint preferred by the user.
 本発明の一態様に係る映像配信方法は、複数のユーザにより異なる視点から撮影された複数の映像のいずれかを端末装置に配信するサーバによる映像配信方法であって、前記複数の映像の一つであり、前記端末装置から要求された第1映像を前記端末装置に配信する配信ステップと、前記複数の映像の一つであり、前記端末装置から次に要求される可能性が高い第2映像を選択する選択ステップと、前記第1映像を前記端末装置に配信している間に前記第2映像の前記端末装置への送信を開始する送信ステップとを含む。 A video distribution method according to an aspect of the present invention is a video distribution method by a server that distributes to a terminal device one of a plurality of videos taken from different viewpoints by a plurality of users, and one of the plurality of videos. A distribution step of distributing the first video requested from the terminal device to the terminal device, and a second video that is one of the plurality of videos and is likely to be requested next from the terminal device And a transmission step of starting transmission of the second video to the terminal device while delivering the first video to the terminal device.
 これによれば、第1映像の表示中に第2映像が予め端末装置に送られる。これにより、端末装置は、第1映像から第2映像への切り替えをスムーズに行なえる。 According to this, the second video is sent in advance to the terminal device during the display of the first video. Thereby, the terminal device can smoothly switch from the first video to the second video.
 例えば、前記選択ステップでは、前記複数の映像のうち、前記第1映像との関連度が高い映像を前記第2映像として選択する。 For example, in the selection step, a video having a high degree of association with the first video is selected as the second video from the plurality of videos.
 これによれば、現在表示中の第1映像に関連度が高い第2映像が選択された場合に、端末装置は、映像の切り替えをスムーズに行なえる。 According to this, when a second video having a high degree of association with the first video currently displayed is selected, the terminal device can smoothly switch the video.
 例えば、前記選択ステップでは、撮影シーンの位置が前記第1映像の撮影シーンの位置に近いほど、前記関連度が高いと判定する。 For example, in the selection step, it is determined that the degree of association is higher as the position of the shooting scene is closer to the position of the shooting scene of the first video.
 例えば、前記選択ステップでは、さらに、前記撮影シーンの広さが前記第1映像の前記撮影シーンの広さに近いほど、前記関連度が高いと判定する。 For example, in the selection step, it is further determined that the degree of association is higher as the width of the shooting scene is closer to the width of the shooting scene of the first video.
 例えば、前記選択ステップでは、前記第1映像に含まれる被写体と、同一の被写体が撮影されている映像の前記関連度を高く設定する。 For example, in the selection step, the degree of association between the subject included in the first video and the video in which the same subject is shot is set high.
 例えば、前記選択ステップでは、前記複数の映像のフレームレート、解像度又はビットレートに基づき、前記第2映像を選択する。 For example, in the selection step, the second video is selected based on a frame rate, resolution, or bit rate of the plurality of videos.
 例えば、前記選択ステップでは、前記複数の映像のうち、他のユーザに選択された回数が多い映像を前記第2映像として選択する。 For example, in the selection step, an image that is frequently selected by another user among the plurality of images is selected as the second image.
 例えば、前記選択ステップでは、ユーザの視聴履歴又は予め登録されている嗜好情報に基づき、前記第2映像を選択する。 For example, in the selection step, the second video is selected based on the user's viewing history or pre-registered preference information.
 また、本発明の一態様に係る映像受信方法は、複数の視点から撮影された複数の映像のいずれかをサーバから受信し、当該映像を表示する端末装置による映像受信方法であって、前記複数の映像から第1映像を選択する選択ステップと、前記第1映像の送信を前記サーバに要求する要求ステップと、前記第1映像を前記サーバから受信する第1受信ステップと、前記第1映像を表示する表示ステップと、前記第1映像の受信中に、前記複数の映像の一つであり、次に選択される可能性が高い第2映像の受信を開始する第2受信ステップとを含む。 A video reception method according to an aspect of the present invention is a video reception method by a terminal device that receives any of a plurality of videos taken from a plurality of viewpoints from a server and displays the videos. Selecting the first video from the video, a requesting step for requesting the server to transmit the first video, a first receiving step for receiving the first video from the server, and the first video A display step of displaying, and a second reception step of starting reception of a second video that is one of the plurality of videos and is likely to be selected next during reception of the first video.
 これによれば、端末装置は、第1映像の表示中に予め第2映像を受信する。これにより、端末装置は、第1映像から第2映像への切り替えをスムーズに行なえる。 According to this, the terminal device receives the second video in advance during the display of the first video. Thereby, the terminal device can smoothly switch from the first video to the second video.
 例えば、前記映像受信方法は、さらに、前記受信された前記第2映像を蓄積するステップと、前記第1映像の表示中に前記第2映像が選択された場合に、前記蓄積されている前記第2映像を表示するステップとを含む。 For example, the video receiving method further includes a step of storing the received second video, and the second video stored when the second video is selected during display of the first video. Displaying two images.
 例えば、前記映像配信方法は、さらに、前記第1映像の表示中に、前記第1映像及び前記第2映像と異なる第3映像が選択された場合、前記第3映像を前記サーバから受信するステップと、前記第3映像を受信するまでの間、前記蓄積されている前記第2映像を表示するステップとを含む。 For example, the video distribution method further includes a step of receiving the third video from the server when a third video different from the first video and the second video is selected during display of the first video. And displaying the stored second video until the third video is received.
 これによれば、端末装置は、第1映像から他の映像への切り替えの待ち時間に第2映像を表示できる。 According to this, the terminal device can display the second video during the waiting time for switching from the first video to another video.
 例えば、前記表示ステップでは、さらに、前記複数の映像が撮影されている場所を俯瞰する画像であり、前記複数の視点の位置を示す複数のアイコンを含む画像を表示する。 For example, in the display step, an image overlooking a place where the plurality of videos are taken is displayed, and an image including a plurality of icons indicating the positions of the plurality of viewpoints is displayed.
 例えば、前記表示ステップでは、前記複数のアイコンのうち、前記第2映像の視点の位置を示すアイコンが強調表示される。 For example, in the display step, an icon indicating the position of the viewpoint of the second video among the plurality of icons is highlighted.
 これによれば、ユーザに第2映像を選択されやすくできる。 This makes it easier for the user to select the second video.
 また、本発明の一態様に係るサーバは、複数のユーザにより異なる視点から撮影された複数の映像のいずれかを端末装置に配信するサーバであって、前記複数の映像の一つであり、前記端末装置から指定された第1映像を前記端末装置に配信する配信部と、前記複数の映像の一つであり、前記端末装置から次に要求される可能性が高い第2映像を選択する選択部と、前記第1映像を前記端末装置に配信している間に前記第2映像の前記端末装置への送信を開始する送信部とを備える。 The server according to an aspect of the present invention is a server that distributes to a terminal device any one of a plurality of videos shot from different viewpoints by a plurality of users, and is one of the plurality of videos, A distribution unit that distributes the first video designated by the terminal device to the terminal device, and a selection for selecting a second video that is one of the plurality of videos and is likely to be requested next from the terminal device And a transmission unit that starts transmission of the second video to the terminal device while delivering the first video to the terminal device.
 これによれば、第1映像の表示中に第2映像が予め端末装置に送られる。これにより、端末装置は、第1映像から第2映像への切り替えをスムーズに行なえる。 According to this, the second video is sent in advance to the terminal device during the display of the first video. Thereby, the terminal device can smoothly switch from the first video to the second video.
 また、本発明の一態様に係る端末装置は、複数の視点から撮影された複数の映像のいずれかをサーバから受信し、当該映像を表示する端末装置であって、前記複数の映像から第1映像を選択する選択部と、前記第1映像の送信を前記サーバに要求する要求部と、前記第1映像を前記サーバから受信する第1受信部と、前記第1映像を表示する表示部と、前記第1映像の受信中に、前記複数の映像の一つであり、次に選択される可能性が高い第2映像の受信を開始する第2受信部とを備える。 A terminal device according to an aspect of the present invention is a terminal device that receives any of a plurality of videos taken from a plurality of viewpoints from a server and displays the videos. A selection unit that selects a video, a request unit that requests the server to transmit the first video, a first reception unit that receives the first video from the server, and a display unit that displays the first video A second receiving unit that starts receiving a second video that is one of the plurality of videos and is likely to be selected next during reception of the first video.
 これによれば、端末装置は、第1映像の表示中に予め第2映像を受信する。これにより、端末装置は、第1映像から第2映像への切り替えをスムーズに行なえる。 According to this, the terminal device receives the second video in advance during the display of the first video. Thereby, the terminal device can smoothly switch from the first video to the second video.
 また、本発明の一態様に係る映像配信システムは、サーバと端末装置を含む。 The video distribution system according to one aspect of the present invention includes a server and a terminal device.
 これによれば、第1映像の表示中に第2映像が予め端末装置に送られる。これにより、端末装置は、第1映像から第2映像への切り替えをスムーズに行なえる。 According to this, the second video is sent in advance to the terminal device during the display of the first video. Thereby, the terminal device can smoothly switch from the first video to the second video.
 なお、これらの包括的または具体的な態様は、システム、方法、集積回路、コンピュータプログラムまたはコンピュータ読み取り可能なCD-ROMなどの記録媒体で実現されてもよく、システム、方法、集積回路、コンピュータプログラム及び記録媒体の任意な組み合わせで実現されてもよい。 Note that these comprehensive or specific modes may be realized by a system, a method, an integrated circuit, a computer program, or a recording medium such as a computer-readable CD-ROM, and the system, method, integrated circuit, and computer program. Also, any combination of recording media may be realized.
 以下、実施の形態について、図面を参照しながら具体的に説明する。なお、以下で説明する実施の形態は、いずれも本発明の一具体例を示すものである。以下の実施の形態で示される数値、形状、材料、構成要素、構成要素の配置位置及び接続形態、ステップ、ステップの順序などは、一例であり、本発明を限定する主旨ではない。また、以下の実施の形態における構成要素のうち、最上位概念を示す独立請求項に記載されていない構成要素については、任意の構成要素として説明される。 Hereinafter, embodiments will be specifically described with reference to the drawings. Note that each of the embodiments described below shows a specific example of the present invention. The numerical values, shapes, materials, constituent elements, arrangement positions and connecting forms of the constituent elements, steps, order of steps, and the like shown in the following embodiments are merely examples, and are not intended to limit the present invention. In addition, among the constituent elements in the following embodiments, constituent elements that are not described in the independent claims indicating the highest concept are described as optional constituent elements.
 (実施の形態1)
 本実施の形態に係る映像配信システムでは、複数の映像のうち一部の映像を、予め端末装置に送信する。これにより、当該一部の映像が次に選択された場合に、シームレスに映像を切り替えることができる。
(Embodiment 1)
In the video distribution system according to the present embodiment, some of the plurality of videos are transmitted to the terminal device in advance. Thereby, when the said one part image | video is selected next, a video | video can be switched seamlessly.
 まず、本実施の形態に係る映像配信システムの構成を説明する。図1は、本実施の形態に係る映像配信システム100の構成を示すブロック図である。この映像配信システム100は、各々がネットワーク104を介して通信可能な、複数のカメラ101と、端末装置102と、サーバ103とを含む。 First, the configuration of the video distribution system according to the present embodiment will be described. FIG. 1 is a block diagram showing a configuration of a video distribution system 100 according to the present embodiment. The video distribution system 100 includes a plurality of cameras 101, a terminal device 102, and a server 103 that can communicate with each other via a network 104.
 複数のカメラ101は、異なる視点から同一のシーンを同一の時間帯に撮影することで複数の映像信号を生成する。各カメラ101は、複数のユーザの各々に携帯される。例えば、複数のカメラ101は、スポーツスタジアムのような場所にいる複数の観客に所有される。複数のカメラ101で撮影された複数の映像信号は、ネットワーク104を経由してサーバ103に送信される。また、この映像信号には、撮影視点(カメラ位置)、カメラの向き、及び倍率等を示す情報が含まれる。 The plurality of cameras 101 generate a plurality of video signals by photographing the same scene from different viewpoints in the same time zone. Each camera 101 is carried by each of a plurality of users. For example, the plurality of cameras 101 are owned by a plurality of spectators in a place such as a sports stadium. A plurality of video signals photographed by the plurality of cameras 101 are transmitted to the server 103 via the network 104. In addition, the video signal includes information indicating a photographing viewpoint (camera position), a camera direction, a magnification, and the like.
 なお、カメラ101は、少なくとも撮影機能を有する機器であればよく、例えば、デジタルスチルカメラ、デジタルビデオカメラ、スマートフォン、又は携帯端末等である。 The camera 101 only needs to be a device having at least a photographing function, such as a digital still camera, a digital video camera, a smartphone, or a mobile terminal.
 端末装置102は、ユーザが使用する端末であり、少なくとも映像を表示する機能を有する。例えば、この端末装置102は、スマートフォン、携帯端末又はパーソナルコンピュータ等である。なお、端末装置102は、カメラ101と同様の機能を有し、当該ユーザが上記観客に含まれていてもよいし、当該ユーザは、上記スタジアム以外の場所から映像を視聴してもよい。 The terminal device 102 is a terminal used by the user, and has at least a function of displaying an image. For example, the terminal device 102 is a smartphone, a portable terminal, a personal computer, or the like. Note that the terminal device 102 has the same function as the camera 101, and the user may be included in the audience, or the user may view a video from a place other than the stadium.
 サーバ103は、複数のカメラ101から送信された複数の映像信号を保持する。また、サーバ103は、端末装置102からの要求に従い、端末装置102に、保持する複数の映像信号の一部を送信する。また、サーバ103は、複数の映像信号の中身を解析し、得られた映像の特徴に基づき、複数の映像信号の関連度を算出する。さらに、サーバ103は、端末装置102から指定された選択映像信号に加え、選択映像信号に関連度の高い関連映像信号を端末装置102に送信する。 The server 103 holds a plurality of video signals transmitted from the plurality of cameras 101. Further, the server 103 transmits a part of the plurality of video signals to be held to the terminal device 102 in accordance with a request from the terminal device 102. In addition, the server 103 analyzes the contents of the plurality of video signals and calculates the relevance of the plurality of video signals based on the obtained video characteristics. Further, the server 103 transmits a related video signal having a high degree of relevance to the selected video signal to the terminal device 102 in addition to the selected video signal designated by the terminal device 102.
 なお、以下では、複数の映像信号が複数のカメラ101からリアルタイムに送信され、当該映像信号を、ユーザが端末装置102を用いてリアルタイムに視聴する例を述べるが、映像の送信及び視聴の少なくとも一方がリアルタイムに行なわれなくてもよい。また、以下で述べる、映像信号(映像)の送信及び受信とは、主に、映像信号が連続的に送信又は受信されるストリーム送信及び受信を意味する。 In the following, an example in which a plurality of video signals are transmitted from a plurality of cameras 101 in real time and the user views the video signals in real time using the terminal device 102 will be described. However, at least one of video transmission and viewing is described. May not be performed in real time. Further, the transmission and reception of video signals (video) described below mainly means stream transmission and reception in which video signals are continuously transmitted or received.
 以下、各装置の構成を説明する。図2はサーバ103の構成を示すブロック図である。このサーバ103は、受信部111と、映像蓄積部112と、制御部113と、送信部114とを備える。 The configuration of each device will be described below. FIG. 2 is a block diagram showing the configuration of the server 103. The server 103 includes a reception unit 111, a video storage unit 112, a control unit 113, and a transmission unit 114.
 受信部111は、複数のカメラ101で同一のシーンが異なる視点から撮影された複数の映像信号151を受信する。また、受信部111は、端末装置102から送信された視点指定信号152を受信する。この視点指定信号152は、複数の映像信号151のうちいずれかを指定する。 The receiving unit 111 receives a plurality of video signals 151 in which the same scene is captured from different viewpoints by the plurality of cameras 101. In addition, the reception unit 111 receives the viewpoint designation signal 152 transmitted from the terminal device 102. This viewpoint designation signal 152 designates one of the plurality of video signals 151.
 映像蓄積部112は、受信部111が受信した複数の映像信号151を蓄積する。 The video storage unit 112 stores a plurality of video signals 151 received by the reception unit 111.
 制御部113は、映像蓄積部112に蓄積されている複数の映像信号151から視点指定信号152で指定される映像信号151を選択映像信号153として選択し、送信部114を経由して、選択映像信号153を端末装置102へ送信する。また、制御部113は、映像蓄積部112に蓄積されている複数の映像信号151から、選択映像信号153に関連する関連映像信号154を選択し、送信部114を経由して、当該関連映像信号154を端末装置102へ送信する。 The control unit 113 selects the video signal 151 designated by the viewpoint designation signal 152 from the plurality of video signals 151 stored in the video storage unit 112 as the selected video signal 153, and selects the selected video via the transmission unit 114. The signal 153 is transmitted to the terminal device 102. In addition, the control unit 113 selects a related video signal 154 related to the selected video signal 153 from the plurality of video signals 151 stored in the video storage unit 112, and the related video signal is transmitted via the transmission unit 114. 154 is transmitted to the terminal device 102.
 図3は端末装置102のブロック図である。この端末装置102は、受信部121と、蓄積部122と、復号部123と、出力部124と、送信部125と、制御部126と、入力部127とを備える。 FIG. 3 is a block diagram of the terminal device 102. The terminal device 102 includes a receiving unit 121, a storage unit 122, a decoding unit 123, an output unit 124, a transmission unit 125, a control unit 126, and an input unit 127.
 受信部121は、サーバ103から送信された選択映像信号153及び関連映像信号154を受信する。蓄積部122は、受信部121が受信した選択映像信号153及び関連映像信号154を一時的に保持する。 The receiving unit 121 receives the selected video signal 153 and the related video signal 154 transmitted from the server 103. The storage unit 122 temporarily holds the selected video signal 153 and the related video signal 154 received by the receiving unit 121.
 復号部123は、選択映像信号153を復号することで復号映像を生成する。出力部124は、復号映像を含む出力映像155を生成し、当該出力映像155を、例えば、端末装置102が備えるディスプレイ等の表示装置に表示する。 The decoding unit 123 generates a decoded video by decoding the selected video signal 153. The output unit 124 generates an output video 155 including the decoded video, and displays the output video 155 on a display device such as a display provided in the terminal device 102, for example.
 入力部127は、ユーザ操作を受け付ける。例えば、入力部127は、端末装置102が備えるタッチパネルに対するユーザの操作を受け付ける。制御部126は、入力部127が、ユーザによる視点変更の操作を受け付けた場合に、変更先の視点を示す視点指定信号152を、送信部125を経由して、サーバ103へ送信する。 The input unit 127 receives a user operation. For example, the input unit 127 receives a user operation on the touch panel provided in the terminal device 102. When the input unit 127 receives an operation of changing the viewpoint by the user, the control unit 126 transmits a viewpoint designation signal 152 indicating the viewpoint to be changed to the server 103 via the transmission unit 125.
 次に、映像配信システム100の動作を説明する。図4は、映像配信システム100における映像配信処理のシーケンス図である。なお、図4では、複数の映像信号151が既にサーバ103に保持されている。なお、複数の映像信号151は、上記スタジアムの例においてユーザがスタジアムの観客である場合のように、複数のカメラ101からリアルタイムにアップデートされている映像であってもよいし、予めサーバ103に保持された過去の映像であってもよい。 Next, the operation of the video distribution system 100 will be described. FIG. 4 is a sequence diagram of video distribution processing in the video distribution system 100. In FIG. 4, a plurality of video signals 151 are already held in the server 103. Note that the plurality of video signals 151 may be videos updated in real time from the plurality of cameras 101 as in the case of the stadium, in which the user is a stadium spectator. It may be a past video.
 まず、端末装置102は、例えば、ユーザ操作に従いアプリケーションプログラム(アプリ)を起動する(S101)。次に、端末装置102は、初期画面を表示する(S102)。具体的には、端末装置102は、サーバ103から初期情報として、複数の映像信号151が撮影された際の複数のカメラ101の位置(視点位置)を示す情報を受信し、複数のカメラ101の位置を示す情報を初期画面として表示する。 First, for example, the terminal device 102 starts an application program (application) in accordance with a user operation (S101). Next, the terminal device 102 displays an initial screen (S102). Specifically, the terminal apparatus 102 receives information indicating the positions (viewpoint positions) of the plurality of cameras 101 when the plurality of video signals 151 are captured as initial information from the server 103. Information indicating the position is displayed as an initial screen.
 図5は、この初期画面の一例を示す図である。背景画像201には、複数の映像が撮影されている場所を俯瞰する画像が使用される。また、視聴可能な映像の視点位置であり、当該映像を撮影したカメラ101の位置を示すカメラアイコン202が背景画像201上に表示される。 FIG. 5 is a diagram showing an example of this initial screen. As the background image 201, an image overlooking a place where a plurality of videos are taken is used. Further, a camera icon 202 indicating the position of the viewable video and the position of the camera 101 that shot the video is displayed on the background image 201.
 なお、カメラアイコン202の代わりに、又は、カメラアイコン202に加え、サムネイルが表示されてもよい。さらに、初期画面が拡大された場合に、カメラアイコン202の変わりにサムネイルが表示されてもよい。 Note that thumbnails may be displayed instead of the camera icon 202 or in addition to the camera icon 202. Furthermore, a thumbnail may be displayed instead of the camera icon 202 when the initial screen is enlarged.
 また、映像の数が多い場合には、後述する関連度等に基づき、ユーザへのお薦め度が高い映像のカメラアイコン202又はサムネイルだけが表示されてもよい。また、サムネイルが表示され場合には、サムネイルはカメラアイコン202より大きく表示されてもよい。 Also, when the number of videos is large, only the camera icon 202 or the thumbnail of the video having a high recommendation level for the user may be displayed based on the degree of relevance described later. When thumbnails are displayed, the thumbnails may be displayed larger than the camera icon 202.
 また、映像の数が多い場合には、関連性の高い映像がグループ化され、各グループ又は各グループの代表映像に対してカメラアイコン202が表示されてもよい。ここで代表映像は、例えば、映像の特性(解像度、フレームレート又はビットレート等)に基づき決定される。例えば解像度が一番高い映像、フレームレートが一番高い映像、又はビットレートが一番高い映像が代表映像に決定される。 Further, when the number of videos is large, highly relevant videos may be grouped, and the camera icon 202 may be displayed for each group or a representative video of each group. Here, the representative video is determined based on, for example, video characteristics (resolution, frame rate, bit rate, or the like). For example, a video having the highest resolution, a video having the highest frame rate, or a video having the highest bit rate is determined as the representative video.
 また、各グループの関連内容を示す情報が、カメラアイコン202と合せて表示されてもよい。また、カメラアイコン202の代わりに、各グループの代表映像のサムネイル、又は、縮小された映像が表示されてもよい。 Also, information indicating the related contents of each group may be displayed together with the camera icon 202. Further, instead of the camera icon 202, thumbnails of representative videos of each group or reduced videos may be displayed.
 ここで、各グループの代表映像はクリックされる可能性が高い。このため、端末装置102は、予め代表映像をサーバ103から受信しておいてもよい。つまり、端末装置102は、初期画面の表示時に各グループの代表映像を全て受信してもよい。または、端末装置102は、あるグループ又は代表映像が選択された場合に、そのグループに含まれる映像の一部又は全てをサーバ103から受信してもよい。 Here, the representative video of each group is highly likely to be clicked. For this reason, the terminal apparatus 102 may receive the representative video from the server 103 in advance. That is, the terminal device 102 may receive all the representative videos of each group when the initial screen is displayed. Alternatively, when a certain group or representative video is selected, the terminal device 102 may receive a part or all of the video included in the group from the server 103.
 また、端末装置102は、受信し始めてからしばらく経過し蓄積部122に十分データを蓄積した映像に対応するカメラアイコン202のみを選択可能な状態に設定してもよい。 Further, the terminal device 102 may set only the camera icon 202 corresponding to the video that has been stored for a while in the storage unit 122 after a while from the start of reception.
 また、端末装置102は、ユーザ操作に応じて画面が拡大又は縮小されても、表示されるカメラアイコン202の数が一定になるように、表示する視点を選択してもよい。 Further, the terminal device 102 may select the viewpoint to be displayed so that the number of camera icons 202 to be displayed is constant even when the screen is enlarged or reduced according to a user operation.
 また、初期画面の背景画像201は、ユーザが現在いる位置によって切替えられてもよい。例えば、ユーザがスタジアムの内野側スタンドにいる場合は、内野側スタンドから見える風景の画像が背景画像201に設定され、ユーザが外野側スタンドにいる場合は外野スタンドから見える風景の画像が背景画像201に設定されてもよい。 Further, the background image 201 of the initial screen may be switched depending on the position where the user is currently present. For example, when the user is on the infield side stand of the stadium, the landscape image seen from the infield side stand is set as the background image 201, and when the user is on the outfield side stand, the landscape image seen from the outfield stand is the background image 201. May be set.
 また、初期画面に表示するカメラアイコン202が背景画像201に合せて切替えられてもよい。また、カメラアイコン202はユーザの位置に応じて切替えられてもよい。例えば、ユーザが内野側スタンドにいる場合は、内野側スタンドから見える風景の画像が背景画像201に設定され、その風景内に存在する撮影視点を示すカメラアイコン202が表示されてもよい。 Also, the camera icon 202 displayed on the initial screen may be switched according to the background image 201. The camera icon 202 may be switched according to the position of the user. For example, when the user is on the infield side stand, an image of a landscape seen from the infield side stand may be set as the background image 201, and a camera icon 202 indicating a shooting viewpoint existing in the landscape may be displayed.
 この際、ユーザの位置に応じて、予め受信する映像が切り替えられてもよい。例えば、ユーザが内野側スタンドにいる場合は、端末装置102は、外野側スタンドから撮影された映像を予め受信してもよい。 At this time, the video to be received in advance may be switched according to the position of the user. For example, when the user is at the infield side stand, the terminal device 102 may receive in advance an image shot from the outfield side stand.
 また、初期画面又は予め受信される映像は、全ユーザ又は一部のユーザの視聴状況によって切り替えらえてもよい。例えば、視聴中のユーザの数が多い、又は、過去に視聴された数が多い映像が優先的に受信されてもよい。 Further, the initial screen or the video received in advance may be switched depending on the viewing status of all users or some users. For example, a video with a large number of viewing users or a large number of users viewed in the past may be preferentially received.
 再度、図4を用いて説明を行なう。初期画面において、ユーザによりいずれかのカメラアイコン202が選択されると(S103)、端末装置102は、選択された視点を示す視点指定信号152をサーバ103に送信する(S104)。 Again, the description will be made with reference to FIG. When one of the camera icons 202 is selected by the user on the initial screen (S103), the terminal apparatus 102 transmits a viewpoint designation signal 152 indicating the selected viewpoint to the server 103 (S104).
 視点指定信号152を受信したサーバ103は、視点指定信号152で指定される選択映像信号153の端末装置102への送信を開始する(S105)。選択映像信号153を受信した端末装置102は、選択映像信号153を復号し、得られた映像の表示を開始する(S106)。 The server 103 that has received the viewpoint designation signal 152 starts transmission of the selected video signal 153 designated by the viewpoint designation signal 152 to the terminal device 102 (S105). The terminal device 102 that has received the selected video signal 153 decodes the selected video signal 153 and starts displaying the obtained video (S106).
 また、視点指定信号152を受信したサーバ103は、選択映像信号153に関連する関連映像信号154を選択し(S107)、関連映像信号154の端末装置102への送信を開始する(S108)。なお、ここでは、選択映像信号153の送信の開始(S105)後に関連映像信号154の選択(S107)が行なわれているが、これらの処理順は任意でよく、一部が並列に行なわれてもよい。 Further, the server 103 that has received the viewpoint designation signal 152 selects the related video signal 154 related to the selected video signal 153 (S107), and starts transmitting the related video signal 154 to the terminal device 102 (S108). Here, the selection of the related video signal 154 (S107) is performed after the transmission of the selected video signal 153 is started (S105). However, these processing orders may be arbitrary, and some of them are performed in parallel. Also good.
 この関連映像の選択処理(S108)について説明する。サーバ103は、関連映像の選択処理として以下に示す複数の方法の少なくとも一つを用いる。また、以下に示す複数の方法の各々において、各映像の関連度が設定され、最終的な関連度が最も高い映像が関連映像として選択される。なお、優先度が高い側から順に複数の映像が関連映像として選択されてもよい。 The related video selection process (S108) will be described. The server 103 uses at least one of the following methods as related video selection processing. In each of the following methods, the relevance level of each video is set, and the video with the highest final relevance level is selected as the related video. A plurality of videos may be selected as related videos in order from the higher priority side.
 図6~図11は、この選択処理のフローチャートである。 6 to 11 are flowcharts of this selection process.
 図6に示す例では、サーバ103は、選択映像の撮影シーン(映像に写っている領域)の位置を算出し(S151)、選択映像の撮影シーンの位置に撮影シーンが近い映像の関連度を増加させる(S152)。具体的には、サーバ103は、カメラ101に送信された映像信号151に含まれる情報を用いて各映像の撮影シーンの位置を算出する。より具体的には、映像信号151には、当該映像を撮影した視点位置、カメラ101の方向及びズーム倍率などの情報が含まれる。サーバ103は、これらの情報を用いて、カメラ101が撮影している撮影シーンの位置を算出する。 In the example shown in FIG. 6, the server 103 calculates the position of the shooting scene of the selected video (area shown in the video) (S151), and determines the relevance of the video whose shooting scene is close to the position of the shooting scene of the selected video. Increase (S152). Specifically, the server 103 calculates the position of the shooting scene of each video using information included in the video signal 151 transmitted to the camera 101. More specifically, the video signal 151 includes information such as the viewpoint position where the video is shot, the direction of the camera 101, and the zoom magnification. The server 103 uses these pieces of information to calculate the position of the shooting scene that the camera 101 is shooting.
 なお、サーバ103は、各映像の撮影シーンの位置の算出を予め行なっておいてもよく、映像信号151の受信後であれば任意のタイミングで行なってよい。 Note that the server 103 may calculate the position of the shooting scene of each video in advance, or may be performed at an arbitrary timing as long as the video signal 151 is received.
 また、サーバ103は、撮影シーンの位置に加え、選択映像の撮影シーンの広さに撮影シーンの広さが近い映像の関連度を増加させてもよい。 In addition to the position of the shooting scene, the server 103 may increase the relevance of the video whose shooting scene is close to the shooting scene of the selected video.
 また、サーバ103は、選択映像の撮影シーンの位置に撮像シーンの位置が非常に近い(ほぼ同じ)映像に関しては関連度を増加させなくてもよい。 Also, the server 103 does not need to increase the degree of relevance for a video whose position of the imaging scene is very close (almost the same) to the position of the shooting scene of the selected video.
 図7に示す例では、サーバ103は、選択映像中の被写体(例えば選手)を特定し(S161)、選択映像中の被写体と同じ被写体が写っている映像の関連度を増加させる(S162)。例えば、カメラ101は、画像解析(顔認証等)により、映像中の被写体を特定し、当該被写体を示す情報を含む映像信号151をサーバ103に送信する。サーバ103は、当該情報を用いて、各映像中の被写体を判定する。なお、上記画像解析は、サーバ103で行なわれてもよい。また、被写体とは特定の人物に限らず、特定のチームなどであってもよい。 In the example shown in FIG. 7, the server 103 identifies a subject (for example, a player) in the selected video (S161), and increases the relevance of a video in which the same subject as the subject in the selected video is captured (S162). For example, the camera 101 identifies a subject in the video by image analysis (face authentication or the like), and transmits a video signal 151 including information indicating the subject to the server 103. The server 103 determines a subject in each video using the information. The image analysis may be performed by the server 103. The subject is not limited to a specific person, but may be a specific team.
 このように、サーバ103は、カメラ101又はサーバ103が、カメラ101が撮影した映像、及びカメラに付属するセンサで取得された情報の少なくとも一方を用いて生成された情報を用いて関連度を算出する。 As described above, the server 103 calculates the degree of association using the information generated by the camera 101 or the server 103 using at least one of the video captured by the camera 101 and the information acquired by the sensor attached to the camera. To do.
 図8に示す例では、サーバ103は、複数の映像の人気度を取得し(S171)、人気度の高い映像の関連度を増加させる(S172)。ここで人気度とは、例えば、現在又は過去のある一定時間内に映像が視聴された回数、又は映像を視聴したユーザの数を示す。なお、この人気度は、例えば、サーバ103において、複数のユーザの視聴状況に基づき順次算出される。 In the example shown in FIG. 8, the server 103 acquires the popularity of a plurality of videos (S171), and increases the relevance of videos with high popularity (S172). Here, the degree of popularity indicates, for example, the number of times that a video has been viewed within a certain time in the present or the past, or the number of users who have viewed the video. Note that the degree of popularity is sequentially calculated in the server 103 based on the viewing status of a plurality of users, for example.
 図9に示す例では、サーバ103は、ユーザの嗜好情報を取得し(S181)、ユーザの嗜好にあった映像の関連度を増加させる(S182)。ここで、嗜好情報とは、ユーザの視聴履歴、又は予め登録されている、ユーザの嗜好又は趣味を示す登録情報などである。例えば、サーバ103は、ユーザが過去に特定の選手又はチームが写っている映像を多く視聴している場合には、当該選手又はチームが写っている映像の関連度を増加させる。また、サーバ103は、登録情報にユーザが応援する選手又はチームが示されている場合には、当該選手又はチームが写っている映像の関連度を増加させる。 In the example shown in FIG. 9, the server 103 acquires user preference information (S181), and increases the degree of relevance of the video that matches the user preference (S182). Here, the preference information is a user's viewing history or registered information indicating a user's preference or hobby registered in advance. For example, when the user has watched many videos of a specific player or team in the past, the server 103 increases the degree of relevance of the videos of the player or team. Moreover, when the player or team which a user supports is shown by registration information, the server 103 increases the relevance degree of the image | video which the said player or team is reflected.
 図10に示す例では、サーバ103は、端末装置102が使用できる通信帯域を示す通信帯域情報を取得し(S191)、当該通信帯域に応じて関連度を変更する(S192)。具体的には、サーバ103は、端末装置102が使用できる通信帯域に適した、ビットレート、フレームレート又は解像度の映像の関連度を増加させる。例えば、端末装置102が使用できる通信帯域が十分に広い場合には、サーバ103は、ビットレート、フレームレート又は解像度が高い映像の関連度を増加させる。 In the example shown in FIG. 10, the server 103 acquires communication band information indicating a communication band that can be used by the terminal device 102 (S191), and changes the degree of association according to the communication band (S192). Specifically, the server 103 increases the degree of relevance of the video having the bit rate, the frame rate, or the resolution suitable for the communication band that can be used by the terminal device 102. For example, when the communication band that can be used by the terminal device 102 is sufficiently wide, the server 103 increases the relevance of the video having a high bit rate, frame rate, or resolution.
 なお、サーバ103は、カメラ101から送信された映像信号151の解像度又はフレームレートを変換することで複数のビットレートの映像信号を生成し、この複数の映像信号を格納しておいてもよい。 The server 103 may generate a plurality of bit rate video signals by converting the resolution or the frame rate of the video signal 151 transmitted from the camera 101, and store the plurality of video signals.
 また、視聴中に端末装置102が使用できる帯域幅が変動した場合、使用できる帯域幅にあわせて、選択映像又は関連映像を切り替えてもよい。 Also, when the bandwidth that can be used by the terminal device 102 changes during viewing, the selected video or the related video may be switched according to the available bandwidth.
 図11に示す例では、サーバ103は、端末装置102が使用できる通信帯域を示す通信帯域情報を取得し(S191)、当該通信帯域に応じて関連映像の数を決定する(S193)。具体的には、サーバ103は、通信帯域が広いほど、関連映像の数を増加させる。 In the example shown in FIG. 11, the server 103 acquires communication band information indicating a communication band that can be used by the terminal apparatus 102 (S191), and determines the number of related videos according to the communication band (S193). Specifically, the server 103 increases the number of related videos as the communication band is wider.
 このように、サーバ103は、複数の映像のうち、選択映像(第1映像)との関連度が高い映像を関連映像(第2映像)として選択する。具体的には、サーバ103は、撮影シーンの位置が選択映像の撮影シーンの位置に近いほど関連度が高いと判定する。また、サーバ103は、撮影シーンの広さが選択映像の撮影シーンの広さに近いほど関連度が高いと判定する。また、サーバ103は、選択映像に含まれる被写体と、同一の被写体が撮影されている映像の関連度を高く設定する。 Thus, the server 103 selects, as a related video (second video), a video having a high degree of association with the selected video (first video) from among a plurality of videos. Specifically, the server 103 determines that the degree of association is higher as the position of the shooting scene is closer to the position of the shooting scene of the selected video. Further, the server 103 determines that the degree of association is higher as the width of the shooting scene is closer to the width of the shooting scene of the selected video. Further, the server 103 sets a high degree of association between the subject included in the selected video and the video in which the same subject is captured.
 また、サーバ103は、複数の映像のフレームレート、解像度又はビットレートに基づき、関連映像を選択する。また、サーバ103は、複数の映像のうち、他のユーザに選択された回数が多い映像を関連映像として選択する。また、サーバ103は、ユーザの視聴履歴又は予め登録されている嗜好情報に基づき、関連映像を選択する。 Also, the server 103 selects related videos based on the frame rate, resolution, or bit rate of a plurality of videos. In addition, the server 103 selects, as a related video, a video that has been frequently selected by another user from a plurality of videos. Further, the server 103 selects a related video based on the viewing history of the user or the preference information registered in advance.
 図12は、映像が選択された後の表示画面の一例を示す図である。図12に示すように表示画面には、選択された映像である選択映像211と、俯瞰画像212と、上面画像213と、操作ボタン214~216とが表示される。 FIG. 12 is a diagram showing an example of the display screen after the video is selected. As shown in FIG. 12, a selection video 211 that is a selected video, an overhead image 212, a top image 213, and operation buttons 214 to 216 are displayed on the display screen.
 俯瞰画像212は、撮影シーンを俯瞰する画像であり、カメラアイコン202を含む。この俯瞰画像212は、初期画面で表示されていた画像と同様である。上面画像213は、撮影シーン全体を上から見た図であり、カメラアイコン202を含む。 The bird's-eye view image 212 is an image for bird's-eye view of the shooting scene, and includes a camera icon 202. This overhead image 212 is the same as the image displayed on the initial screen. The top image 213 is a view of the entire shooting scene viewed from above, and includes a camera icon 202.
 操作ボタン214~216はユーザが操作を行うためのボタンである。操作ボタン214が選択されることで、表示が初期画面に戻る。操作ボタン215又は216が操作されることで、他の視点映像に表示映像が切り替わる。なお、この時、選択映像との関連度が高い映像が優先的に選ばれる。 Operation buttons 214 to 216 are buttons for the user to operate. When the operation button 214 is selected, the display returns to the initial screen. When the operation button 215 or 216 is operated, the display video is switched to another viewpoint video. At this time, a video having a high degree of association with the selected video is preferentially selected.
 例えば、操作ボタン215が操作されることで、選択映像の撮影シーンの位置に、撮影シーンの位置が最も近い映像に表示映像が切り替えられる。 For example, when the operation button 215 is operated, the display image is switched to the image where the position of the shooting scene is closest to the position of the shooting scene of the selected image.
 また、操作ボタン216が操作させることで、最もお薦め度の高い映像に表示映像が切り替えられる。これにより、一度選択して視聴した映像がユーザの好みにあわなくても、ユーザは容易な操作で、その時点の試合を一番楽しむことができる映像に表示映像を切り替えて当該表示映像を視聴できる。 In addition, when the operation button 216 is operated, the display image is switched to the image with the highest recommendation. As a result, even if the video that has been selected and viewed does not meet the user's preference, the user can easily switch the display video to the video that can enjoy the game at that time and view the displayed video. it can.
 また、ユーザが俯瞰画像212又は上面画像213に含まれるカメラアイコン202を選択することで、選択されたカメラアイコン202に対応する映像に表示映像が切り替えられる。 In addition, when the user selects the camera icon 202 included in the overhead image 212 or the top image 213, the display image is switched to the image corresponding to the selected camera icon 202.
 なお、図12に示す各画像及び操作ボタンの配置は一例であり、この例に限定されない。また、複数の画像及び複数の操作ボタンの全てが表示される必要はなく一部のみが表示されてもよい。 In addition, arrangement | positioning of each image and operation button shown in FIG. 12 is an example, and is not limited to this example. Further, it is not necessary to display all of the plurality of images and the plurality of operation buttons, and only a part of them may be displayed.
 ここで、本実施の形態では、選択映像との関連度に応じて、カメラアイコン202の表示が変更される。例えば、選択映像と関連度の高い映像に対応するカメラアイコン202が強調表示される。なお、複数の映像のうち、選択映像と関連度の高い映像に対応するカメラアイコン202のみが表示されてもよい。また、関連度に応じてカメラアイコン202の表示方法が連続的又は段階的に変更されてもよい。また、カメラアイコン202の近傍に関連度を示す情報が表示されてもよい。 Here, in the present embodiment, the display of the camera icon 202 is changed according to the degree of association with the selected video. For example, a camera icon 202 corresponding to a video having a high degree of association with the selected video is highlighted. Note that only the camera icon 202 corresponding to a video having a high degree of association with the selected video among the plurality of videos may be displayed. Further, the display method of the camera icon 202 may be changed continuously or stepwise according to the degree of association. Information indicating the degree of association may be displayed near the camera icon 202.
 また、別の実施の形態として、ボールにセンサが組み込まれており、当該センサで検知された情報に基づき、ボールがどのように飛んで行ったかが判別されてもよい。そして、ボールの軌跡が、俯瞰画像212又は上面画像213に重畳表示されてもよい。 As another embodiment, a sensor may be incorporated in the ball, and it may be determined how the ball flew based on information detected by the sensor. Then, the trajectory of the ball may be superimposed on the overhead image 212 or the top image 213.
 さらには、ボールが飛んで行った先にカメラアイコン202がある場合には、端末装置102は、ボールの位置に近い視点位置の映像信号を予めサーバ103が受信しておいてもよい。 Furthermore, when the camera icon 202 is present at the point where the ball flew, the terminal device 102 may receive the video signal of the viewpoint position close to the position of the ball in advance by the server 103.
 つまり、当該システムは、試合の流れなどをなんらかの手段(ボールのセンサ等)から取得し、その情報を基にユーザが見たいと思うカメラアイコン202をあらかじめ推定し、端末装置102は、推定された映像を予め受信しておいてもよい。 That is, the system obtains the flow of the game from some means (such as a ball sensor) and preliminarily estimates the camera icon 202 that the user wants to see based on the information, and the terminal device 102 is estimated. The video may be received in advance.
 また、サーバ103は、このような試合の流れ等のその場の状況、又はユーザの位置等に基づき、複数の映像に優先度を設定してもよい。 Further, the server 103 may set priorities for a plurality of videos based on the current situation such as the flow of the game or the position of the user.
 再度、図4を用いて説明を行なう。図6に示す表示画面において、視点切替の操作が行なわれる(S109)。なお、ここでは、関連映像が選択されるとする。この場合、端末装置102は、関連映像信号154を予め受信しているので、当該関連映像信号154を復号し、関連映像を表示する(S110)。このように、端末装置102は、次に選択される可能性の高い関連映像を予め受信することで、映像の切替をシームレスに行うことができる。 Again, the description will be made with reference to FIG. On the display screen shown in FIG. 6, a viewpoint switching operation is performed (S109). Here, it is assumed that the related video is selected. In this case, since the terminal device 102 has received the related video signal 154 in advance, the terminal device 102 decodes the related video signal 154 and displays the related video (S110). In this way, the terminal device 102 can seamlessly switch the video by receiving in advance the related video that is likely to be selected next.
 また、端末装置102は、選択された視点を示す視点指定信号152をサーバ103に送信する(S111)。また、視点指定信号152を受信したサーバ103は、視点指定信号152で指定される選択映像信号153を端末装置102へ送信する。つまり、サーバ103は、先の関連映像信号154の送信を、選択映像信号153の送信として継続する(S112)。また、サーバ103は、新たな選択映像信号153に関連する関連映像信号154を選択し(S113)、関連映像信号154の端末装置102への送信を開始する(S114)。 Also, the terminal device 102 transmits a viewpoint designation signal 152 indicating the selected viewpoint to the server 103 (S111). In addition, the server 103 that has received the viewpoint designation signal 152 transmits a selection video signal 153 designated by the viewpoint designation signal 152 to the terminal device 102. That is, the server 103 continues transmission of the related video signal 154 as transmission of the selected video signal 153 (S112). Further, the server 103 selects the related video signal 154 related to the new selected video signal 153 (S113), and starts transmitting the related video signal 154 to the terminal device 102 (S114).
 なお、映像表示(S110)と、視点指定信号152の送信(S111)との順序は任意でよく、一部が並列に行なわれてもよい。 Note that the order of the video display (S110) and the transmission of the viewpoint designation signal 152 (S111) may be arbitrary, and some of them may be performed in parallel.
 次に、端末装置102の動作の流れを説明する。図13は、端末装置102の動作の流れ示すフローチャートである。なお、図13は、ある視点映像が表示されている状態における端末装置102の処理を示す。 Next, the operation flow of the terminal device 102 will be described. FIG. 13 is a flowchart illustrating the operation flow of the terminal apparatus 102. FIG. 13 shows processing of the terminal device 102 in a state where a certain viewpoint video is displayed.
 端末装置102は、ユーザの操作により視点切替が指示されたかを判定する(S121)。視点切替が指示された場合(S121でYes)、端末装置102は、視点指定信号152をサーバ103に送信する(S122)。 The terminal apparatus 102 determines whether or not the viewpoint switching is instructed by the user's operation (S121). When the viewpoint switching is instructed (Yes in S121), the terminal device 102 transmits the viewpoint designation signal 152 to the server 103 (S122).
 また、端末装置102は、視点切替先の選択映像が関連映像であるかを判定する(S123)。選択映像が関連映像でない場合(S123でNo)、端末装置102は、視点指定信号152に応じてサーバ103が送信する選択映像を受信するまで待機し(S124)、選択映像を受信した場合(S124でYes)、当該選択映像を表示する(S125)。 Further, the terminal device 102 determines whether the viewpoint switching destination selected video is a related video (S123). If the selected video is not related video (No in S123), the terminal apparatus 102 waits until receiving the selected video transmitted by the server 103 in response to the viewpoint designation signal 152 (S124), and if the selected video is received (S124). Yes), the selected video is displayed (S125).
 一方、選択映像が関連映像である場合(S123でYes)、端末装置102は、既に蓄積している関連映像を選択映像として表示する(S125)。 On the other hand, when the selected video is a related video (Yes in S123), the terminal device 102 displays the already stored related video as the selected video (S125).
 ここで、本システムがライブ映像を表示する場合、復号映像の切替は、ランダムアクセスフレームの復号が完了した時刻で行われてもよい。その際、ユーザの視点切替指示が発生した時刻から、切替わる時刻まで待ち時間が発生するが、端末装置102は、この待ち時間において、切り替え前の映像を再生し続けてもよいし、待ち画面を表示してもよい。 Here, when the system displays live video, switching of decoded video may be performed at the time when decoding of the random access frame is completed. At this time, a waiting time is generated from the time when the user's viewpoint switching instruction is generated until the switching time, and the terminal device 102 may continue to play the video before switching during this waiting time, May be displayed.
 また、本システムがライブ映像ではなくハイライト映像を表示する場合は、端末装置102は、切替前の映像の再生時刻から最も近い時刻のランダムアクセスポイントを探し、そこから映像を復号し表示してもよい。 When the system displays a highlight video instead of a live video, the terminal device 102 searches for a random access point at a time closest to the playback time of the video before switching, and decodes and displays the video from there. Also good.
 次に、端末装置102は、新たな選択映像に関連する関連映像を受信した場合(S126でYes)、受信した関連映像を蓄積部122に順次蓄積する(S127)。なお、表示された後の選択映像のデータ及び、受信後に一定期間使用されなかった関連映像のデータは、順次蓄積部122から削除される。 Next, when receiving the related video related to the new selected video (Yes in S126), the terminal device 102 sequentially stores the received related video in the storage unit 122 (S127). Note that the data of the selected video after being displayed and the data of the related video that has not been used for a certain period after reception are sequentially deleted from the storage unit 122.
 次に、端末装置102は、新たに受信した関連映像の情報を表示する(S128)。具体的には、端末装置102は、関連映像のカメラアイコン202を強調表示する。例えば、関連映像のカメラアイコン202は、他のカメラアイコン202よりも大きく表示される。また、関連映像のカメラアイコン202の輪郭線が、他のカメラアイコン202の輪郭線よりも太く表示される。または、関連映像のカメラアイコン202の色が、赤色などの目立つ色に変更される。なお、強調表示の方法はこれに限定されない。 Next, the terminal apparatus 102 displays the newly received related video information (S128). Specifically, the terminal device 102 highlights the camera icon 202 of the related video. For example, the related video camera icon 202 is displayed larger than the other camera icons 202. In addition, the contour line of the camera icon 202 of the related video is displayed thicker than the contour lines of the other camera icons 202. Alternatively, the color of the camera icon 202 of the related video is changed to a conspicuous color such as red. The highlighting method is not limited to this.
 なお、端末装置102は、図14又は図15に示す処理を行なってもよい。図14及び図15は、端末装置102の動作の変形例の流れ示すフローチャートである。 Note that the terminal apparatus 102 may perform the processing shown in FIG. 14 and 15 are flowcharts illustrating a flow of a modification example of the operation of the terminal device 102.
 図14に示す処理は、図13に示す処理に対してステップS129が追加されている。つまり、選択映像が関連映像でない場合(S123でNo)、端末装置102は、選択映像を受信するまでの期間において関連映像を表示する(S129)。また、端末装置102が複数の関連映像を蓄積している場合には、端末装置102は、蓄積している複数の関連映像のうち、新たな選択映像に関連度が最も高い関連映像を表示してもよい。 In the process shown in FIG. 14, step S129 is added to the process shown in FIG. That is, when the selected video is not related video (No in S123), the terminal apparatus 102 displays the related video in a period until the selected video is received (S129). In addition, when the terminal device 102 stores a plurality of related videos, the terminal device 102 displays a related video having the highest degree of association with the newly selected video among the plurality of stored related videos. May be.
 また、図15に示す処理は、図13に示す処理に対してステップS130が追加されている。つまり、選択映像が関連映像でない場合(S123でNo)、端末装置102は、選択映像を受信するまでの期間において三次元構成データを表示する(S130)。ここで、三次元構成データとは、複数の映像が撮影されている場所の三次元構成データであり、図5に示す例では、野球場の三次元構成データである。また、この三次元構成データは、複数の映像信号151を用いてサーバ103で生成され、予め端末装置102に送信されている。 Further, in the process shown in FIG. 15, step S130 is added to the process shown in FIG. That is, when the selected video is not a related video (No in S123), the terminal device 102 displays the three-dimensional configuration data in a period until the selected video is received (S130). Here, the three-dimensional configuration data is the three-dimensional configuration data of a place where a plurality of videos are taken, and in the example shown in FIG. 5, is the three-dimensional configuration data of a baseball field. The three-dimensional configuration data is generated by the server 103 using a plurality of video signals 151 and transmitted to the terminal device 102 in advance.
 なお、端末装置102は、三次元構成データを用いて、この期間において表示する映像を生成してもよい。例えば、端末装置102は、三次元構成データにおいて直前の表示映像の視点位置から選択映像の視点位置に視点位置が連続的に変更される映像を生成し、当該映像を上記期間に表示してもよい。また、このような視覚効果は、蓄積部122に映像データが蓄積されている場合にも用いられてもよい。さらに、この視覚効果が用いられるか否かは、直前の表示映像の視点位置と選択映像の視点位置との距離に応じて切り替えられてもよい。例えば、上記距離が短い場合には視覚効果が用いられず、上記距離が長い場合には視覚効果が用いられる。 Note that the terminal device 102 may generate an image to be displayed during this period using the three-dimensional configuration data. For example, the terminal apparatus 102 generates a video in which the viewpoint position is continuously changed from the viewpoint position of the immediately preceding display video in the three-dimensional configuration data to the viewpoint position of the selected video, and displays the video during the above period. Good. Such visual effects may also be used when video data is stored in the storage unit 122. Furthermore, whether or not this visual effect is used may be switched according to the distance between the viewpoint position of the immediately preceding display image and the viewpoint position of the selected image. For example, when the distance is short, the visual effect is not used, and when the distance is long, the visual effect is used.
 また、ここでは、選択映像を受信するまでの待ち時間において、関連映像又は三次元構成データが表示される例を述べたが、端末装置102は、通信エラー等により選択映像を受信できない場合など、何らかのエラーで選択映像を受信できない場合に、関連映像又は三次元構成データを表示してもよい。 In addition, here, an example in which related video or three-dimensional configuration data is displayed in the waiting time until receiving the selected video has been described. However, when the terminal device 102 cannot receive the selected video due to a communication error or the like, When the selected video cannot be received due to some error, the related video or the three-dimensional configuration data may be displayed.
 また、端末装置102は、選択映像を受信できない場合であって、当該映像を撮影しているカメラ101が、ユーザの近くに存在する場合は、近距離通信などの別の通信方法を用いてカメラ101から直接映像信号を受信してもよい。 If the terminal device 102 cannot receive the selected video and the camera 101 that is capturing the video is near the user, the terminal device 102 uses another communication method such as near field communication. The video signal may be received directly from the 101.
 以上のように、端末装置102は、複数の視点から撮影された複数の映像のいずれかをサーバ103から受信し、当該映像を表示する。まず、端末装置102は、複数の映像から選択映像(第1映像)を選択する(S121)。次に、端末装置102は、選択映像の送信をサーバ103に要求する(S122)。次に、端末装置102は、選択映像をサーバ103から受信し(S124)、当該選択映像を表示する(S125)。次に、端末装置102は、選択映像の受信及び表示中に、複数の映像の一つであり、選択映像と異なる映像であって、次に選択される可能性が高い関連映像の受信を開始する(S126)。 As described above, the terminal apparatus 102 receives any of a plurality of videos taken from a plurality of viewpoints from the server 103 and displays the videos. First, the terminal device 102 selects a selection video (first video) from a plurality of videos (S121). Next, the terminal device 102 requests the server 103 to transmit the selected video (S122). Next, the terminal device 102 receives the selected video from the server 103 (S124), and displays the selected video (S125). Next, during reception and display of the selected video, the terminal device 102 starts receiving a related video that is one of a plurality of videos and is different from the selected video and is likely to be selected next. (S126).
 また、端末装置102は、受信された関連映像を蓄積する(S127)。端末装置102は、選択映像の表示中に関連映像が選択された場合(S123でYes)に、蓄積されている関連映像を表示する(S125)。 Also, the terminal device 102 accumulates the received related video (S127). When the related video is selected during the display of the selected video (Yes in S123), the terminal device 102 displays the stored related video (S125).
 また、端末装置102は、選択映像の表示中に、選択映像及び関連映像と異なる第3映像が選択された場合(S123でNo)、第3映像をサーバ103から受信する(S124)。端末装置102は、第3映像を受信するまでの間、蓄積されている関連映像を表示する(S129)。 Further, when the third video different from the selected video and the related video is selected during the display of the selected video (No in S123), the terminal device 102 receives the third video from the server 103 (S124). The terminal apparatus 102 displays the related video that has been accumulated until the third video is received (S129).
 次に、サーバ103の動作の流れを説明する。図16は、サーバ103の動作の流れを示すフローチャートである。 Next, the operation flow of the server 103 will be described. FIG. 16 is a flowchart showing the operation flow of the server 103.
 まず、サーバ103は、視点指定信号152を端末装置102から受信したかを判定する(S141)。サーバ103は、視点指定信号152を受信した場合(S141でYes)、蓄積している複数の映像信号から、視点指定信号152で示される映像信号を選択映像信号153として選択し、当該選択映像信号153を端末装置102へ送信する(S142)。 First, the server 103 determines whether the viewpoint designation signal 152 is received from the terminal device 102 (S141). When the server 103 receives the viewpoint designation signal 152 (Yes in S141), the server 103 selects the video signal indicated by the viewpoint designation signal 152 as the selected video signal 153 from the accumulated video signals, and the selected video signal 153 is transmitted to the terminal device 102 (S142).
 また、サーバ103は、上述したように、蓄積している複数の映像信号151から、優先度に基づき選択映像に関連度が高い関連映像信号154を選択し(S143)、関連映像信号154を端末装置102へ送信する(S144)。 Further, as described above, the server 103 selects the related video signal 154 having a high degree of relevance to the selected video based on the priority from the plurality of stored video signals 151 (S143), and uses the related video signal 154 as a terminal. The data is transmitted to the device 102 (S144).
 以上のように、サーバ103は、複数のユーザにより異なる視点から撮影された複数の映像のいずれかを端末装置102に配信する。まず、サーバ103は、複数の映像の一つであり、端末装置102から要求された選択映像(第1映像)を端末装置102に配信する(S142)。次に、サーバ103は、複数の映像から、選択映像と異なる映像であって、端末装置102から次に要求される可能性が高い関連映像(第2映像)を選択する(S143)。言い換えると、関連映像は、端末装置102から要求されていない映像である。次に、サーバ103は、選択映像を端末装置102に配信している間に関連映像の端末装置102への送信を開始する(S144)。 As described above, the server 103 delivers any of a plurality of videos taken from different viewpoints by a plurality of users to the terminal device 102. First, the server 103 distributes the selected video (first video) requested by the terminal device 102 to the terminal device 102, which is one of a plurality of videos (S142). Next, the server 103 selects, from a plurality of videos, a related video (second video) that is different from the selected video and is highly likely to be requested next from the terminal device 102 (S143). In other words, the related video is a video that is not requested from the terminal device 102. Next, the server 103 starts transmitting the related video to the terminal device 102 while delivering the selected video to the terminal device 102 (S144).
 以上、実施の形態に係る映像配信方法、映像受信方法及び映像配信システムついて説明したが、本発明は、この実施の形態に限定されるものではない。 The video distribution method, video reception method, and video distribution system according to the embodiment have been described above, but the present invention is not limited to this embodiment.
 また、上記実施の形態に係る映像配信システムに含まれる各装置に含まれる各処理部は典型的には集積回路であるLSIとして実現される。これらは個別に1チップ化されてもよいし、一部又は全てを含むように1チップ化されてもよい。 Further, each processing unit included in each device included in the video distribution system according to the above embodiment is typically realized as an LSI which is an integrated circuit. These may be individually made into one chip, or may be made into one chip so as to include a part or all of them.
 また、集積回路化はLSIに限るものではなく、専用回路又は汎用プロセッサで実現してもよい。LSI製造後にプログラムすることが可能なFPGA(Field Programmable Gate Array)、又はLSI内部の回路セルの接続や設定を再構成可能なリコンフィギュラブル・プロセッサを利用してもよい。 Further, the integration of circuits is not limited to LSI, and may be realized by a dedicated circuit or a general-purpose processor. An FPGA (Field Programmable Gate Array) that can be programmed after manufacturing the LSI or a reconfigurable processor that can reconfigure the connection and setting of circuit cells inside the LSI may be used.
 上記各実施の形態において、各構成要素は、専用のハードウェアで構成されるか、各構成要素に適したソフトウェアプログラムを実行することによって実現されてもよい。各構成要素は、CPU又はプロセッサなどのプログラム実行部が、ハードディスク又は半導体メモリなどの記録媒体に記録されたソフトウェアプログラムを読み出して実行することによって実現されてもよい。 In each of the above embodiments, each component may be configured by dedicated hardware or may be realized by executing a software program suitable for each component. Each component may be realized by a program execution unit such as a CPU or a processor reading and executing a software program recorded on a recording medium such as a hard disk or a semiconductor memory.
 言い換えると、映像配信システムに含まれる各装置は、処理回路(processing circuitry)と、当該処理回路に電気的に接続された(当該処理回路からアクセス可能な)記憶装置(storage)とを備える。処理回路は、専用のハードウェア及びプログラム実行部の少なくとも一方を含む。また、記憶装置は、処理回路がプログラム実行部を含む場合には、当該プログラム実行部により実行されるソフトウェアプログラムを記憶する。処理回路は、記憶装置を用いて、上記実施の形態に係る映像配信方法又波映像受信方法を実行する。 In other words, each device included in the video distribution system includes a processing circuit and a storage device (storage) electrically connected to the processing circuit (accessible from the processing circuit). The processing circuit includes at least one of dedicated hardware and a program execution unit. Further, when the processing circuit includes a program execution unit, the storage device stores a software program executed by the program execution unit. The processing circuit uses the storage device to execute the video distribution method or the wave video reception method according to the above embodiment.
 さらに、本発明は上記ソフトウェアプログラムであってもよいし、上記プログラムが記録された非一時的なコンピュータ読み取り可能な記録媒体であってもよい。また、上記プログラムは、インターネット等の伝送媒体を介して流通させることができるのは言うまでもない。 Furthermore, the present invention may be the software program or a non-transitory computer-readable recording medium on which the program is recorded. Needless to say, the program can be distributed via a transmission medium such as the Internet.
 また、上記で用いた数字は、全て本発明を具体的に説明するために例示するものであり、本発明は例示された数字に制限されない。 Further, all the numbers used above are illustrated for specifically explaining the present invention, and the present invention is not limited to the illustrated numbers.
 また、上記の映像配信方法又は映像受信方法に含まれるステップが実行される順序は、本発明を具体的に説明するために例示するためのものであり、上記以外の順序であってもよい。また、上記ステップの一部が、他のステップと同時(並列)に実行されてもよい。 Further, the order in which the steps included in the above video distribution method or video reception method are executed is for illustrating the present invention specifically, and may be in an order other than the above. Also, some of the above steps may be executed simultaneously (in parallel) with other steps.
 以上、本発明の一つ又は複数の態様に係る映像配信方法、映像受信方法、映像配信システム、サーバ及び端末装置について、実施の形態に基づいて説明したが、本発明は、この実施の形態に限定されるものではない。本発明の趣旨を逸脱しない限り、当業者が思いつく各種変形を本実施の形態に施したものや、異なる実施の形態における構成要素を組み合わせて構築される形態も、本発明の一つ又は複数の態様の範囲内に含まれてもよい。 As described above, the video distribution method, the video reception method, the video distribution system, the server, and the terminal device according to one or more aspects of the present invention have been described based on the embodiment. It is not limited. Unless it deviates from the gist of the present invention, the embodiment in which various modifications conceived by those skilled in the art have been made in the present embodiment, and forms constructed by combining components in different embodiments are also applicable to one or more of the present invention. It may be included within the scope of the embodiments.
 (実施の形態2)
 上記各実施の形態で示した画像処理方法及び装置の構成の他の応用例とそれを用いたシステムを説明する。当該システムは、インテリジェント化と対象空間の広域化とが進む映像システムに適用でき、例えば、(1)店舗或いは工場のセキュリティカメラ、又は警察の車載カメラなどに実装される監視システム、(2)個人所有のカメラ或いは各車載カメラ、又は道路に備えられたカメラなどを用いた交通情報システム、(3)ドローンなど遠隔操作又は自動制御可能な装置を用いた環境調査又は配送システム、及び(4)エンターテイメント施設又はスタジアム等における設置カメラ、ドローン等の移動カメラ、又は個人所有のカメラなどを用いた映像などのコンテンツ送受信システムなどに適用できる。
(Embodiment 2)
Another application example of the configuration of the image processing method and apparatus shown in the above embodiments and a system using the same will be described. The system can be applied to a video system in which intelligentization and widening of the target space are progressing. For example, (1) a surveillance system implemented in a security camera in a store or a factory, an in-vehicle camera in a police, or Traffic information system using own camera or each on-vehicle camera or camera provided on road, (3) Environmental survey or delivery system using remote control or automatic control device such as drone, and (4) Entertainment The present invention can be applied to a content transmission / reception system such as a video using an installation camera in a facility or a stadium, a mobile camera such as a drone, or a personally owned camera.
 図17は、本実施の形態における映像情報処理システムex100の構成を示す図である。本実施の形態においては、死角の発生を防止する例、及び特定の領域を撮影禁止にする例について説明する。 FIG. 17 is a diagram showing a configuration of the video information processing system ex100 in the present embodiment. In this embodiment, an example of preventing the generation of blind spots and an example of prohibiting photographing in a specific area will be described.
 図17に示す映像情報処理システムex100は、映像情報処理装置ex101と、複数のカメラex102と、映像受信装置ex103とを含む。なお、映像受信装置ex103は、必ずしも映像情報処理システムex100に含まれる必要はない。 17 includes a video information processing apparatus ex101, a plurality of cameras ex102, and a video reception apparatus ex103. Note that the video receiving device ex103 is not necessarily included in the video information processing system ex100.
 映像情報処理装置ex101は、保存部ex111と、解析部ex112とを備える。N個のカメラex102のそれぞれは、映像を撮影する機能と撮影した映像データを映像情報処理装置ex101に送信する機能とを有する。また、カメラex102は、撮影中の映像を表示する機能を有する場合もある。なお、カメラex102は、撮影された映像信号をHEVC又はH.264のような符号化方式を用いてエンコードしたうえで映像情報処理装置ex101に送信してよいし、エンコードされていない映像データを映像情報処理装置ex101に送信してもよい。 The video information processing apparatus ex101 includes a storage unit ex111 and an analysis unit ex112. Each of the N cameras ex102 has a function of capturing video and a function of transmitting captured video data to the video information processing apparatus ex101. In addition, the camera ex102 may have a function of displaying an image being shot. The camera ex102 converts the captured video signal into HEVC or H.264. The encoded information may be encoded using an encoding method such as H.264 and transmitted to the video information processing apparatus ex101, or unencoded video data may be transmitted to the video information processing apparatus ex101.
 ここで、各カメラex102は、監視カメラ等の固定カメラ、無人飛行型ラジコンや車等に搭載された移動カメラ、又は、ユーザが所持するユーザカメラである。 Here, each camera ex102 is a fixed camera such as a surveillance camera, a moving camera mounted on an unmanned flight type radio control or a car, or a user camera possessed by the user.
 移動カメラは、映像情報処理装置ex101から送信された指示信号を受信し、受信された指示信号に応じて、移動カメラ自体の位置又は撮影方向を変更する。 The moving camera receives the instruction signal transmitted from the video information processing apparatus ex101, and changes the position or shooting direction of the moving camera itself according to the received instruction signal.
 また、撮影開示前に複数のカメラex102の時刻が、サーバ又は基準カメラの時刻情報などを用いてキャリブレーションされる。また、複数のカメラex102の空間位置が、撮影対象となる空間のオブジェクトの写り方又は基準カメラからの相対位置に基づいてキャリブレーションされる。 Also, the time of the plurality of cameras ex102 is calibrated using the time information of the server or the reference camera, etc. before the disclosure of photographing. Further, the spatial positions of the plurality of cameras ex102 are calibrated based on how the objects in the space to be imaged are captured or relative positions from the reference camera.
 情報処理装置ex101に含まれる保存部ex111は、N個のカメラex102から送信された映像データを保存する。 The storage unit ex111 included in the information processing apparatus ex101 stores video data transmitted from the N cameras ex102.
 解析部ex112は、保存部ex111に保存された映像データから死角を検出し、死角の発生を防止するための移動カメラへの指示を示す指示信号を移動カメラへ送信する。移動カメラは指示信号に従って移動を行い、撮影を継続する。 The analysis unit ex112 detects a blind spot from the video data stored in the storage unit ex111, and transmits an instruction signal indicating an instruction to the mobile camera for preventing the generation of the blind spot to the mobile camera. The moving camera moves in accordance with the instruction signal and continues shooting.
 解析部ex112は、例えば、SfM(Structure from Motion)を用いて死角検出を行う。SfMとは、異なる位置から撮影された複数の映像から被写体の三次元形状を復元する手法であり、被写体形状及びカメラ位置を同時に推定する形状復元技術として広く知られている。例えば、解析部ex112は、SfMを用いて、保存部ex111に保存された映像データから施設内又はスタジアム内の三次元形状を復元し、復元できない領域を死角として検出する。 The analysis unit ex112 performs blind spot detection using, for example, SfM (Structure from Motion). SfM is a technique for restoring the three-dimensional shape of a subject from a plurality of videos taken from different positions, and is widely known as a shape restoration technique for simultaneously estimating the subject shape and the camera position. For example, the analysis unit ex112 restores the three-dimensional shape in the facility or the stadium from the video data saved in the saving unit ex111 using SfM, and detects an area that cannot be restored as a blind spot.
 なお、解析部ex112は、カメラex102の位置及び撮影方向が固定であり、位置及び撮影方向の情報が既知の場合は、これらの既知の情報を用いてSfMを行ってもよい。また、移動カメラの位置及び撮影方向が、移動カメラが備えるGPS及び角度センサ等により取得できる場合は、移動カメラは、当該移動カメラの位置及び撮影方向の情報を解析部ex112に送信し、解析部ex112は、送信された位置及び撮影方向の情報を用いてSfMを行ってもよい。 Note that if the position and shooting direction of the camera ex102 are fixed and information on the position and shooting direction is known, the analysis unit ex112 may perform SfM using these known information. Further, when the position and shooting direction of the moving camera can be acquired by a GPS and an angle sensor provided in the moving camera, the moving camera transmits information on the position and shooting direction of the moving camera to the analysis unit ex112, and the analysis unit The ex 112 may perform SfM using the transmitted position and shooting direction information.
 なお、死角検出の方法は上述したSfMを用いた方法に限られるものではない。例えば、解析部ex112は、レーザレンジファインダなどのデプスセンサの情報を用いることで、撮影対象であるオブジェクトの空間距離を把握してもよい。また、解析部ex112は、カメラ位置、撮影方向及びズーム倍率等の情報を、空間内の予め設定したマーカ又は特定のオブジェクトが画像に含まれるか、含まれる場合にはそのサイズ等から検出してもよい。このように、解析部ex112は、各カメラの撮影領域を検出できる任意の方法を用いて、死角の検出を行う。また、解析部ex112は、複数の撮影対象について互いの位置関係等の情報を映像データ又は近接距離センサ等から取得し、取得した位置関係に基づいて死角が発生する可能性の高い領域を特定してもよい。 Note that the method of detecting the blind spot is not limited to the method using SfM described above. For example, the analysis unit ex112 may grasp the spatial distance of the object to be imaged by using information of a depth sensor such as a laser range finder. In addition, the analysis unit ex112 detects information such as a camera position, a shooting direction, and a zoom magnification from an image that includes a preset marker or a specific object in the space, or the size of the marker or the like. Also good. As described above, the analysis unit ex112 performs blind spot detection using an arbitrary method capable of detecting the imaging region of each camera. In addition, the analysis unit ex112 acquires information such as a mutual positional relationship for a plurality of imaging targets from video data or a proximity distance sensor, and identifies an area where a blind spot is likely to occur based on the acquired positional relationship. May be.
 ここで死角とは、撮影したい領域中で映像が存在しない部分だけでなく、他の部分と比較して画質の悪い部分、及び予め定められた画質を得られていない部分などを含む。この検出対象の部分は、当該システムの構成又は目的に応じて適宜設定されればよい。例えば、撮影される空間中の特定の被写体について、要求される画質が高く設定されてもよい。また、逆に撮影空間中の特定の領域について、要求される画質が低く設定されてもよいし、映像が撮影されていなくても死角と判定しないように設定されてもよい。 Here, the blind spot includes not only a portion where an image does not exist in a region to be photographed, but also a portion having a poor image quality compared to other portions and a portion where a predetermined image quality is not obtained. This detection target portion may be set as appropriate according to the configuration or purpose of the system. For example, the required image quality may be set high for a specific subject in the space where the image is taken. Conversely, for a specific area in the shooting space, the required image quality may be set low, or it may be set not to be determined as a blind spot even if no video is shot.
 なお、上述した画質とは、映像中の撮影対象となる被写体が占める面積(例えばピクセル数)、又は撮影対象となる被写体にピントが合っているかといった映像に関する様々な情報を含むものであり、それらの情報又はその組み合わせを基準に死角であるか否かが判定されればよい。 The above-mentioned image quality includes various information related to the video such as the area occupied by the subject to be photographed in the video (for example, the number of pixels) or whether the subject to be photographed is in focus. Whether or not it is a blind spot may be determined based on the information or the combination thereof.
 なお、上記の説明では、実際に死角となっている領域の検出について説明したが、死角の発生を防止するために検出する必要のある領域は実際に死角となっている領域に限定されない。例えば、複数の撮影対象が存在し、少なくともその一部が移動している場合には、ある撮影対象とカメラとの間に別の撮影対象が入ることによって新たな死角が生じる可能性がある。これに対し、解析部ex112は、例えば撮影された映像データ等から複数の撮影対象の動きを検出し、検出された複数の撮影対象の動きとカメラex102の位置情報に基づいて、新たに死角となる可能性のある領域を推定してもよい。この場合、映像情報処理装置ex101は、死角となる可能性のある領域を撮影するように移動カメラに指示信号を送信し、死角の発生を防止してもよい。 In the above description, detection of a region that is actually a blind spot has been described. However, a region that needs to be detected in order to prevent the generation of a blind spot is not limited to a region that is actually a blind spot. For example, when there are a plurality of shooting targets and at least some of them are moving, there is a possibility that a new blind spot may be generated when another shooting target enters between a shooting target and the camera. On the other hand, the analysis unit ex112 detects movements of a plurality of shooting targets from, for example, shot video data and the like, and based on the detected movements of the plurality of shooting targets and position information of the camera ex102, a new blind spot and It is also possible to estimate a possible region. In this case, the video information processing apparatus ex101 may transmit an instruction signal to the moving camera so as to capture an area that may become a blind spot, and prevent the generation of a blind spot.
 なお、移動カメラが複数ある場合、映像情報処理装置ex101は、死角、又は死角となる可能性がある領域を撮影させるために指示信号を送信する移動カメラを選択する必要がある。また、移動カメラ及び死角、又は死角となる可能性がある領域がそれぞれ複数存在する場合、映像情報処理装置ex101は、複数の移動カメラのそれぞれについて、どの死角、又は死角となる可能性がある領域を撮影させるかを決定する必要がある。例えば、映像情報処理装置ex101は、死角、又は死角となる可能性のある領域と各移動カメラが撮影中の領域の位置とに基づいて、死角、又は死角となる領域に最も近い移動カメラを選択する。また、映像情報処理装置ex101は、各移動カメラについて、当該移動カメラが現在撮影中の映像データが得られない場合に新たに死角が発生するか否かを判定し、現在撮影中の映像データが得られなくても死角が発生しないと判断された移動カメラを選択してもよい。 When there are a plurality of moving cameras, the video information processing apparatus ex101 needs to select a moving camera that transmits an instruction signal in order to capture a blind spot or an area that may become a blind spot. In addition, when there are a plurality of moving cameras and blind spots, or areas that may become blind spots, the video information processing apparatus ex101 determines which dead spots or areas that may become blind spots for each of the plurality of moving cameras. It is necessary to decide whether to shoot. For example, the video information processing apparatus ex101 selects a moving camera that is closest to the blind spot or the area that is the blind spot based on the blind spot or the area that may be the blind spot and the position of the area that each moving camera is capturing. To do. Further, the video information processing apparatus ex101 determines, for each moving camera, whether or not a blind spot is newly generated when the moving camera cannot obtain the video data currently being shot. If it is not obtained, a moving camera determined not to generate a blind spot may be selected.
 以上の構成により、映像情報処理装置ex101は、死角を検出し、死角を防止するように移動カメラに対して指示信号を送信することにより、死角の発生を防止できる。 With the above configuration, the video information processing apparatus ex101 can prevent the generation of a blind spot by detecting a blind spot and transmitting an instruction signal to the moving camera so as to prevent the blind spot.
 (変形例1)
 なお、上記説明では、移動カメラに移動を指示する指示信号が送信される例を述べたが、指示信号は、ユーザカメラのユーザに移動を指示するための信号であってもよい。例えば、ユーザカメラは、指示信号に基づき、ユーザにカメラの方向を変更するように指示する指示画像を表示する。なお、ユーザカメラは、ユーザの移動の指示として、地図上に移動経路を示した指示画像を表示してもよい。また、ユーザカメラは、取得される画像の質を向上させるために撮影方向、角度、画角、画質、及び撮影領域の移動など詳細な撮影の指示を表示してもよく、さらに映像情報処理装置ex101側で制御可能であれば、映像情報処理装置ex101は、そのような撮影に関するカメラex102の特徴量を自動で制御してもよい。
(Modification 1)
In the above description, an example in which an instruction signal for instructing movement is transmitted to the moving camera has been described. However, the instruction signal may be a signal for instructing the user of the user camera to move. For example, the user camera displays an instruction image that instructs the user to change the direction of the camera based on the instruction signal. Note that the user camera may display an instruction image indicating a movement route on a map as an instruction to move the user. The user camera may display detailed shooting instructions such as shooting direction, angle, angle of view, image quality, and movement of the shooting area in order to improve the quality of the acquired image. If control is possible on the ex101 side, the video information processing apparatus ex101 may automatically control the feature amount of the camera ex102 regarding such shooting.
 ここで、ユーザカメラは、例えば、スタジアム内の観客又は施設内の警備員が持つスマートフォン、タブレット型端末、ウェアラブル端末、又はHMD(Head Mounted Display)である。 Here, the user camera is, for example, a smartphone, a tablet terminal, a wearable terminal, or an HMD (Head Mounted Display) held by a spectator in the stadium or a guard in the facility.
 また、指示画像を表示する表示端末は、映像データを撮影するユーザカメラと同一である必要はない。例えば、ユーザカメラに予め対応付けられた表示端末に対して、ユーザカメラが指示信号又は指示画像を送信し、当該表示端末が指示画像を表示してもよい。また、ユーザカメラに対応する表示端末の情報が、予め映像情報処理装置ex101に登録されてもよい。この場合は、映像情報処理装置ex101は、ユーザカメラに対応する表示端末に対して指示信号を直接送信することで、表示端末に指示画像を表示させてもよい。 Also, the display terminal that displays the instruction image need not be the same as the user camera that captures the video data. For example, the user camera may transmit an instruction signal or an instruction image to a display terminal associated with the user camera in advance, and the display terminal may display the instruction image. Further, information on the display terminal corresponding to the user camera may be registered in advance in the video information processing apparatus ex101. In this case, the video information processing apparatus ex101 may display the instruction image on the display terminal by directly transmitting the instruction signal to the display terminal corresponding to the user camera.
 (変形例2)
 解析部ex112は、例えばSfMを用いて、保存部ex111に保存された映像データから施設内又はスタジアム内の三次元形状を復元することで自由視点映像(三次元再構成データ)を生成してもよい。この自由視点映像は、保存部ex111に保存される。映像情報処理装置ex101は、映像受信装置ex103から送信される視野情報(及び/又は、視点情報)に応じた映像データを保存部ex111から読み出して、映像受信装置ex103に送信する。なお、映像受信装置ex103は、複数のカメラ111の一つであってもよい。
(Modification 2)
The analysis unit ex112 may generate a free viewpoint video (three-dimensional reconstruction data) by restoring the three-dimensional shape in the facility or the stadium from the video data stored in the storage unit ex111 using, for example, SfM. Good. This free viewpoint video is stored in the storage unit ex111. The video information processing apparatus ex101 reads video data corresponding to the visual field information (and / or viewpoint information) transmitted from the video reception apparatus ex103 from the storage unit ex111 and transmits the video data to the video reception apparatus ex103. Note that the video reception device ex103 may be one of the plurality of cameras 111.
 (変形例3)
 映像情報処理装置ex101は、撮影禁止領域を検出してもよい。この場合、解析部ex112は撮影画像を解析し、移動カメラが撮影禁止領域を撮影している場合には移動カメラに対して撮影禁止信号を送信する。移動カメラは撮影禁止信号を受信している間は撮影を停止する。
(Modification 3)
The video information processing apparatus ex101 may detect a shooting prohibited area. In this case, the analysis unit ex112 analyzes the photographed image, and transmits a photographing prohibition signal to the moving camera when the mobile camera is photographing the photographing prohibition region. The mobile camera stops shooting while receiving the shooting prohibition signal.
 解析部ex112は、例えば、SfMを用いて復元された三次元の仮想空間と、撮影映像とのマッチングを取ることで、空間内で予め設定されている移動カメラが撮影禁止領域を撮影中かを判定する。または、解析部ex112は、空間内に配置されたマーカ又は特徴的なオブジェクトをトリガーとして移動カメラが撮影禁止領域を撮影中かを判定する。撮影禁止領域とは、例えば施設内又はスタジアム内のトイレなどである。 The analysis unit ex112, for example, matches the three-dimensional virtual space restored using SfM with the captured image, thereby determining whether the mobile camera set in advance in the space is capturing the prohibited image area. judge. Alternatively, the analysis unit ex112 determines whether the moving camera is shooting the shooting prohibited area using a marker or a characteristic object arranged in the space as a trigger. The photographing prohibited area is, for example, a toilet in a facility or a stadium.
 また、ユーザカメラが撮影禁止領域を撮影している場合には、ユーザカメラは、無線又は有線で接続されるディスプレイ等にメッセージを表示したり、スピーカ又はイヤホンから音又は音声を出力したりすることで、現在の場所が撮影禁止場所であることをユーザに知らせてもよい。 In addition, when the user camera is shooting a shooting prohibited area, the user camera displays a message on a display or the like connected wirelessly or by wire, or outputs a sound or sound from a speaker or an earphone. Thus, the user may be informed that the current location is a shooting prohibited location.
 例えば、上記メッセージとして、現在カメラを向けている方向が撮影禁止である旨が表示される。または、表示される地図上に撮影禁止領域と現在の撮影領域とが示される。また、撮影の再開は、例えば、撮影禁止信号が出力されなくなれば自動的に行われる。または、撮影禁止信号が出力されておらず、かつ、ユーザが撮影再開を行う操作をした場合に、撮影が再開されてもよい。また、撮影の停止と再開とが短期間で複数回起こった場合には、再度キャリブレーションが行われてもよい。または、ユーザに現在位置を確認したり移動を促したりするための通知が行われてもよい。 For example, as the above message, it is displayed that the direction in which the camera is currently facing is prohibited from being shot. Alternatively, the shooting prohibited area and the current shooting area are shown on the displayed map. In addition, the resumption of photographing is automatically performed when, for example, the photographing prohibition signal is not output. Alternatively, photographing may be resumed when the photographing prohibition signal is not output and the user performs an operation to resume photographing. In addition, when the stop and restart of shooting occur a plurality of times in a short period, calibration may be performed again. Alternatively, notification for confirming the current position or prompting the user to move may be performed.
 また、警察など特別な業務の場合には、記録のためこのような機能をオフにするパスコード又は指紋認証などが用いられてもよい。さらに、そのような場合であっても撮影禁止領域の映像が外部に表示されたり保存される場合には自動でモザイクなど画像処理が行われてもよい。 Also, in the case of special operations such as police, a passcode or fingerprint authentication that turns off such a function for recording may be used. Further, even in such a case, image processing such as mosaicing may be automatically performed when a video in the photographing prohibited area is displayed or stored outside.
 以上の構成により、映像情報処理装置ex101は、撮影禁止の判定を行い、撮影を停止するようにユーザに通知することで、ある領域を撮影禁止に設定できる。 With the above configuration, the video information processing apparatus ex101 can determine that shooting is prohibited and notify the user to stop shooting, thereby setting a certain region to shooting prohibited.
 (変形例4)
 映像から三次元の仮想空間を構築するためには、複数視点の映像を集める必要があるため、映像情報処理システムex100は、撮影映像を転送したユーザに対してインセンティブを設定する。例えば、映像情報処理装置ex101は、映像を転送したユーザに対し、無料又は割引料金で映像配信を行ったり、オンライン又はオフラインの店又はゲーム内で使用できるような金銭的な価値、又はゲームなどのバーチャル空間での社会的地位など非金銭的な価値のあるポイントを付与する。また、映像情報処理装置ex101は、リクエストが多いなど価値のある視野(及び/又は、視点)の撮影映像を転送したユーザに対しては特に高いポイントを付与する。
(Modification 4)
In order to construct a three-dimensional virtual space from videos, it is necessary to collect videos from a plurality of viewpoints. Therefore, the video information processing system ex100 sets an incentive for the user who transferred the shot video. For example, the video information processing apparatus ex101 delivers a video value to a user who has transferred video at a free or discounted rate, a monetary value that can be used in an online or offline store or game, a game, etc. Points that have non-monetary value such as social status in virtual space. In addition, the video information processing apparatus ex101 gives a particularly high point to a user who has transferred a captured video of a valuable field of view (and / or viewpoint) such as many requests.
 (変形例5)
 映像情報処理装置ex101は、解析部ex112の解析結果に基づき、ユーザカメラに対して付加情報を送信してもよい。この場合、ユーザカメラは撮影映像に付加情報を重畳して、画面に表示する。付加情報とは、例えば、スタジアムでの試合が撮影されている場合には、選手名又は身長などの選手の情報であり、映像内の各選手に対応付けて当該選手の名前又は顔写真などが表示される。なお、映像情報処理装置ex101は、映像データの一部又は全部の領域に基づきインターネット経由の検索により、付加情報を抽出してもよい。また、カメラex102は、Bluetooth(登録商標)をはじめとする近距離無線通信又は、スタジアム等の照明から可視光通信によりそのような付加情報を受け取り、受け取った付加情報を、映像データにマッピングしてもよい。また、カメラex102は、このマッピングを、カメラex102に有線又は無線により接続される記憶部に保持されるテーブルであって、可視光通信技術により得られる情報と付加情報との対応関係を示すテーブルなどの一定規則に基づいて行なってもよいし、インターネット検索により最も確からしい組み合わせの結果を用いて行なってもよい。
(Modification 5)
The video information processing apparatus ex101 may transmit additional information to the user camera based on the analysis result of the analysis unit ex112. In this case, the user camera superimposes additional information on the captured video and displays it on the screen. The additional information is, for example, information on players such as a player name or height when a game in a stadium is being shot, and the name or face photo of the player is associated with each player in the video. Is displayed. Note that the video information processing apparatus ex101 may extract additional information by searching via the Internet based on part or all of the video data area. The camera ex102 receives such additional information by short-range wireless communication including Bluetooth (registered trademark) or visible light communication from lighting such as a stadium, and maps the received additional information to video data. Also good. The camera ex102 is a table in which this mapping is stored in a storage unit connected to the camera ex102 by wire or wirelessly, and shows a correspondence relationship between information obtained by visible light communication technology and additional information, etc. It may be performed based on a certain rule of the above, or may be performed using the most probable combination result by Internet search.
 また、監視システムにおいては、施設内の警備員が持つユーザカメラに対して、例えば注意人物の情報が重畳されることで、監視システムの高精度化を図ることができる。 Also, in the monitoring system, for example, information of a caution person is superimposed on a user camera held by a guard in the facility, so that the monitoring system can be highly accurate.
 (変形例5)
 解析部ex112は,自由視点映像とユーザカメラの撮影映像とのマッチングを取ることで、ユーザカメラが施設内又はスタジアム内のどの領域を撮影中かを判定してもよい。なお、撮影領域の判定方法はこれに限られず、上述した各実施の形態で説明した様々な撮影領域の判定方法又はその他の撮影領域の判定方法を用いられてもよい。
(Modification 5)
The analysis unit ex112 may determine which area in the facility or stadium the user camera is capturing by matching the free viewpoint image and the captured image of the user camera. Note that the imaging region determination method is not limited to this, and various imaging region determination methods or other imaging region determination methods described in the above-described embodiments may be used.
 映像情報処理装置ex101は、解析部ex112の解析結果に基づき、ユーザカメラに対して過去映像を送信する。ユーザカメラは撮影映像に過去映像を重畳して、又は撮影映像を過去映像に置換して、画面に表示する。 The video information processing apparatus ex101 transmits the past video to the user camera based on the analysis result of the analysis unit ex112. The user camera displays the past video on the screen by superimposing the past video on the shot video or replacing the shot video with the past video.
 例えば、ハーフタイム中に、過去映像として前半のハイライトシーンが表示される。これにより、ユーザはハーフタイム中に、前半のハイライトシーンを自分が見ている方向の映像として楽しむことができる。なお過去映像は、前半のハイライトシーンに限らず、そのスタジアムで行われた過去の試合のハイライトシーンなどでもよい。また、映像情報処理装置ex101が過去映像を配信するタイミングはハーフタイム中に限らず、例えば試合終了後でも、試合中でもよい。特に試合中の場合には、解析部ex112の解析結果に基づき、映像情報処理装置ex101はユーザが見逃した重要と考えられるシーンを配信してもよい。また、映像情報処理装置ex101はユーザからリクエストがあった場合のみ過去映像を配信してもよく、又は過去映像の配信前に配信許可のメッセージを配信してもよい。 For example, during the halftime, the highlight scene of the first half is displayed as a past video. Accordingly, the user can enjoy the highlight scene of the first half as a video in the direction in which he / she is viewing during the halftime. The past video is not limited to the highlight scene in the first half, but may be a highlight scene of a past game held at the stadium. The timing at which the video information processing apparatus ex101 delivers the past video is not limited to half time, and may be, for example, after the match or during the match. Particularly during a game, based on the analysis result of the analysis unit ex112, the video information processing apparatus ex101 may deliver a scene that is considered important and missed by the user. In addition, the video information processing apparatus ex101 may distribute the past video only when requested by the user, or may distribute a distribution permission message before the past video is distributed.
 (変形例6)
 映像情報処理装置ex101は、解析部ex112の解析結果に基づき、ユーザカメラに対して広告情報を送信してもよい。ユーザカメラは撮影映像に広告情報を重畳して、画面に表示する。
(Modification 6)
The video information processing apparatus ex101 may transmit advertisement information to the user camera based on the analysis result of the analysis unit ex112. The user camera superimposes advertisement information on the captured video and displays it on the screen.
 広告情報は例えば変形例5で示した、ハーフタイム中又は試合終了後の過去映像配信直前に配信されてもよい。これにより、配信業者は広告主からの広告料を得ることができ、ユーザに安価又は無料で映像配信サービスを提供できる。また、映像情報処理装置ex101は、広告情報の配信直前に広告配信許可のメッセージを配信してもよいし、ユーザが広告を視聴した場合のみ無料でサービスを提供してもよいし、広告を視聴しない場合より安価にサービスを提供してもよい。 The advertisement information may be distributed immediately before the past video distribution during the half time or after the match, as shown in, for example, Modification 5. Accordingly, the distributor can obtain an advertisement fee from the advertiser, and can provide a video distribution service to the user at a low cost or free of charge. In addition, the video information processing apparatus ex101 may distribute an advertisement distribution permission message immediately before distribution of the advertisement information, may provide a service for free only when the user views the advertisement, or views the advertisement. Service may be provided at a lower cost than when not.
 また、広告に従ってユーザが「今すぐ注文する」などをクリックすると、当該システム又は何らかの位置情報に基づいてユーザの位置を把握しているスタッフ又は会場の自動の配送システムが注文された飲み物を席まで届けてくれる。決裁はスタッフへの手渡しでもよいし、予めモバイル端末のアプリ等に設定されているクレジットカード情報に基づいて行われてもよい。また、広告にはeコマースサイトへのリンクが含まれ、通常の自宅配送等のオンラインショッピングが可能な状態になっていてもよい。 In addition, when the user clicks “Order now” according to the advertisement, the system or the staff who knows the location of the user based on some location information or the automatic delivery system of the venue will bring the ordered drink to the seat Will deliver. The decision may be handed to the staff or may be made based on credit card information set in advance in the mobile terminal application or the like. Further, the advertisement may include a link to an e-commerce site, and online shopping such as normal home delivery may be possible.
 (変形例7)
 映像受信装置ex103は、カメラex102(ユーザカメラ)の一つであってもよい。
(Modification 7)
The video receiving device ex103 may be one of the cameras ex102 (user camera).
 この場合、解析部ex112は、自由視点映像とユーザカメラの撮影映像とのマッチングを取ることで、ユーザカメラが施設内又はスタジアム内のどの領域を撮影中かを判定する。なお、撮影領域の判定方法はこれに限らない。 In this case, the analysis unit ex112 determines which area in the facility or stadium the user camera is shooting by matching the free viewpoint video and the video shot by the user camera. Note that the method for determining the imaging region is not limited to this.
 例えば、ユーザが、画面に表示されている矢印の方向にスワイプ操作をすると、ユーザカメラはその方向へ視点を移動させることを示す視点情報を生成する。映像情報処理装置ex101は、解析部ex112が判定したユーザカメラの撮影領域から視点情報の分だけ移動させた領域を撮影した映像データを保存部ex111から読み出し、当該映像データのユーザカメラへの送信を開始する。そしてユーザカメラは撮影映像ではなく、映像情報処理装置ex101から配信された映像を表示する。 For example, when the user performs a swipe operation in the direction of the arrow displayed on the screen, the user camera generates viewpoint information indicating that the viewpoint is moved in that direction. The video information processing apparatus ex101 reads the video data obtained by shooting the area moved by the viewpoint information from the shooting area of the user camera determined by the analysis unit ex112 from the storage unit ex111, and transmits the video data to the user camera. Start. The user camera displays the video distributed from the video information processing apparatus ex101 instead of the captured video.
 以上により、施設内又はスタジアム内のユーザは、画面スワイプのような簡易な動作で、好きな視点からの映像を視聴できる。例えば野球場の3塁側で観戦している観客が、1塁側の視点からの映像を視聴できる。また、監視システムにおいては、施設内の警備員が画面スワイプのような簡易な動作で、自身が確認したい視点又はセンターからの割り込みとして注視すべき映像などを、視点を適用的に変えながら視聴することができるので、監視システムの高精度化を図ることができる。 As described above, the users in the facility or the stadium can view the video from a favorite viewpoint with a simple operation like a screen swipe. For example, a spectator watching on the third base side of a baseball field can view a video from the first base side viewpoint. In addition, in the surveillance system, the security guards in the facility can watch the video that should be watched as an interrupt from the viewpoint or the center that they want to confirm by a simple operation like a screen swipe while changing the viewpoint appropriately. Therefore, it is possible to increase the accuracy of the monitoring system.
 また、施設内又はスタジアム内のユーザへの映像の配信は、例えばユーザカメラと撮影対象との間に障害物が存在し、見えない領域がある場合等にも有効である。この場合、ユーザカメラは、ユーザカメラの撮影領域のうち障害物が含まれる一部の領域の映像を、撮影映像から、映像情報処理装置ex101からの配信映像に切り替えて表示してもよいし、画面全体を撮影映像から配信映像に切り替えて表示してもよい。また、ユーザカメラは、撮影映像と配信映像とを合成して障害物を透過して視聴対象が見えているような映像を表示してもよい。この構成によると、障害物の影響でユーザの位置から撮影対象が見えない場合にも、映像情報処理装置ex101から配信された映像を視聴することができるので、障害物の影響を軽減することができる。 Also, distribution of video to users in the facility or stadium is also effective when there are obstacles between the user camera and the shooting target and there is an invisible area, for example. In this case, the user camera may switch and display the video of a part of the shooting area of the user camera including the obstacle from the shot video to the distribution video from the video information processing apparatus ex101. The entire screen may be switched from the captured video to the distributed video and displayed. In addition, the user camera may display an image in which the object to be viewed is seen through the obstacle by combining the captured image and the distribution image. According to this configuration, it is possible to view the video distributed from the video information processing apparatus ex101 even when the shooting target cannot be seen from the position of the user due to the influence of the obstacle, so that the influence of the obstacle can be reduced. it can.
 また、障害物により見えない領域の映像として配信映像を表示する場合は、上述した画面スワイプのようなユーザによる入力処理に応じた表示の切り替え制御とは異なる表示の切り替え制御が行われてもよい。例えば、ユーザカメラの移動及び撮影方向の情報、並びに予め得られている障害物の位置情報に基づいて撮影領域に障害物が含まれると判定される場合に、撮影映像から配信映像への表示の切り替えが自動的に行われもよい。また、撮影映像データの解析により撮影対象ではない障害物が映っていると判定された場合に、撮影映像から配信映像への表示の切り替えが自動的に行われてもよい。また、撮影映像に含まれる障害物の面積(例えばピクセル数)が所定の閾値を超えた場合、又は撮影対象の面積に対する障害物の面積の比が所定の割合を超えた場合に、撮影映像から配信映像への表示の切り替えが自動的に行われてもよい。 Further, when the distribution video is displayed as a video of an area that cannot be seen due to an obstacle, display switching control different from the display switching control according to the input process by the user, such as the screen swipe described above, may be performed. . For example, when it is determined that an obstacle is included in the shooting area based on the information on the movement and shooting direction of the user camera and the position information of the obstacle obtained in advance, the display from the shot video to the distribution video is performed. Switching may be performed automatically. In addition, when it is determined that an obstacle that is not a shooting target is reflected by analysis of the shot video data, display switching from the shot video to the distribution video may be automatically performed. Further, when the area of the obstacle (for example, the number of pixels) included in the photographed image exceeds a predetermined threshold, or when the ratio of the area of the obstacle to the area to be photographed exceeds a predetermined ratio, The display switching to the distribution video may be automatically performed.
 なお、ユーザの入力処理に応じて撮影映像から配信映像への表示の切り替え及び配信映像から撮影映像への表示の切り替えが行われてもよい。 It should be noted that the display switching from the captured video to the distribution video and the display switching from the distribution video to the captured video may be performed in accordance with the user input processing.
 (変形例8)
 各カメラex102で撮影された映像データの重要度に基づき映像データを映像情報処理装置ex101に転送する速度が指示されてもよい。
(Modification 8)
The speed at which the video data is transferred to the video information processing apparatus ex101 may be instructed based on the importance of the video data captured by each camera ex102.
 この場合、解析部ex112は保存部ex111に保存された映像データ、又は当該映像データを撮影したカメラex102の重要度を判定する。ここでの重要度の判定は、例えば映像中に含まれる人の数或いは移動物体の数、映像データの画質などの情報、又はその組み合わせに基づいて行われる。 In this case, the analysis unit ex112 determines the importance of the video data stored in the storage unit ex111 or the camera ex102 that captured the video data. The determination of the importance here is performed based on, for example, information such as the number of people or moving objects included in the video, the image quality of the video data, or a combination thereof.
 また、映像データの重要度の判定は、映像データが撮影されたカメラex102の位置又は映像データが撮影している領域に基づいてもよい。例えば、対象のカメラex102の近くに撮影中の他のカメラex102が複数存在する場合に、対象のカメラex102で撮影された映像データの重要度を低くする。また、対象のカメラex102の位置が他のカメラex102から離れていても同じ領域を撮影している他のカメラex102が複数存在する場合に、対象のカメラex102で撮影された映像データの重要度を低くする。また、映像データの重要度の判定は、映像配信サービスにおけるリクエストの多さに基づいて行われてもよい。なお、重要度の判定方法は、上述したものやその組み合わせに限られず、監視システム又は映像配信システムの構成又は目的に応じた方法であればよい。 Also, the determination of the importance of the video data may be based on the position of the camera ex102 where the video data is shot or the area where the video data is shot. For example, when there are a plurality of other cameras ex102 being shot near the target camera ex102, the importance of the video data shot by the target camera ex102 is reduced. In addition, even when the position of the target camera ex102 is far from the other camera ex102, when there are a plurality of other cameras ex102 shooting the same area, the importance of the video data shot by the target camera ex102 is set. make low. The determination of the importance of the video data may be performed based on the number of requests in the video distribution service. The importance determination method is not limited to the method described above or a combination thereof, and may be any method according to the configuration or purpose of the monitoring system or the video distribution system.
 また、重要度の判定は撮影された映像データに基づくものでなくてもよい。例えば、映像情報処理装置ex101以外の端末へ映像データを送信するカメラex102の重要度が高く設定されてもよい。逆に、映像情報処理装置ex101以外の端末へ映像データを送信するカメラex102の重要度が低く設定されてもよい。これにより、例えば、映像データの伝送を必要とする複数のサービスが通信帯域を共有している場合に、各サービスの目的又は特性に応じた通信帯域の制御の自由度が高くなる。これにより、必要な映像データが得られないことによる各サービスの品質の劣化を防止できる。 Also, the determination of the importance may not be based on the captured video data. For example, the importance of the camera ex102 that transmits video data to a terminal other than the video information processing apparatus ex101 may be set high. Conversely, the importance of the camera ex102 that transmits video data to a terminal other than the video information processing apparatus ex101 may be set low. Thereby, for example, when a plurality of services that require transmission of video data share a communication band, the degree of freedom in controlling the communication band according to the purpose or characteristics of each service is increased. Thereby, it is possible to prevent the quality of each service from deteriorating due to the lack of necessary video data.
 また、解析部ex112は、自由視点映像とカメラex102の撮影映像とを用いて、映像データの重要度を判定してもよい。 Also, the analysis unit ex112 may determine the importance of the video data using the free viewpoint video and the video shot by the camera ex102.
 映像情報処理装置ex101は、解析部ex112で行われた重要度の判定結果に基づき、カメラex102に対して通信速度指示信号を送信する。映像情報処理装置ex101は、例えば、重要度が高い映像を撮影しているカメラex102に対して高い通信速度を指示する。また、映像情報処理装置ex101は、速度の制御だけではなく、重要な情報については、欠落によるデメリットを低減するために複数回送るような方式を指示する信号を送信してもよい。これにより、施設内又はスタジアム内全体の通信を効率的に行うことができる。なお、カメラex102と映像情報処理装置ex101との通信は、有線通信であっても無線通信であってもよい。また、映像情報処理装置ex101は、有線通信及び無線通信のいずれか一方のみを制御してもよい。 The video information processing apparatus ex101 transmits a communication speed instruction signal to the camera ex102 based on the importance determination result performed by the analysis unit ex112. For example, the video information processing apparatus ex101 instructs a high communication speed to the camera ex102 that captures a video with high importance. Further, the video information processing apparatus ex101 may transmit not only the speed control but also a signal instructing a method in which important information is transmitted a plurality of times in order to reduce a disadvantage caused by the lack. Thereby, communication within the facility or the entire stadium can be performed efficiently. Communication between the camera ex102 and the video information processing apparatus ex101 may be wired communication or wireless communication. The video information processing apparatus ex101 may control only one of wired communication and wireless communication.
 カメラex102は、通信速度指示信号に従った通信速度で、撮影映像データを映像情報処理装置ex101に送信する。なお、カメラex102は所定の回数再送が失敗した場合には、その撮影映像データの再送を停止し、次の撮影映像データの転送を開始してもよい。これにより、施設内又はスタジアム内全体の通信を効率的に行うことができ、解析部ex112における処理の高速化を実現できる。 The camera ex102 transmits the captured video data to the video information processing apparatus ex101 at a communication speed according to the communication speed instruction signal. Note that if the retransmission of the camera ex102 fails a predetermined number of times, the camera ex102 may stop the retransmission of the captured video data and start the transfer of the next captured video data. As a result, communication within the facility or the entire stadium can be efficiently performed, and high-speed processing in the analysis unit ex112 can be realized.
 また、カメラex102は、それぞれに割り当てられた通信速度が撮影した映像データを転送するために十分な帯域でない場合は、撮影した映像データを、割り当てられた通信速度で送信可能なビットレートの映像データに変換し、変換後の映像データを送信してもよし、映像データの転送を中止してもよい。 In addition, when the communication speed assigned to each camera ex102 is not a sufficient band for transferring the captured video data, the video data of the bit rate capable of transmitting the captured video data at the assigned communication speed. The converted video data may be transmitted, or the video data transfer may be stopped.
 また、上述したように死角の発生を防止するために映像データが使用される場合、撮影された映像データに含まれる撮影領域のうちの一部の領域のみが死角を埋めるために必要である可能性がある。この場合、カメラex102は、少なくとも、映像データから、死角の発生を防止するために必要とされる領域のみを抽出することで抽出映像データを生成し、生成された抽出映像データを映像情報処理装置ex101に送信してもよい。この構成によると、死角の発生の抑制をより少ない通信帯域で実現できる。 In addition, as described above, when video data is used to prevent the generation of blind spots, only a part of the shooting area included in the captured video data may be necessary to fill the blind spots. There is sex. In this case, the camera ex102 generates the extracted video data by extracting at least the area necessary for preventing the generation of the blind spot from the video data, and the generated extracted video data is used as the video information processing apparatus. You may transmit to ex101. According to this configuration, the occurrence of blind spots can be suppressed with a smaller communication band.
 また、例えば、付加情報の重畳表示又は映像配信が行われる場合には、カメラex102は、映像情報処理装置ex101にカメラex102の位置情報及び撮影方向の情報を送信する必要がある。この場合、映像データを転送するためには十分ではない帯域しか割り当てられなかったカメラex102は、カメラex102で検出された位置情報及び撮影方向の情報のみを送信してもよい。また、映像情報処理装置ex101においてカメラex102の位置情報及び撮影方向の情報を推定する場合は、カメラex102は、撮影した映像データを、位置情報及び撮影方向の情報の推定に必要な解像度に変換し、変換された映像データを映像情報処理装置ex101に送信してもよい。この構成によると、少ない通信帯域しか割り当てられなかったカメラex102に対しても、付加情報の重畳表示又は映像配信のサービスを提供できる。また、映像情報処理装置ex101は、より多くのカメラex102から撮影領域の情報を取得できるため、例えば注目されている領域を検出する等の目的で、撮影領域の情報を利用するような場合においても有効である。 Further, for example, when additional information is superimposed and displayed or video distribution is performed, the camera ex102 needs to transmit the position information of the camera ex102 and the shooting direction information to the video information processing apparatus ex101. In this case, the camera ex102 to which only a bandwidth that is not sufficient for transferring the video data may be transmitted, only the position information detected by the camera ex102 and the information on the shooting direction. Further, when the video information processing apparatus ex101 estimates position information and shooting direction information of the camera ex102, the camera ex102 converts the shot video data to a resolution necessary for estimating the position information and shooting direction information. The converted video data may be transmitted to the video information processing apparatus ex101. According to this configuration, it is possible to provide an additional information superimposed display or video distribution service even for the camera ex102 to which only a small communication band is allocated. In addition, since the video information processing apparatus ex101 can acquire shooting area information from a larger number of cameras ex102, for example, when the shooting area information is used for the purpose of detecting a focused area, for example. It is valid.
 なお、上述した割り当てられた通信帯域に応じた映像データの転送処理の切り替えは、通知された通信帯域に基づいてカメラex102が行ってもよいし、映像情報処理装置ex101が各カメラex102の動作を決定し、決定された動作を示す制御信号を各カメラex102に通知してもよい。これにより、動作の切り替えの判定に必要な計算量、カメラex102の処理能力、及び必要となる通信帯域等に応じて、適切に処理の分担を行える。 Note that the switching of the video data transfer process according to the allocated communication band described above may be performed by the camera ex102 based on the notified communication band, or the video information processing apparatus ex101 performs the operation of each camera ex102. The control signal indicating the determined operation may be notified to each camera ex102. As a result, the processing can be appropriately shared according to the calculation amount necessary for determining the switching of the operation, the processing capability of the camera ex102, the necessary communication band, and the like.
 (変形例9)
 解析部ex112は、映像受信装置ex103から送信された視野情報(及び/又は、視点情報)に基づき、映像データの重要度を判定してもよい。例えば、解析部ex112は、視野情報(及び/又は、視点情報)が示す領域を多く含む撮影映像データの重要度を高く設定する。また、解析部ex112は、映像中に含まれる人の数、又は移動物体の数を考慮して、映像データの重要度を判定してもよい。なお、重要度の判定方法はこれに限らない。
(Modification 9)
The analysis unit ex112 may determine the importance of the video data based on the visual field information (and / or viewpoint information) transmitted from the video reception device ex103. For example, the analysis unit ex112 sets the importance of captured video data including many areas indicated by the visual field information (and / or viewpoint information) to be high. The analysis unit ex112 may determine the importance of the video data in consideration of the number of people included in the video or the number of moving objects. Note that the importance determination method is not limited to this.
 なお、本実施の形態で説明した通信制御方法は、必ずしも複数の映像データから三次元形状の再構築を行うシステムにおいて用いられる必要はない。例えば複数のカメラex102が存在する環境において、映像データを選択的又は伝送速度に差をつけて有線通信及び/又は無線通信で送信する場合であれば、本実施の形態で説明した通信制御方法は有効である。 Note that the communication control method described in the present embodiment is not necessarily used in a system that reconstructs a three-dimensional shape from a plurality of video data. For example, in the environment where a plurality of cameras ex102 exist, if the video data is transmitted selectively or with a difference in transmission speed by wired communication and / or wireless communication, the communication control method described in the present embodiment is It is valid.
 (変形例10)
 映像配信システムにおいて、映像情報処理装置ex101は、撮影シーンの全体を示す概観映像を映像受信装置ex103に送信してもよい。
(Modification 10)
In the video distribution system, the video information processing apparatus ex101 may transmit an overview video showing the entire shooting scene to the video receiving apparatus ex103.
 具体的には、映像情報処理装置ex101は、映像受信装置ex103から送信された配信リクエストを受信した場合、保存部ex111から施設内又はスタジアム内全体の概観映像を読み出し、当該外観映像を映像受信装置ex103に送信する。この概観映像は更新間隔が長くてもよく(低フレームレートでもよく)、また画質が低くてもよい。視聴者は、映像受信装置ex103の画面上に表示された概観映像中で、見たい部分をタッチする。これにより、映像受信装置ex103は、タッチされた部分に対応する視野情報(及び/又は、視点情報)を映像情報処理装置ex101に送信する。 Specifically, when the video information processing apparatus ex101 receives the distribution request transmitted from the video receiving apparatus ex103, the video information processing apparatus ex101 reads an overview video of the entire facility or stadium from the storage unit ex111, and the external video is received by the video receiving apparatus. send to ex103. The overview video may have a long update interval (may be a low frame rate) or may have a low image quality. The viewer touches a portion to be seen in the overview video displayed on the screen of the video receiving device ex103. Accordingly, the video reception device ex103 transmits visual field information (and / or viewpoint information) corresponding to the touched portion to the video information processing device ex101.
 映像情報処理装置ex101は、視野情報(及び/又は、視点情報)に応じた映像データを保存部ex111から読み出し、当該映像データを映像受信装置ex103に送信する。 The video information processing apparatus ex101 reads video data corresponding to the visual field information (and / or viewpoint information) from the storage unit ex111, and transmits the video data to the video receiving apparatus ex103.
 また、解析部ex112は、視野情報(及び/又は、視点情報)で示される領域に対して優先的に三次元形状の復元(三次元再構成)を行うことで自由視点映像を生成する。解析部ex112は、施設内又はスタジアム内全体の三次元形状を、概観を示す程度の精度で復元する。これにより、映像情報処理装置ex101は、三次元形状の復元を効率的に行うことができる。その結果、視聴者が見たい領域の自由視点映像の高フレームレート化、及び高画質を実現できる。 In addition, the analysis unit ex112 generates a free viewpoint video by preferentially restoring the three-dimensional shape (three-dimensional reconstruction) on the region indicated by the visual field information (and / or viewpoint information). The analysis unit ex112 restores the three-dimensional shape of the entire facility or the stadium with an accuracy that shows an overview. Thereby, the video information processing apparatus ex101 can efficiently restore the three-dimensional shape. As a result, it is possible to realize a high frame rate and high image quality of a free viewpoint video in an area desired by the viewer.
 (変形例11)
 なお、映像情報処理装置ex101は、例えば、設計図面などから事前に生成された施設又はスタジアムの三次元形状復元データを事前映像として、予め保存しておいてもよい。なお、事前映像はこれに限らず、デプスセンサから得られる空間の凹凸と、過去又はキャリブレーション時の画像又は映像データから導出されるピクチャとをオブジェクトごとにマッピングした仮想空間データであってもよい。
(Modification 11)
Note that the video information processing apparatus ex101 may store in advance, for example, three-dimensional shape restoration data of a facility or a stadium generated in advance from a design drawing or the like as a preliminary video. The prior image is not limited to this, and may be virtual space data obtained by mapping, for each object, the unevenness of the space obtained from the depth sensor and the picture derived from the image or the image data at the past or during calibration.
 例えば、スタジアムでサッカーが行われている場合、解析部ex112は、選手及びボールのみに限定して三次元形状の復元を行い、得られた復元データと事前映像とを合成することで自由視点映像を生成してもよい。あるいは、解析部ex112は、選手及びボールに対して優先して三次元形状の復元を行ってもよい。これにより、映像情報処理装置ex101は、三次元形状の復元を効率的に行うことができる。その結果、視聴者が注目する選手及びボールに関する自由視点映像の高フレームレート化及び高画質化を実現できる。また、監視システムにおいては、解析部ex112は、人物及び移動物体のみに限定して、又はそれらを優先して三次元形状の復元を行ってもよい。 For example, when soccer is being performed in a stadium, the analysis unit ex112 performs reconstruction of a three-dimensional shape limited to only players and balls, and combines the obtained restoration data and a prior image to generate a free viewpoint video. May be generated. Alternatively, the analysis unit ex112 may preferentially restore the three-dimensional shape with respect to the player and the ball. Thereby, the video information processing apparatus ex101 can efficiently restore the three-dimensional shape. As a result, it is possible to realize a high frame rate and high image quality of a free viewpoint video related to the player and the ball that the viewer pays attention to. In the monitoring system, the analysis unit ex112 may perform the reconstruction of the three-dimensional shape by limiting to only the person and the moving object or giving priority to them.
 (変形例12)
 各装置の時刻は、サーバの基準時刻等に基づき、撮影開始時にキャリブレーションされてもよい。解析部ex112は、複数のカメラex102で撮影された複数の撮影映像データのうち、時刻設定の精度に応じて、予め設定された時間範囲内に属する時刻に撮影された複数の映像データを用いて、三次元形状の復元を行う。この時刻の検出には、例えば撮影映像データが保存部ex111に格納された時刻が用いられる。なお、時刻の検出方法はこれに限らない。これにより、映像情報処理装置ex101は、三次元形状の復元を効率的に行うことができるので、自由視点映像の高フレームレート化及び高画質化を実現できる。
(Modification 12)
The time of each device may be calibrated at the start of shooting based on the reference time of the server. The analysis unit ex112 uses a plurality of video data captured at a time that falls within a preset time range according to the accuracy of time setting among a plurality of captured video data captured by the plurality of cameras ex102. 3D shape restoration. For the detection of this time, for example, the time when the captured video data is stored in the storage unit ex111 is used. The time detection method is not limited to this. As a result, the video information processing apparatus ex101 can efficiently restore the three-dimensional shape, thereby realizing a high frame rate and high image quality of the free viewpoint video.
 または、解析部ex112は、保存部ex111に保存された複数の映像データのうち、高画質データのみを用いて、又は高画質データを優先的に用いて、三次元形状の復元を行ってもよい。 Alternatively, the analysis unit ex112 may restore the three-dimensional shape using only the high-quality data or using the high-quality data preferentially among the plurality of video data stored in the storage unit ex111. .
 (変形例13)
 解析部ex112は,カメラ属性情報を用いて、三次元形状の復元を行ってもよい。この場合、カメラex102は、撮影映像データとカメラ属性情報とを映像情報処理装置ex101に送信する。カメラ属性情報は、例えば、撮影位置、撮影角度、撮影時刻、又はズーム倍率などである。
(Modification 13)
The analysis unit ex112 may restore the three-dimensional shape using the camera attribute information. In this case, the camera ex102 transmits the captured video data and camera attribute information to the video information processing apparatus ex101. The camera attribute information is, for example, a shooting position, a shooting angle, a shooting time, or a zoom magnification.
 これにより、映像情報処理装置ex101は、三次元形状の復元を効率的に行うことができるので、自由視点映像の高フレームレート化及び高画質化を実現できる。 Thereby, since the video information processing apparatus ex101 can efficiently restore the three-dimensional shape, it is possible to realize a high frame rate and high image quality of the free viewpoint video.
 具体的には、カメラex102は、施設内又はスタジアム内に三次元座標を定義し、カメラex102がどのあたりの座標をどの角度から、どれ位のズームで、どの時間に撮ったかという情報を映像と共にカメラ属性情報として映像情報処理装置ex101に送信する。また、カメラex102の起動時に、施設内又はスタジアム内の通信ネットワーク上の時計とカメラ内の時計との同期がとられ、時間情報が生成される。 Specifically, the camera ex102 defines three-dimensional coordinates in the facility or in the stadium, and information about which coordinates the camera ex102 took from which angle, how much zoom, and at what time, along with the video. It transmits to the video information processing apparatus ex101 as camera attribute information. Further, when the camera ex102 is activated, the clock on the communication network in the facility or stadium is synchronized with the clock in the camera, and time information is generated.
 また、カメラex102の起動時又は任意のタイミングで施設内又はスタジアム内の特定のポイントにカメラex102を向けることにより、カメラex102の位置及び角度情報が取得される。図18は、カメラex102に起動時に、カメラex102の画面上に表示される通知の一例を示す図である。ユーザがこの通知に従い、スタジアム北側の広告中のサッカーボール中心にある「+」に、画面中央に表示された「+」を合わせて、カメラex102のディスプレイをタッチすると、カメラex102は、カメラex102から広告までのベクトル情報を取得しカメラ位置及び角度の基準を特定する。その後、カメラex102のモーション情報からその時々のカメラ座標及び角度が特定される。もちろん、この表示に限るものではなく、矢印等を用いて撮影期間中も座標、角度、又は撮影領域の移動速度等を指示するような表示が用いられてもよい。 Also, the position and angle information of the camera ex102 is acquired by pointing the camera ex102 to a specific point in the facility or stadium when the camera ex102 is activated or at an arbitrary timing. FIG. 18 is a diagram illustrating an example of a notification displayed on the screen of the camera ex102 when the camera ex102 is activated. According to this notification, when the user touches the display of the camera ex102 by aligning the “+” displayed at the center of the soccer ball in the advertisement on the north side of the stadium with the “+” displayed in the center of the screen, the camera ex102 is moved from the camera ex102. The vector information up to the advertisement is acquired and the reference of the camera position and angle is specified. Thereafter, the camera coordinates and angle at that time are specified from the motion information of the camera ex102. Needless to say, the display is not limited to this, and a display that uses an arrow or the like to indicate coordinates, an angle, a moving speed of the imaging region, or the like during the imaging period may be used.
 カメラex102の座標の特定は、GPS、WiFi(登録商標)、3G、LTE(Long Term Evolution)、及び5G(無線LAN)の電波を用いて行われてもよいし、ビーコン(Bluetooth(登録商標)、超音波)など近距離無線を利用して行われてもよい。また、施設内又はスタジアム内のどの基地局に撮影映像データが届いたかという情報が用いられてもよい。 The coordinates of the camera ex102 may be specified using GPS, WiFi (registered trademark), 3G, LTE (Long Term Evolution), and 5G (wireless LAN) radio waves, or a beacon (Bluetooth (registered trademark)). , Ultrasonic), or the like. Further, information on which base station in the facility or stadium the captured video data has reached may be used.
 (変形例14)
 当該システムはスマートフォン等のモバイル端末上で動作するアプリケーションとして提供されてもよい。
(Modification 14)
The system may be provided as an application that operates on a mobile terminal such as a smartphone.
 上記システムへのログインには、各種SNS等のアカウントが用いられてもよい。なお、アプリ専用のアカウント、又は機能が制限されたゲストアカウントが用いられてもよい。このようにアカウントが用いられることで、好みの映像又は好みのアカウント等を評価することができる。また、撮影中又は視聴中の映像データに類似した映像データ、撮影中又は視聴中の映像データの視点に類似した視点の映像データなどに優先的に帯域を割り振ることで、これらの映像データの解像度を高めることができる。これにより、これらの視点からの三次元形状の復元をより精度よく行うことができる。 An account such as various SNSs may be used to log in to the system. An application-dedicated account or a guest account with limited functions may be used. By using the account in this way, it is possible to evaluate a favorite video or a favorite account. In addition, by preferentially allocating bandwidth to video data similar to the video data being shot or viewed, video data having a viewpoint similar to the viewpoint of the video data being shot or viewed, the resolution of these video data Can be increased. Thereby, it is possible to restore the three-dimensional shape from these viewpoints with higher accuracy.
 また、ユーザは、当該アプリケーションで、好みの画像映像を選択し、相手方をフォローすることで、選択した画像を他のユーザよりも優先して見たり、相手方の承認などを条件にテキストチャット等でつながりをもつことができる。このように、新たなコミュニティの生成が可能である。 In addition, the user can select a preferred image video in the application and follow the other party, so that the selected image can be viewed with priority over other users, or can be used for text chat, etc., subject to the other party's approval. You can have a connection. In this way, a new community can be generated.
 このようにユーザ同士がコミュニティ内でつながることにより、撮影自体、また撮影した画像の共有などが活発化し、より精度の高い三次元形状の復元を促すことができる。 Thus, by connecting users within the community, shooting itself and sharing of the shot image are activated, and it is possible to promote restoration of a more accurate three-dimensional shape.
 また、コミュニティ内のつながりの設定に応じて、ユーザは、他人が撮影した画像又は映像を編集したり、他人の画像と自分の画像とをコラージュして新たな画像又は映像を作成したりできる。これにより、新たな画像又は映像を当該コミュニティ内の人のみでシェアするなど、新たな映像作品のシェアが可能になる。また、この編集においてCGのキャラクタを挿入するなどにより、拡張現実(Augmented Reality)のゲーム等にも映像作品を利用できる。 Also, according to the connection setting within the community, the user can edit an image or video taken by another person or create a new image or video by collaging the image of another person with his own image. This makes it possible to share a new video work, such as sharing a new image or video only with people in the community. Also, a video work can be used for augmented reality games by inserting a CG character in this editing.
 また、当該システムによると三次元モデルデータが逐次出力可能になるため、ゴールシーンなどの特徴的なシーンでの三次元モデルデータに基づき、施設が有する3Dプリンタなどが立体オブジェクトを出力することができる。これにより、試合後に、その試合中のシーンに基づくオブジェクトをキーホルダーのようなお土産として売ったり、参加ユーザに配布することも可能である。もちろん通常の写真として、もっとも良い視点からの画像をプリントすることも可能である。 In addition, according to the system, 3D model data can be sequentially output, so that a 3D printer or the like of a facility can output a 3D object based on 3D model data in a characteristic scene such as a goal scene. . Thereby, after a game, an object based on the scene during the game can be sold as a souvenir such as a key holder, or distributed to participating users. Of course, it is also possible to print an image from the best viewpoint as a normal photograph.
 (変形例15)
 上記システムを用いて、例えば、警察の車載カメラ、及び警察官のウェアラブルカメラの映像などから、地域全体の大雑把な状態を、当該システムに接続されたセンターで管理することができる。
(Modification 15)
By using the system described above, for example, the rough state of the entire region can be managed by the center connected to the system from the video of the police car wearable camera and the police officer wearable camera.
 一般のパトロールの時は、例えば数分おきで静止画の送受信が行なわれる。また、センターは、過去の犯罪データ等を用いて分析した結果に基づいた犯罪マップに基づいて犯罪発生の可能性が高い地域を特定する、もしくはこのように特定された犯罪発生確率に関連する地域データを保持している。特定された犯罪発生確率の高い地域では、画像の送受信の頻度を上げたり、画像を動画に変更したりしてもよい。また、事件発生時は、動画、又はSfM等を用いた三次元再構成データが用いられてもよい。また、センターもしくは各端末が、同時にデプスセンサ又はサーモセンサなど他のセンサの情報を用いて画像又は仮想空間を補正することで、警察官は、より正確に状況を把握できる。 In general patrol, for example, still images are transmitted and received every few minutes. In addition, the center identifies areas where there is a high possibility of crimes based on crime maps based on the results of analysis using past crime data, etc., or areas related to the crime occurrence probability identified in this way Holds data. In an identified area where the crime occurrence probability is high, the frequency of image transmission / reception may be increased, or the image may be changed to a moving image. Further, when an incident occurs, a moving image or three-dimensional reconstruction data using SfM or the like may be used. The center or each terminal simultaneously corrects an image or virtual space using information from other sensors such as a depth sensor or a thermo sensor, so that the police officer can grasp the situation more accurately.
 また、センターは、三次元再構成データを用いることで、複数の端末にそのオブジェクトの情報をフィードバックできる。これにより、各端末を持つ個々人がオブジェクトをトラッキングできる。 Also, the center can feed back the object information to a plurality of terminals by using the 3D reconstruction data. This allows individuals with each terminal to track the object.
 また、最近では、建造物或いは環境の調査、又はスポーツなどの臨場感ある撮影等の目的で、クワッドコプター、ドローンなどの飛行可能な装置による空中からの撮影が行なわれる。このような自律移動装置による撮影は、画像がブレるということが問題になりやすいが、SfMは位置及び傾きによりそのブレを補正しながら三次元化を行なうことが可能である。これにより、画質の向上、及び空間の復元精度の向上を実現できる。 Also, recently, for the purpose of investigating buildings or the environment, or shooting with a sense of reality such as sports, shooting from the air is performed with a flightable device such as a quadcopter or drone. Although photographing with such an autonomous mobile device tends to cause a problem that the image is blurred, SfM can be three-dimensionalized while correcting the blur based on the position and the inclination. As a result, it is possible to improve the image quality and the accuracy of space restoration.
 また、車外を撮影する車載カメラの設置が、国によっては義務付けられている。このような車載カメラにおいても、複数の画像からモデル化された三次元データを用いることで、行き先の方向の天気及び路面の状態、並びに渋滞度合い等をより精度よく把握できる。 Also, the installation of an in-vehicle camera that takes pictures outside the vehicle is obligatory in some countries. Even in such an in-vehicle camera, by using three-dimensional data modeled from a plurality of images, it is possible to more accurately grasp the weather in the direction of the destination, the state of the road surface, the degree of traffic congestion, and the like.
 (実施の形態3)
 上記各実施の形態で示した画像処理方法の構成を実現するためのプログラムを記憶メディアに記録することにより、上記各実施の形態で示した処理を独立したコンピュータシステムにおいて簡単に実施することが可能となる。記憶メディアは、磁気ディスク、光ディスク、光磁気ディスク、ICカード、半導体メモリ等、プログラムを記録できるものであればよい。
(Embodiment 3)
By recording a program for realizing the configuration of the image processing method described in each of the above embodiments on a storage medium, the processing described in each of the above embodiments can be easily performed in an independent computer system. It becomes. The storage medium may be any medium that can record a program, such as a magnetic disk, an optical disk, a magneto-optical disk, an IC card, and a semiconductor memory.
 さらにここで、上記各実施の形態で示した画像処理方法の応用例とそれを用いたシステムを説明する。当該システムは、画像処理方法を用いた装置を有することを特徴とする。システムにおける他の構成について、場合に応じて適切に変更することができる。 Furthermore, application examples of the image processing method shown in the above embodiments and a system using the same will be described here. The system includes an apparatus using an image processing method. Other configurations in the system can be appropriately changed according to circumstances.
 図19は、コンテンツ配信サービスを実現するコンテンツ供給システムex200の全体構成を示す図である。通信サービスの提供エリアを所望の大きさに分割し、各セル内にそれぞれ固定無線局である基地局ex206、ex207、ex208、ex209、ex210が設置されている。 FIG. 19 is a diagram showing an overall configuration of a content supply system ex200 that realizes a content distribution service. The communication service providing area is divided into desired sizes, and base stations ex206, ex207, ex208, ex209, and ex210, which are fixed wireless stations, are installed in each cell.
 このコンテンツ供給システムex200は、インターネットex201にインターネットサービスプロバイダex202および通信網ex204、および基地局ex206からex210を介して、コンピュータex211、PDA(Personal Digital Assistant)ex212、カメラex213、スマートフォンex214、ゲーム機ex215などの各機器が接続される。 This content supply system ex200 includes a computer ex211, a PDA (Personal Digital Assistant) ex212, a camera ex213, a smartphone ex214, a game machine ex215, etc. via the Internet ex201, the Internet service provider ex202, the communication network ex204, and the base stations ex206 to ex210. Are connected.
 しかし、コンテンツ供給システムex200は図19のような構成に限定されず、いずれかの要素を組合せて接続するようにしてもよい。また、固定無線局である基地局ex206からex210を介さずに、各機器が電話線、ケーブルテレビ、又は光通信などの通信網ex204に直接接続されてもよい。また、各機器が近距離無線等を介して直接相互に接続されていてもよい。 However, the content supply system ex200 is not limited to the configuration shown in FIG. 19 and may be connected by combining any of the elements. In addition, each device may be directly connected to a communication network ex204 such as a telephone line, cable television, or optical communication without going through the base stations ex206 to ex210 which are fixed wireless stations. In addition, the devices may be directly connected to each other via short-range wireless or the like.
 カメラex213はデジタルビデオカメラ等の動画撮影が可能な機器であり、カメラex216はデジタルカメラ等の静止画撮影、動画撮影が可能な機器である。また、スマートフォンex214は、GSM(登録商標)(Global System for Mobile Communications)方式、CDMA(Code Division Multiple Access)方式、W-CDMA(Wideband-Code Division Multiple Access)方式、若しくはLTE(Long Term Evolution)方式、HSPA(High Speed Packet Access)、又は高周波帯域を利用した通信方式などに対応するスマートフォン機、またはPHS(Personal Handyphone System)等であり、いずれでも構わない。 The camera ex213 is a device that can shoot a moving image such as a digital video camera, and the camera ex216 is a device that can shoot a still image and a moving image such as a digital camera. In addition, the smartphone ex214 is a GSM (registered trademark) (Global System for Mobile Communications) method, a CDMA (Code Division Multiple Access) method, a W-CDMA (Wideband-Code Division MultipleL method, or a Multiple Acceleration method). , HSPA (High Speed Packet Access), a smartphone corresponding to a communication method using a high frequency band, or a PHS (Personal Handyphone System), and any of them may be used.
 コンテンツ供給システムex200では、カメラex213等が基地局ex209、通信網ex204を通じてストリーミングサーバex203に接続されることで、ライブ配信等が可能になる。ライブ配信では、ユーザがカメラex213を用いて撮影するコンテンツ(例えば、音楽ライブの映像等)に対して符号化処理を行い、ストリーミングサーバex203に送信する。一方、ストリーミングサーバex203は要求のあったクライアントに対して送信されたコンテンツデータをストリーム配信する。クライアントとしては、上記符号化処理されたデータを復号化することが可能な、コンピュータex211、PDAex212、カメラex213、スマートフォンex214、ゲーム機ex215等がある。配信されたデータを受信した各機器では、受信したデータを復号化処理して再生する。 In the content supply system ex200, the camera ex213 and the like are connected to the streaming server ex203 through the base station ex209 and the communication network ex204, thereby enabling live distribution and the like. In live distribution, content (for example, music live video) that the user captures using the camera ex213 is encoded and transmitted to the streaming server ex203. On the other hand, the streaming server ex203 streams the content data transmitted to the requested client. Examples of the client include a computer ex211, a PDA ex212, a camera ex213, a smartphone ex214, and a game machine ex215 that can decode the encoded data. Each device that receives the distributed data decodes the received data and reproduces it.
 なお、撮影したデータの符号化処理はカメラex213で行っても、データの送信処理をするストリーミングサーバex203で行ってもよいし、互いに分担して行ってもよい。同様に配信されたデータの復号化処理はクライアントで行っても、ストリーミングサーバex203で行ってもよいし、互いに分担して行ってもよい。また、カメラex213に限らず、カメラex216で撮影した静止画像および/または動画像データを、コンピュータex211を介してストリーミングサーバex203に送信してもよい。この場合の符号化処理はカメラex216、コンピュータex211、ストリーミングサーバex203のいずれで行ってもよいし、互いに分担して行ってもよい。さらに復号された画像の表示についても、システムにつながった複数の機器が連動して同じ画像を表示してもよいし、大きな表示部を有する装置で全体の画像を表示し、スマートフォンex214等では画像の一部の領域を拡大して表示してもよい。 The encoded processing of the captured data may be performed by the camera ex213, the streaming server ex203 that performs the data transmission processing, or may be performed in a shared manner. Similarly, the decryption processing of the distributed data may be performed by the client, the streaming server ex203, or may be performed in common with each other. In addition to the camera ex213, still images and / or moving image data captured by the camera ex216 may be transmitted to the streaming server ex203 via the computer ex211. The encoding process in this case may be performed by any of the camera ex216, the computer ex211, and the streaming server ex203, or may be performed in a shared manner. Further, with respect to the display of the decoded image, a plurality of devices connected to the system may be linked to display the same image, or the entire image is displayed on a device having a large display unit, and the smartphone ex214 or the like displays the image. A part of the area may be enlarged and displayed.
 また、これら符号化・復号化処理は、一般的にコンピュータex211や各機器が有するLSIex500において処理する。LSIex500は、ワンチップであっても複数チップからなる構成であってもよい。なお、動画像符号化・復号化用のソフトウェアをコンピュータex211等で読み取り可能な何らかの記録メディア(CD-ROM、フレキシブルディスク、ハードディスクなど)に組み込み、そのソフトウェアを用いて符号化・復号化処理を行ってもよい。さらに、スマートフォンex214がカメラ付きである場合には、そのカメラで取得した動画データを送信してもよい。このときの動画データはスマートフォンex214が有するLSIex500で符号化処理されたデータである。 Also, these encoding / decoding processes are generally performed in the computer ex211 and the LSI ex500 included in each device. The LSI ex500 may be configured as a single chip or a plurality of chips. It should be noted that moving image encoding / decoding software is incorporated into some recording media (CD-ROM, flexible disk, hard disk, etc.) that can be read by the computer ex211 etc., and encoding / decoding processing is performed using the software. May be. Furthermore, when the smartphone ex214 has a camera, moving image data acquired by the camera may be transmitted. The moving image data at this time is data encoded by the LSI ex500 included in the smartphone ex214.
 また、ストリーミングサーバex203は複数のサーバや複数のコンピュータであって、データを分散して処理したり記録したり配信するものであってもよい。 Further, the streaming server ex203 may be a plurality of servers or a plurality of computers, and may process, record, and distribute data in a distributed manner.
 以上のようにして、コンテンツ供給システムex200では、符号化されたデータをクライアントが受信して再生することができる。このようにコンテンツ供給システムex200では、ユーザが送信した情報をリアルタイムでクライアントが受信して復号化し、再生することができ、特別な権利や設備を有さないユーザでも個人放送を実現できる。 As described above, in the content supply system ex200, the client can receive and reproduce the encoded data. As described above, in the content supply system ex200, the information transmitted by the user can be received, decrypted and reproduced by the client in real time, and even a user who does not have special rights or facilities can realize personal broadcasting.
 なお、コンテンツ供給システムex200の例に限らず、図20に示すように、デジタル放送用システムex300にも、上記各実施の形態を適用してもよい。具体的には、放送局ex301では映像データに音楽データなどが多重化された多重化データが電波を介して通信または衛星ex302に伝送される。この映像データは上記各実施の形態で説明した動画像符号化方法により符号化されたデータである。これを受けた放送衛星ex302は、放送用の電波を発信し、この電波を衛星放送の受信が可能な家庭のアンテナex304が受信する。受信した多重化データを、テレビ(受信機)ex400またはセットトップボックス(STB)ex317等の装置が復号化して再生する。 Note that the above embodiments may be applied not only to the example of the content supply system ex200 but also to the digital broadcast system ex300 as shown in FIG. Specifically, in the broadcast station ex301, multiplexed data obtained by multiplexing music data and the like on video data is transmitted to a communication or satellite ex302 via radio waves. This video data is data encoded by the moving image encoding method described in the above embodiments. Receiving this, the broadcasting satellite ex302 transmits a radio wave for broadcasting, and this radio wave is received by a home antenna ex304 capable of receiving satellite broadcasting. The received multiplexed data is decoded and reproduced by a device such as the television (receiver) ex400 or the set top box (STB) ex317.
 また、DVD、BD等の記録メディアex315、もしくはSDなどのメモリex316に記録した多重化データを読み取り復号化する、または記録メディアex315もしくはメモリex316に映像信号を符号化し、さらに場合によっては音楽信号と多重化して書き込むリーダ/レコーダex318にも上記各実施の形態で示した動画像復号化装置または動画像符号化装置を実装することが可能である。この場合、再生された映像信号はモニタex319に表示され、多重化データが記録された記録メディアex315、又はメモリex316により他の装置やシステムにおいて映像信号を再生することができる。また、ケーブルテレビ用のケーブルex303または衛星/地上波放送のアンテナex304に接続されたセットトップボックスex317内に動画像復号化装置を実装し、これをテレビのモニタex319で表示してもよい。このときセットトップボックスではなく、テレビ内に動画像復号化装置を組み込んでもよい。 Also, it reads and decodes multiplexed data recorded in a recording medium ex315 such as DVD or BD, or a memory ex316 such as SD, or encodes a video signal in the recording medium ex315 or memory ex316, and in some cases, a music signal It is possible to mount the moving picture decoding apparatus or moving picture encoding apparatus described in each of the above embodiments in the reader / recorder ex318 that writes in a multiplexed manner. In this case, the reproduced video signal is displayed on the monitor ex319, and the video signal can be reproduced in another device or system by the recording medium ex315 in which the multiplexed data is recorded or the memory ex316. In addition, a moving picture decoding apparatus may be mounted in a set-top box ex317 connected to a cable ex303 for cable television or an antenna ex304 for satellite / terrestrial broadcasting, and this may be displayed on a monitor ex319 of the television. At this time, the moving picture decoding apparatus may be incorporated in the television instead of the set top box.
 図21は、スマートフォンex214を示す図である。また、図22は、スマートフォンex214の構成例を示す図である。スマートフォンex214は、基地局ex210との間で電波を送受信するためのアンテナex450、映像、静止画を撮ることが可能なカメラ部ex465、カメラ部ex465で撮像した映像、アンテナex450で受信した映像等が復号化されたデータを表示する液晶ディスプレイ等の表示部ex458を備える。スマートフォンex214は、さらに、タッチパネル等である操作部ex466、音声を出力するためのスピーカ等である音声出力部ex457、音声を入力するためのマイク等である音声入力部ex456、撮影した映像、静止画、録音した音声、または受信した映像、静止画、メール等の符号化されたデータもしくは復号化されたデータを保存可能なメモリ部ex467、又は図20に例示されたメモリex316、もしくはユーザを特定し、ネットワークをはじめ各種データへのアクセスの認証をするためのSIMex468とのインタフェース部であるスロット部ex464を備える。 FIG. 21 is a diagram showing the smartphone ex214. FIG. 22 is a diagram illustrating a configuration example of the smartphone ex214. The smartphone ex214 includes an antenna ex450 for transmitting and receiving radio waves to and from the base station ex210, a camera unit ex465 that can take a video and a still image, a video captured by the camera unit ex465, a video received by the antenna ex450, and the like. A display unit ex458 such as a liquid crystal display for displaying the decrypted data is provided. The smartphone ex214 further includes an operation unit ex466 such as a touch panel, an audio output unit ex457 such as a speaker for outputting audio, an audio input unit ex456 such as a microphone for inputting audio, a captured video, a still image , A memory portion ex467 that can store the recorded audio, or the encoded data or the decoded data such as received video, still image, or mail, or the memory ex316 illustrated in FIG. And a slot part ex464 which is an interface part with the SIMex 468 for authenticating access to various data including the network.
 スマートフォンex214は、表示部ex458及び操作部ex466等を統括的に制御する主制御部ex460に対して、電源回路部ex461、操作入力制御部ex462、映像信号処理部ex455、カメラインタフェース部ex463、LCD(Liquid Crystal Display)制御部ex459、変調/復調部ex452、多重/分離部ex453、音声信号処理部ex454、スロット部ex464、メモリ部ex467がバスex470を介して互いに接続されている。 The smartphone ex214 controls the power supply circuit ex461, the operation input control unit ex462, the video signal processing unit ex455, the camera interface unit ex463, the LCD (for the main control unit ex460 that comprehensively controls the display unit ex458, the operation unit ex466, and the like. A Liquid Crystal Display) control unit ex459, a modulation / demodulation unit ex452, a multiplexing / demultiplexing unit ex453, an audio signal processing unit ex454, a slot unit ex464, and a memory unit ex467 are connected to each other via a bus ex470.
 電源回路部ex461は、ユーザの操作により終話及び電源キーがオン状態にされると、バッテリパックから各部に対して電力を供給することによりスマートフォンex214を動作可能な状態に起動する。 When the end of call and the power key are turned on by a user operation, the power supply circuit unit ex461 starts up the smartphone ex214 in an operable state by supplying power from the battery pack to each unit.
 スマートフォンex214は、CPU、ROM、RAM等を有する主制御部ex460の制御に基づいて、音声通話モード時に音声入力部ex456で収音した音声信号を音声信号処理部ex454でデジタル音声信号に変換し、これを変調/復調部ex452でスペクトラム拡散処理し、送信/受信部ex451でデジタルアナログ変換処理および周波数変換処理を施した後にアンテナex450を介して送信する。またスマートフォンex214は、音声通話モード時にアンテナex450を介して受信した受信データを増幅して周波数変換処理およびアナログデジタル変換処理を施し、変調/復調部ex452でスペクトラム逆拡散処理し、音声信号処理部ex454でアナログ音声信号に変換した後、これを音声出力部ex457から出力する。 The smartphone ex214 converts the audio signal collected by the audio input unit ex456 in the audio call mode into a digital audio signal by the audio signal processing unit ex454 based on the control of the main control unit ex460 having a CPU, a ROM, a RAM, and the like. This is subjected to spectrum spread processing by the modulation / demodulation unit ex452, and is subjected to digital analog conversion processing and frequency conversion processing by the transmission / reception unit ex451, and then transmitted via the antenna ex450. In addition, the smartphone ex214 amplifies reception data received via the antenna ex450 in the voice call mode, performs frequency conversion processing and analog-digital conversion processing, performs spectrum despreading processing in the modulation / demodulation unit ex452, and performs voice signal processing unit ex454. After being converted into an analog audio signal, the audio output unit ex457 outputs it.
 さらにデータ通信モード時に電子メールを送信する場合、本体部の操作部ex466等の操作によって入力された電子メールのテキストデータは操作入力制御部ex462を介して主制御部ex460に送出される。主制御部ex460は、テキストデータを変調/復調部ex452でスペクトラム拡散処理をし、送信/受信部ex451でデジタルアナログ変換処理および周波数変換処理を施した後にアンテナex450を介して基地局ex210へ送信する。電子メールを受信する場合は、受信したデータに対してこのほぼ逆の処理が行われ、表示部ex458に出力される。 Further, when an e-mail is transmitted in the data communication mode, the text data of the e-mail input by the operation of the operation unit ex466 of the main unit is sent to the main control unit ex460 via the operation input control unit ex462. The main control unit ex460 performs spread spectrum processing on the text data in the modulation / demodulation unit ex452, performs digital analog conversion processing and frequency conversion processing in the transmission / reception unit ex451, and then transmits the text data to the base station ex210 via the antenna ex450. . In the case of receiving an e-mail, almost the reverse process is performed on the received data and output to the display unit ex458.
 データ通信モード時に映像、静止画、または映像と音声を送信する場合、映像信号処理部ex455は、カメラ部ex465から供給された映像信号を上記各実施の形態で示した動画像符号化方法によって圧縮符号化し、符号化された映像データを多重/分離部ex453に送出する。また、音声信号処理部ex454は、映像、静止画等をカメラ部ex465で撮像中に音声入力部ex456で収音した音声信号を符号化し、符号化された音声データを多重/分離部ex453に送出する。 When transmitting video, still images, or video and audio in the data communication mode, the video signal processing unit ex455 compresses the video signal supplied from the camera unit ex465 by the moving image encoding method described in each of the above embodiments. The encoded video data is sent to the multiplexing / demultiplexing unit ex453. The audio signal processing unit ex454 encodes the audio signal picked up by the audio input unit ex456 while the camera unit ex465 captures video, still images, and the like, and sends the encoded audio data to the multiplexing / separating unit ex453. To do.
 多重/分離部ex453は、映像信号処理部ex455から供給された符号化された映像データと音声信号処理部ex454から供給された符号化された音声データを所定の方式で多重化し、その結果得られる多重化データを変調/復調部(変調/復調回路部)ex452でスペクトラム拡散処理をし、送信/受信部ex451でデジタルアナログ変換処理及び周波数変換処理を施した後にアンテナex450を介して送信する。 The multiplexing / demultiplexing unit ex453 multiplexes the encoded video data supplied from the video signal processing unit ex455 and the encoded audio data supplied from the audio signal processing unit ex454 by a predetermined method, and is obtained as a result. The multiplexed data is subjected to spread spectrum processing by a modulation / demodulation unit (modulation / demodulation circuit unit) ex452, and subjected to digital analog conversion processing and frequency conversion processing by a transmission / reception unit ex451, and then transmitted through an antenna ex450.
 データ通信モード時にホームページ等にリンクされた動画像ファイルのデータを受信する場合、または映像およびもしくは音声が添付された電子メールを受信する場合、アンテナex450を介して受信された多重化データを復号化するために、多重/分離部ex453は、多重化データを分離することにより映像データのビットストリームと音声データのビットストリームとに分け、同期バスex470を介して符号化された映像データを映像信号処理部ex455に供給するとともに、符号化された音声データを音声信号処理部ex454に供給する。映像信号処理部ex455は、上記各実施の形態で示した動画像符号化方法に対応した動画像復号化方法によって復号化することにより映像信号を復号し、LCD制御部ex459を介して表示部ex458から、例えばホームページにリンクされた動画像ファイルに含まれる映像、静止画が表示される。また音声信号処理部ex454は、音声信号を復号し、音声出力部ex457から音声が出力される。 Decode multiplexed data received via antenna ex450 when receiving data of moving image files linked to websites in data communication mode or when receiving e-mails with video and / or audio attached For this, the multiplexing / separating unit ex453 separates the multiplexed data into a video data bit stream and an audio data bit stream, and performs video signal processing on the video data encoded via the synchronization bus ex470. The encoded audio data is supplied to the audio signal processing unit ex454 while being supplied to the unit ex455. The video signal processing unit ex455 decodes the video signal by decoding using a video decoding method corresponding to the video encoding method shown in each of the above embodiments, and the display unit ex458 via the LCD control unit ex459. From, for example, video and still images included in a moving image file linked to a home page are displayed. The audio signal processing unit ex454 decodes the audio signal, and the audio is output from the audio output unit ex457.
 また、上記スマートフォンex214等の端末は、テレビex400と同様に、符号化器・復号化器を両方持つ送受信型端末の他に、符号化器のみの送信端末、復号化器のみの受信端末という3通りの実装形式が考えられる。さらに、デジタル放送用システムex300において、映像データに音楽データなどが多重化された多重化データを受信、送信するとして説明したが、音声データ以外に映像に関連する文字データなどが多重化されたデータであってもよいし、多重化データではなく映像データ自体であってもよい。 In addition to the transmission / reception terminal having both the encoder and the decoder, the terminal such as the smartphone ex214 is a transmission terminal having only an encoder and a receiving terminal having only a decoder, as well as the television ex400. A possible implementation format is possible. Furthermore, in the digital broadcasting system ex300, it has been described that multiplexed data in which music data or the like is multiplexed with video data is received and transmitted. However, data in which character data related to video is multiplexed in addition to audio data It may be video data itself instead of multiplexed data.
 また、本発明はかかる上記実施の形態に限定されるものではなく、本発明の範囲を逸脱することなく種々の変形または修正が可能である。 Further, the present invention is not limited to the above-described embodiment, and various changes and modifications can be made without departing from the scope of the present invention.
 本発明は、複数のカメラで撮影された映像を配信する映像配信システムに適用でききる。 The present invention can be applied to a video distribution system that distributes video shot by a plurality of cameras.
 100 映像配信システム
 101 カメラ
 102 端末装置
 103 サーバ
 104 ネットワーク
 111,121 受信部
 112 映像蓄積部
 113,126 制御部
 114,125 送信部
 122 蓄積部
 123 復号部
 124 出力部
 127 入力部
 151 映像信号
 152 視点指定信号
 153 選択映像信号
 154 関連映像信号
 155 出力映像
 201 背景画像
 202 カメラアイコン
 211 選択映像
 212 俯瞰画像
 213 上面画像
 214,215,216 操作ボタン
DESCRIPTION OF SYMBOLS 100 Video delivery system 101 Camera 102 Terminal device 103 Server 104 Network 111,121 Reception part 112 Video storage part 113,126 Control part 114,125 Transmission part 122 Storage part 123 Decoding part 124 Output part 127 Input part 151 Video signal 152 Viewpoint designation | designated Signal 153 Selection video signal 154 Related video signal 155 Output video 201 Background image 202 Camera icon 211 Selection video 212 Overhead image 213 Top image 214, 215, 216 Operation buttons

Claims (17)

  1.  複数のユーザにより異なる視点から撮影された複数の映像のいずれかを端末装置に配信するサーバによる映像配信方法であって、
     前記複数の映像の一つであり、前記端末装置から要求された第1映像を前記端末装置に配信する配信ステップと、
     前記複数の映像の一つであり、前記端末装置から次に要求される可能性が高い第2映像を選択する選択ステップと、
     前記第1映像を前記端末装置に配信している間に前記第2映像の前記端末装置への送信を開始する送信ステップとを含む
     映像配信方法。
    A video delivery method by a server that delivers any of a plurality of videos taken from different viewpoints by a plurality of users to a terminal device,
    A delivery step of delivering the first video requested by the terminal device to the terminal device, the one of the plurality of videos;
    A selection step of selecting a second video that is one of the plurality of videos and is likely to be requested next from the terminal device;
    A transmission step of starting transmission of the second video to the terminal device while distributing the first video to the terminal device.
  2.  前記選択ステップでは、前記複数の映像のうち、前記第1映像との関連度が高い映像を前記第2映像として選択する
     請求項1記載の映像配信方法。
    The video distribution method according to claim 1, wherein in the selection step, a video having a high degree of association with the first video is selected as the second video from the plurality of videos.
  3.  前記選択ステップでは、撮影シーンの位置が前記第1映像の撮影シーンの位置に近いほど、前記関連度が高いと判定する
     請求項2記載の映像配信方法。
    The video distribution method according to claim 2, wherein the selection step determines that the degree of association is higher as the position of the shooting scene is closer to the position of the shooting scene of the first video.
  4.  前記選択ステップでは、さらに、前記撮影シーンの広さが前記第1映像の前記撮影シーンの広さに近いほど、前記関連度が高いと判定する
     請求項3記載の映像配信方法。
    The video distribution method according to claim 3, wherein the selection step further determines that the degree of association is higher as the width of the shooting scene is closer to the width of the shooting scene of the first video.
  5.  前記選択ステップでは、前記第1映像に含まれる被写体と、同一の被写体が撮影されている映像の前記関連度を高く設定する
     請求項2記載の映像配信方法。
    The video distribution method according to claim 2, wherein, in the selection step, the degree of association between a subject included in the first video and a video in which the same subject is shot is set high.
  6.  前記選択ステップでは、前記複数の映像のフレームレート、解像度又はビットレートに基づき、前記第2映像を選択する
     請求項1記載の映像配信方法。
    The video distribution method according to claim 1, wherein, in the selection step, the second video is selected based on a frame rate, resolution, or bit rate of the plurality of videos.
  7.  前記選択ステップでは、前記複数の映像のうち、他のユーザに選択された回数が多い映像を前記第2映像として選択する
     請求項1記載の映像配信方法。
    The video distribution method according to claim 1, wherein, in the selection step, a video that has been frequently selected by another user among the plurality of videos is selected as the second video.
  8.  前記選択ステップでは、ユーザの視聴履歴又は予め登録されている嗜好情報に基づき、前記第2映像を選択する
     請求項1記載の映像配信方法。
    The video distribution method according to claim 1, wherein, in the selection step, the second video is selected based on a user's viewing history or pre-registered preference information.
  9.  複数の視点から撮影された複数の映像のいずれかをサーバから受信し、当該映像を表示する端末装置による映像受信方法であって、
     前記複数の映像から第1映像を選択する選択ステップと、
     前記第1映像の送信を前記サーバに要求する要求ステップと、
     前記第1映像を前記サーバから受信する第1受信ステップと、
     前記第1映像を表示する表示ステップと、
     前記第1映像の受信中に、前記複数の映像の一つであり、次に選択される可能性が高い第2映像の受信を開始する第2受信ステップとを含む
     映像受信方法。
    A video reception method by a terminal device that receives any of a plurality of videos taken from a plurality of viewpoints from a server and displays the videos,
    A selection step of selecting a first video from the plurality of videos;
    A requesting step for requesting the server to transmit the first video;
    A first receiving step of receiving the first video from the server;
    A display step for displaying the first video;
    A second receiving step of starting reception of a second video that is one of the plurality of videos and is likely to be selected next during reception of the first video;
  10.  前記映像受信方法は、さらに、
     前記受信された前記第2映像を蓄積するステップと、
     前記第1映像の表示中に前記第2映像が選択された場合に、前記蓄積されている前記第2映像を表示するステップとを含む
     請求項9記載の映像受信方法。
    The video receiving method further includes:
    Storing the received second video;
    The video receiving method according to claim 9, further comprising: displaying the stored second video when the second video is selected during display of the first video.
  11.  前記映像配信方法は、さらに、
     前記第1映像の表示中に、前記第1映像及び前記第2映像と異なる第3映像が選択された場合、前記第3映像を前記サーバから受信するステップと、
     前記第3映像を受信するまでの間、前記蓄積されている前記第2映像を表示するステップとを含む
     請求項10記載の映像受信方法。
    The video distribution method further includes:
    Receiving a third video from the server when a third video different from the first video and the second video is selected during the display of the first video;
    The video receiving method according to claim 10, further comprising: displaying the stored second video until the third video is received.
  12.  前記表示ステップでは、さらに、前記複数の映像が撮影されている場所を俯瞰する画像であり、前記複数の視点の位置を示す複数のアイコンを含む画像を表示する
     請求項9~11のいずれか1項に記載の映像受信方法。
    The display step further displays an image that is a bird's-eye view of a place where the plurality of videos are taken and includes a plurality of icons indicating the positions of the plurality of viewpoints. The video receiving method according to the item.
  13.  前記表示ステップでは、前記複数のアイコンのうち、前記第2映像の視点の位置を示すアイコンが強調表示される
     請求項12記載の映像受信方法。
    The video receiving method according to claim 12, wherein in the display step, an icon indicating a position of a viewpoint of the second video is highlighted among the plurality of icons.
  14.  複数のユーザにより異なる視点から撮影された複数の映像のいずれかを端末装置に配信するサーバであって、
     前記複数の映像の一つであり、前記端末装置から指定された第1映像を前記端末装置に配信する配信部と、
     前記複数の映像の一つであり、前記端末装置から次に要求される可能性が高い第2映像を選択する選択部と、
     前記第1映像を前記端末装置に配信している間に前記第2映像の前記端末装置への送信を開始する送信部とを備える
     サーバ。
    A server for delivering any of a plurality of videos taken from different viewpoints by a plurality of users to a terminal device,
    A distribution unit that is one of the plurality of images and distributes the first image designated by the terminal device to the terminal device;
    A selection unit that selects a second video that is one of the plurality of videos and is likely to be requested next from the terminal device;
    A transmission unit configured to start transmission of the second video to the terminal device while delivering the first video to the terminal device.
  15.  複数の視点から撮影された複数の映像のいずれかをサーバから受信し、当該映像を表示する端末装置であって、
     前記複数の映像から第1映像を選択する選択部と、
     前記第1映像の送信を前記サーバに要求する要求部と、
     前記第1映像を前記サーバから受信する第1受信部と、
     前記第1映像を表示する表示部と、
     前記第1映像の受信中に、前記複数の映像の一つであり、次に選択される可能性が高い第2映像の受信を開始する第2受信部とを備える
     端末装置。
    A terminal device that receives any of a plurality of videos taken from a plurality of viewpoints from a server and displays the videos,
    A selection unit for selecting a first video from the plurality of videos;
    A request unit that requests the server to transmit the first video;
    A first receiver for receiving the first video from the server;
    A display unit for displaying the first video;
    A terminal device comprising: a second receiving unit that starts receiving a second video that is one of the plurality of videos and is likely to be selected next during reception of the first video.
  16.  請求項14記載のサーバと、
     請求項15記載の端末装置とを含む
     映像配信システム。
    A server according to claim 14;
    A video distribution system comprising the terminal device according to claim 15.
  17.  請求項9記載の映像受信方法をコンピュータに実行させるための
     プログラム。
    A program for causing a computer to execute the video receiving method according to claim 9.
PCT/JP2015/001655 2014-04-14 2015-03-24 Image delivery method, image reception method, server, terminal apparatus, and image delivery system WO2015159487A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP15779927.1A EP3133819A1 (en) 2014-04-14 2015-03-24 Image delivery method, image reception method, server, terminal apparatus, and image delivery system
US15/285,736 US10271082B2 (en) 2014-04-14 2016-10-05 Video distribution method, video reception method, server, terminal apparatus, and video distribution system

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
JP2014-082774 2014-04-14
JP2014082774A JP2015204512A (en) 2014-04-14 2014-04-14 Information processing apparatus, information processing method, camera, reception device, and reception method
US201462015601P 2014-06-23 2014-06-23
US62/015,601 2014-06-23
JP2015045352A JP6607433B2 (en) 2014-06-23 2015-03-06 Video distribution method and server
JP2015-045352 2015-03-06

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/285,736 Continuation US10271082B2 (en) 2014-04-14 2016-10-05 Video distribution method, video reception method, server, terminal apparatus, and video distribution system

Publications (1)

Publication Number Publication Date
WO2015159487A1 true WO2015159487A1 (en) 2015-10-22

Family

ID=54323715

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2015/001655 WO2015159487A1 (en) 2014-04-14 2015-03-24 Image delivery method, image reception method, server, terminal apparatus, and image delivery system

Country Status (1)

Country Link
WO (1) WO2015159487A1 (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105915839A (en) * 2015-12-07 2016-08-31 乐视云计算有限公司 Multi-channel video display method of broadcast instructing platform and multi-channel video display device thereof
CN105959596A (en) * 2016-05-24 2016-09-21 深圳市华泰敏信息技术有限公司 Backup method and device of video source
JP2017139725A (en) * 2016-02-03 2017-08-10 パナソニックIpマネジメント株式会社 Image display method and image display device
CN109644265A (en) * 2016-05-25 2019-04-16 佳能株式会社 Control device, control method and storage medium
JP2019514236A (en) * 2016-02-03 2019-05-30 ソニー株式会社 System and method for capturing still and / or moving scenes using multiple camera networks
JP2021182443A (en) * 2019-08-07 2021-11-25 キヤノン株式会社 Transmission device and transmission method, and program
JP2021185695A (en) * 2015-12-28 2021-12-09 日本電気株式会社 Monitoring support system, monitoring support method, and program
US11223821B2 (en) 2016-02-03 2022-01-11 Panasonic Intellectual Property Management Co., Ltd. Video display method and video display device including a selection of a viewpoint from a plurality of viewpoints
JP7474521B2 (en) 2022-04-08 2024-04-25 三栄通信工業株式会社 Information processing device and program

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH10322680A (en) * 1997-05-16 1998-12-04 Nippon Telegr & Teleph Corp <Ntt> Chained multiple viewpoint video reproducing method
JP2011135138A (en) * 2009-12-22 2011-07-07 Canon Inc Video reproducing device and method of controlling the same
JP2011254181A (en) * 2010-05-31 2011-12-15 Nippon Telegr & Teleph Corp <Ntt> Distributed moving picture quality selection device and moving picture distribution device and method and program
JP2012034083A (en) * 2010-07-29 2012-02-16 Canon Inc Video processing apparatus and control method of the same
JP2013183209A (en) * 2012-02-29 2013-09-12 Nagoya Univ System and method for viewing multi-viewpoint video stream

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH10322680A (en) * 1997-05-16 1998-12-04 Nippon Telegr & Teleph Corp <Ntt> Chained multiple viewpoint video reproducing method
JP2011135138A (en) * 2009-12-22 2011-07-07 Canon Inc Video reproducing device and method of controlling the same
JP2011254181A (en) * 2010-05-31 2011-12-15 Nippon Telegr & Teleph Corp <Ntt> Distributed moving picture quality selection device and moving picture distribution device and method and program
JP2012034083A (en) * 2010-07-29 2012-02-16 Canon Inc Video processing apparatus and control method of the same
JP2013183209A (en) * 2012-02-29 2013-09-12 Nagoya Univ System and method for viewing multi-viewpoint video stream

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP3133819A4 *

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105915839A (en) * 2015-12-07 2016-08-31 乐视云计算有限公司 Multi-channel video display method of broadcast instructing platform and multi-channel video display device thereof
JP2021185695A (en) * 2015-12-28 2021-12-09 日本電気株式会社 Monitoring support system, monitoring support method, and program
JP7444228B2 (en) 2015-12-28 2024-03-06 日本電気株式会社 program
JP7363942B2 (en) 2015-12-28 2023-10-18 日本電気株式会社 Programs, information transmission methods, computer-readable storage media, and information transmission systems
JP7188513B2 (en) 2015-12-28 2022-12-13 日本電気株式会社 MONITORING SYSTEM, MONITORING METHOD, AND PROGRAM
EP3413570A4 (en) * 2016-02-03 2019-01-23 Panasonic Intellectual Property Management Co., Ltd. Video display method and video display device
JP2019514236A (en) * 2016-02-03 2019-05-30 ソニー株式会社 System and method for capturing still and / or moving scenes using multiple camera networks
US11223821B2 (en) 2016-02-03 2022-01-11 Panasonic Intellectual Property Management Co., Ltd. Video display method and video display device including a selection of a viewpoint from a plurality of viewpoints
JP2017139725A (en) * 2016-02-03 2017-08-10 パナソニックIpマネジメント株式会社 Image display method and image display device
CN105959596A (en) * 2016-05-24 2016-09-21 深圳市华泰敏信息技术有限公司 Backup method and device of video source
CN109644265A (en) * 2016-05-25 2019-04-16 佳能株式会社 Control device, control method and storage medium
JP2021182443A (en) * 2019-08-07 2021-11-25 キヤノン株式会社 Transmission device and transmission method, and program
JP7204843B2 (en) 2019-08-07 2023-01-16 キヤノン株式会社 Transmission device, transmission method, and program
JP7474521B2 (en) 2022-04-08 2024-04-25 三栄通信工業株式会社 Information processing device and program

Similar Documents

Publication Publication Date Title
JP6607433B2 (en) Video distribution method and server
JP6948624B2 (en) Video distribution method and server
WO2018030206A1 (en) Camerawork generating method and video processing device
JP7113294B2 (en) Multi-view imaging system
US10271082B2 (en) Video distribution method, video reception method, server, terminal apparatus, and video distribution system
JP7203356B2 (en) Imaging system
JP7223978B2 (en) Calibration device and calibration method
JP6820527B2 (en) Video synchronization device and video synchronization method
WO2015159487A1 (en) Image delivery method, image reception method, server, terminal apparatus, and image delivery system
WO2017134706A1 (en) Video display method and video display device
US10862977B2 (en) Method for sharing photographed images between users
JP7122694B2 (en) Imaging system and calibration method
JP6460105B2 (en) Imaging method, imaging system, and terminal device
JP2017139725A (en) Image display method and image display device
WO2015194082A1 (en) Image processing method and image processing system
WO2015182034A1 (en) Image shooting method, image shooting system, server, image shooting apparatus, and image shooting program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15779927

Country of ref document: EP

Kind code of ref document: A1

REEP Request for entry into the european phase

Ref document number: 2015779927

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2015779927

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: DE