WO2016088437A1 - 情報処理装置、情報処理方法およびプログラム - Google Patents
情報処理装置、情報処理方法およびプログラム Download PDFInfo
- Publication number
- WO2016088437A1 WO2016088437A1 PCT/JP2015/077216 JP2015077216W WO2016088437A1 WO 2016088437 A1 WO2016088437 A1 WO 2016088437A1 JP 2015077216 W JP2015077216 W JP 2015077216W WO 2016088437 A1 WO2016088437 A1 WO 2016088437A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- information processing
- processing apparatus
- space
- target position
- image data
- Prior art date
Links
- 230000010365 information processing Effects 0.000 title claims abstract description 71
- 238000003672 processing method Methods 0.000 title description 5
- 230000000007 visual effect Effects 0.000 claims abstract description 44
- 230000006870 function Effects 0.000 description 20
- 238000004891 communication Methods 0.000 description 17
- 238000000034 method Methods 0.000 description 16
- 238000012545 processing Methods 0.000 description 16
- 238000010586 diagram Methods 0.000 description 12
- 238000003384 imaging method Methods 0.000 description 9
- 230000000694 effects Effects 0.000 description 5
- 210000001508 eye Anatomy 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 239000004065 semiconductor Substances 0.000 description 3
- 235000002673 Dioscorea communis Nutrition 0.000 description 2
- 241000544230 Dioscorea communis Species 0.000 description 2
- 208000035753 Periorbital contusion Diseases 0.000 description 2
- 210000005252 bulbus oculi Anatomy 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 238000005401 electroluminescence Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000001151 other effect Effects 0.000 description 2
- 239000013589 supplement Substances 0.000 description 2
- 230000001133 acceleration Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
- G01B11/02—Measuring arrangements characterised by the use of optical techniques for measuring length, width or thickness
- G01B11/03—Measuring arrangements characterised by the use of optical techniques for measuring length, width or thickness by measuring coordinates of points
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
- G06T11/60—Editing figures and text; Combining figures or text
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
- H04N23/611—Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/66—Remote control of cameras or camera parts, e.g. by remote control devices
- H04N23/661—Transmitting camera control signals through networks, e.g. control via the Internet
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/90—Arrangement of cameras or camera modules, e.g. multiple cameras in TV studios or sports stadiums
Definitions
- This disclosure relates to an information processing apparatus, an information processing method, and a program.
- Patent document 1 is an example of such a technique.
- the present disclosure proposes a new and improved information processing apparatus, information processing method, and program that make it possible to appropriately determine a position in a space that is an object of image generation.
- An information processing apparatus including a processor is provided.
- the information indicating the viewing target position from a plurality of viewpoints existing in the space is collected, and the processor evaluates the position in the space according to the density of the viewing target position.
- the function of collecting information indicating the positions of the visual target from a plurality of viewpoints existing in the space and the function of evaluating the position in the space according to the density of the visual target positions.
- a program to be implemented by a processor is provided.
- FIG. 5 is a schematic flowchart illustrating a process for estimating a spatial audience rating according to an embodiment of the present disclosure. It is a figure for demonstrating notionally spatial audience rating estimation in one Embodiment of this indication. It is a figure which shows the other example of the visual target position estimation in one Embodiment of this indication.
- 3 is a diagram for conceptually explaining an example of image collection according to an embodiment of the present disclosure.
- FIG. 10 is a diagram for conceptually explaining another example of image collection according to an embodiment of the present disclosure.
- FIG. FIG. 10 is a diagram for conceptually explaining still another example of image collection according to an embodiment of the present disclosure.
- FIG. 10 is a diagram for conceptually explaining still another example of image collection according to an embodiment of the present disclosure.
- FIG. 10 is a flowchart illustrating an example of processing according to an embodiment of the present disclosure.
- FIG. 3 is a block diagram illustrating a hardware configuration example of an information processing apparatus according to an embodiment of the present disclosure.
- the spatial rating is estimated based on the density of visual target positions from a plurality of viewpoints existing in the space.
- the spatial audience rating shows how much attention each position in the space is attracting.
- the plurality of viewpoints may include, for example, viewpoints of spectators in the stadium.
- the plurality of viewpoints may include, for example, viewpoints of cameras installed in a stadium.
- the viewing target position from these viewpoints is, for example, a position on the line of sight. It is estimated that the position where the line of sight from the audience and the camera is concentrated, that is, the position where the spatial audience rating is high, is particularly attracting attention in the stadium.
- three-dimensional point cloud (point cloud) data for providing a free viewpoint image is generated.
- an image includes a moving image (video) and a still image.
- FIG. 1 is a schematic flowchart illustrating a process for estimating a spatial audience rating according to an embodiment of the present disclosure.
- the viewer position is estimated (S101).
- the viewer means a subject having a viewpoint existing in the space, such as a spectator or a camera.
- the viewer position is estimated by performing self-position estimation based on a viewpoint image provided from the viewer, for example.
- a technique such as SLAM (Simultaneous Localization and Mapping) can be used for self-position estimation based on an image.
- SLAM Simultaneous Localization and Mapping
- information on landmarks in the space or information such as a three-dimensional model of the space may be used.
- the viewer position may be estimated based on information provided from the viewer himself.
- positioning using radio waves from a base station such as GNSS (Global Navigation Satellite Systems), Wi-Fi, or space-specific position expression (for example, the position of a spectator seat indicated by ticket information such as a stadium) ) can be used.
- GNSS Global Navigation Satellite Systems
- Wi-Fi Wireless Fidelity
- space-specific position expression for example, the position of a spectator seat indicated by ticket information such as a stadium
- the center direction of the viewpoint image provided from the viewer may be regarded as the line-of-sight direction.
- the viewer is a spectator wearing a wearable camera
- more detailed viewpoint estimation is possible.
- an eye image captured by an inward camera which is provided together with a viewpoint image captured by a spectator with an outward camera, is used.
- the line of sight can be estimated from the center position of the black eye, or the line of sight can be estimated from the posture of the eyeball model.
- the viewer's visual target position is estimated based on the viewer position and line of sight estimated in S101 and S103 above (S105).
- the visual target position is specified as a position on the line of sight emitted from the viewer position, for example.
- a position where the line of sight intersects the ground or the floor may be estimated as a viewing target position using a three-dimensional model of space.
- the focal length can be used with a camera or the like
- the visual target position on the line of sight may be estimated based on the focal length.
- the processing up to the estimation of the viewing target position in S105 may be executed by, for example, a client device corresponding to each viewer, or may be executed by a server that collects data such as images from the client device.
- the spatial audience rating is estimated based on the viewing target position estimated in S105.
- the spatial audience rating is estimated based on the density of the visual target position in the space.
- information indicating the viewing target position of each viewer, that is, the viewing target position from a plurality of viewpoints existing in the space has been collected by the server that estimates the spatial audience rating.
- the spatial audience rating may be calculated, for example, for each grid that divides the space into a predetermined size, or may be calculated for each cluster including viewing target positions that are close to each other.
- the spatial audience rating estimated by the above processing is used to generate point cloud data for providing a free viewpoint image. More specifically, for example, point cloud data is generated for a position where the spatial audience rating (the density of the position to be viewed) exceeds a threshold value.
- point cloud data is generated for a position where the spatial audience rating (the density of the position to be viewed) exceeds a threshold value.
- there are many methods for providing a free viewpoint image Even if data for providing a free viewpoint image using such a method is generated based on the spatial data rate, Good.
- a technique for providing a free viewpoint image is described in, for example, Japanese Patent Application Laid-Open No. 2007-133660.
- Free viewpoint images allow you to observe any position in space from any angle.
- a free viewpoint image is provided without limiting the position, for example, a user who views the captured image afterwards does not know which position to view, and therefore can always provide a better experience to the user. Is not limited.
- the data becomes enormous. Therefore, in the present embodiment, by limiting the positions to which the free viewpoint image is provided based on the spatial audience rating, the free viewpoint image is provided in a form that is easy for the user to view, and the free viewpoint image is provided. Prevent data from becoming unnecessarily large.
- FIG. 2 is a diagram for conceptually explaining spatial audience rating estimation according to an embodiment of the present disclosure.
- the viewer position 110 existing in the space SP (the stadium in the illustrated example) is estimated.
- the viewer position 110 may be a spectator providing a viewpoint image or a camera position.
- the line of sight 115 is estimated.
- the line of sight 115 may be regarded as, for example, along the center direction of the viewpoint image, or may be detected using a line-of-sight detection technique when the viewer is a spectator.
- the visual target position 120 in the space SP is estimated.
- the visual target position 120 may be estimated as a position where the line of sight 115 intersects the ground (or floor surface) as in the illustrated example.
- the spatial audience rating is estimated according to the density of the viewing target position 120. In the illustrated example, a relatively high spatial audience rating is estimated at the position P1 and the position P2 where the visual target position 120 is concentrated.
- FIG. 3 is a diagram illustrating another example of visual target position estimation according to an embodiment of the present disclosure.
- the viewing target position is estimated based on the viewpoint image provided by the viewer.
- the image 132 captured by the environment installation camera 130 is separated from the viewpoint image. Based on this, the viewer position 110 and the line of sight 115 are estimated.
- the image 132 includes a viewer who is a spectator (VIEWER) as a subject.
- VIEWER spectator
- the viewer position, viewer face orientation, eye orientation, etc. in the image may be specified, and the line of sight 115 may be estimated based on these.
- the result of machine learning regarding the image may be used for the detection of the viewer and the identification of the face orientation, the eye orientation, and the like.
- FIG. 4 is a diagram for conceptually explaining an example of image collection according to an embodiment of the present disclosure.
- images selected at each of the viewpoints 210 arranged in the space SP are selected based on the space audience rating.
- An image including the position 220 is collected for generation of a free viewpoint image.
- the viewpoint 210 is for a spectator (may be a professional cameraman) who wears a wearable camera or performs shooting with a mobile camera (such as a camera mounted on a normal digital camera or a smartphone). It can be a position.
- the viewpoint 210 may include the position of a camera that is fixedly installed for relaying or the like.
- an image photographed at the viewpoint 210 can be used for estimating the spatial audience rating.
- This image may be used as it is as an image for generating a free viewpoint image.
- images from a larger number of viewpoints 210 exist, so that it is easy to obtain a sufficient number of images for generating free viewpoint images.
- all available images may be collected, and an image including the position 220 having a high space audience rate may be selected afterwards.
- the spatial audience rating estimated based on the density of the viewing target position from the separately collected viewpoint images is obtained.
- An image containing the high position 220 can be selected afterwards.
- FIG. 5 is a diagram for conceptually explaining another example of image collection according to an embodiment of the present disclosure.
- an image including a position 220 with a high spatial audience rating is collected by the environment setting camera 230.
- the environment-installed camera 230 may be controlled in real time based on the estimation result of the spatial audience rating.
- the environment-installed cameras 230 are arranged so that their directions can be changed within a predetermined range, and the areas in the space SP that can be photographed by the respective environment-installed cameras 230 partially overlap.
- the region R1 where many viewing target positions 120 are concentrated and the spatial audience rating is estimated to be high is high-resolution (zoomed in) by a plurality of environment-installed cameras 230a and 230b. Shooting is in progress.
- a region R2 in which several viewing target positions 120 are scattered and the spatial audience rating is estimated to be medium shooting is performed with high resolution (zoomed in) by a single environment-installed camera 230c. Yes.
- shooting is performed at a low resolution (zoomed out) by the single environment-installed camera 230d.
- an image for generating a free viewpoint image can be obtained stably and with a desired quality.
- the relationship between the spatial audience rating and the operation of the environment-installed camera 230 is not limited to the above example. For example, even in an area where the spatial audience rating is estimated to be low, shooting may be executed by at least two environment-installed cameras 230 if there is a possibility that an event that should be noticed suddenly occurs. Further, when there is no possibility of such an event, it is not necessary to perform shooting for an area that is estimated to have a low spatial audience rating.
- FIGS. 6 and 7 are diagrams for conceptually explaining still another example of image collection according to an embodiment of the present disclosure.
- the same image collection as in the example shown in FIG. 5 is performed by a UAV (Unmanned aerial vehicle) for photographing.
- the UAV 240 flies over the space SP (or inside the space SP) and captures an image in the space SP.
- the region R1 is captured by the UAVs 240a and 240b
- the region R2 is captured by the UAV 240c
- the region R3 is captured by the UAV 240d.
- the UAV 240 can change the shooting direction and zoom in / out in the same manner as the environment-installed camera 230 shown in FIG. 5, for example. Furthermore, the UAV 240 can freely move above or inside the space SP. Therefore, for example, for a position or region where the spatial audience rating is estimated to be high, it is possible to perform shooting while concentrating more UAVs 240. Further, as in the example shown in FIG. 7, for example, the position of the UAV 240 is changed according to the space audience rating, and the arrangement of the UAV 240 is adjusted so that the shooting range 242 of each UAV 240 covers the entire space SP. You can also
- FIG. 8 is a diagram illustrating a configuration example of a system according to an embodiment of the present disclosure.
- the system 10 includes a camera 300, a viewer terminal 400, a server 500, and a user terminal 600.
- the viewer means a subject having a viewpoint existing in the space
- the user is a free viewpoint provided based on the spatial audience rate estimated according to the density of the viewing target positions of the viewer. It means the subject who browses images.
- the camera 300 is mounted on, for example, a terminal (which may be the viewer terminal 400) worn or carried by the viewer. Alternatively, the camera 300 may be a subject that automatically performs shooting, that is, a viewer itself as referred to in this specification.
- the camera 300 includes an imaging unit including at least one imaging device, and acquires an outward camera image 310. Furthermore, when the camera 300 is mounted on a terminal worn by a viewer, the inward camera image 320 may be acquired by an imaging unit including another imaging device.
- the viewer terminal 400 is a terminal device that is used by the viewer together with the camera 300 or is the viewer itself.
- the viewer terminal 400 is realized by, for example, a hardware configuration of an information processing apparatus described later, and viewer position estimation realized by a processor such as a CPU (Central Processing unit) operating according to a program stored in a memory or storage. 410, a line-of-sight estimation unit 420, and a visual target position estimation unit 430.
- the viewer terminal 400 can be realized as a smartphone, a digital camera, a tablet, a personal computer, or the like.
- the viewer position estimation unit 410 estimates the viewer position using a technique such as SLAM based on the outward camera image 310 acquired by the camera 300, for example. At this time, the viewer position estimation unit 410 may use information such as the landmark 550 and the three-dimensional model 570 provided from the server 500. Alternatively, the viewer position estimation unit 410 may estimate the viewer position based on positioning information such as the GNSS data 560 or ticket information. In this case, the outward camera image 310 is not necessarily used for estimating the viewer position.
- the gaze estimation unit 420 estimates the viewer's gaze based on, for example, the inward camera image 320 acquired by the camera 300.
- the viewpoint image that is, the center direction of the outward camera image 310 can be regarded as the line-of-sight direction
- the line-of-sight estimation unit 420 does not need to exist in such a case.
- the line-of-sight estimation unit 420 estimates the line of sight from the center position of the black eye based on the eye image included in the inward-facing camera image 320, or the posture of the eyeball model It is possible to estimate the line of sight from
- the visual target position estimation unit 430 estimates the viewer's visual target position based on the viewer position estimated by the viewer position estimation unit 410 and the visual line estimated by the visual line estimation unit 420.
- the visual target position estimation unit 430 estimates the visual target position based on the three-dimensional model 570 of the space provided from the server 500, for example.
- the visual target position estimation unit 430 uploads information indicating the estimated visual target position, for example, coordinates in the space defined in the three-dimensional model 570, to the server 500.
- the server 500 includes one or a plurality of server devices connected to the camera 300 and / or the viewer terminal 400 via a network.
- the server device is realized by, for example, a hardware configuration of an information processing device described later.
- the server 500 can include a spatial audience rating estimation unit 520 and a point cloud generation unit 540 realized by a processor such as a CPU of the server device operating according to a program stored in a memory or storage.
- a processor such as a CPU of the server device operating according to a program stored in a memory or storage.
- data of the viewing target position data 510, the image data 530, the landmark 550, and the three-dimensional model 570 can be stored.
- the server 500 collects information on the viewing target position provided from the viewer terminal 400 as viewing target position data 510.
- the viewer position estimation, the gaze estimation, and / or the visual target position estimation may be performed by the server.
- the server device processor The information indicating the viewing target position from a plurality of viewpoints existing in the group is collected. This information may be, for example, information indicating the estimated viewing target position itself as in the illustrated example, may be information indicating the estimated viewer position or line of sight, and may be the outward camera image 310. And / or inward camera image 320.
- the spatial audience rating estimation unit 520 estimates the spatial audience rating based on the viewing target position data 510.
- the spatial audience rating is an example of an index that is calculated according to the density of the visual target position and evaluates the position in the space. Therefore, it can be said that the space audience rating estimation unit 520 evaluates the position in the space according to the density of the viewing target position.
- the server 500 acquires the outward camera image 310 uploaded from the camera 300 and stores it as image data 530.
- the point cloud generation unit 540 generates point cloud data for providing a free viewpoint image from the image data 530 according to the spatial audience rating.
- the point cloud data is distributed to the user terminal 600.
- the point cloud generation unit 540 realized in the server 500 edits image data obtained by photographing the space based on the result of evaluating the position in the space according to the density of the viewing target position. It is an example of the function to perform.
- the point cloud generation unit 540 selectively generates point cloud data for providing a free viewpoint image according to a spatial audience rating that is an index according to the density of the viewing target position. Selective generation may include, for example, generating point cloud data at a position or region with a high spatial audience rating and not generating point cloud data at a position or region with a low spatial audience rating.
- the processor of the server device that implements the server 500 further captures the space based on the space audience rating. You may implement
- the function of controlling image capturing by the environment-installed camera 230 and the UAV 240 can be realized by the processor of the server device.
- control of the position in the space where the image is to be captured can be performed by changing the orientation of the environment-installed camera 230, moving the UAV 240, or the like.
- the control of the range to be imaged may be performed by zooming in / out of a camera mounted on the environment-installed camera 230 or the UAV 240.
- User terminal 600 is a terminal device used by a user who browses a free viewpoint image.
- the user terminal 600 may be implemented by, for example, a hardware configuration of an information processing apparatus to be described later, and may include a free viewpoint image generation unit 610 that is implemented by a processor such as a CPU operating according to a program stored in a memory or storage.
- the user terminal 600 can be realized as a television, a recorder, a personal computer, a tablet, a smartphone, or the like.
- the free viewpoint image generation unit 610 generates a free viewpoint image based on the point cloud data provided from the server 500.
- the free viewpoint image can be generated according to the viewpoint designation 620 by the user's operation input.
- a user who views a free viewpoint image is not limited to the viewpoint designation 620, and may be able to designate zoom-in / out or the like by an operation input.
- the generated free viewpoint image may be displayed on a display of the user terminal 600 or may be output to an externally connected display.
- the configuration example of the system according to this embodiment has been described above.
- the described configuration is an example, and various modifications are possible.
- the functions described above as being included in the viewer terminal 400 and the user terminal 600 may be realized by the server 500.
- the server 500 may realize the free viewpoint image generation unit 610 and distribute the free viewpoint image generated according to the viewpoint designation 620 or the like to a display device under the user.
- FIG. 9 is a flowchart illustrating an example of processing according to an embodiment of the present disclosure.
- FIG. 9 shows an example of processing executed in the system 10 described above with reference to FIG.
- a landmark is detected from the outward camera image 310 acquired by the camera 300 (S201).
- a landmark is, for example, an object that is installed in a space and can be a landmark in an image.
- a landmark may not be installed for viewer position estimation, such as a line drawn on a stadium field, a pole for a competition, a theater stage, or a guidance display.
- the landmark may be installed for viewer position estimation.
- the landmarks can be arranged with a texture that allows each landmark to be uniquely identified at a position where the landmarks are easily transferred to images taken from various viewpoints.
- a landmark is detected from the outward camera image 310 (S203).
- viewer position estimation using the landmark is performed (S205).
- the viewer position is estimated using SLAM technology with reference to the coordinates of landmarks set in advance in a three-dimensional model of space.
- viewer position estimation using GNSS data is performed (S207). Note that the viewer position may be estimated based on other positioning information instead of the GNSS data.
- the viewer position estimation based on the positioning information or the like may be preferentially performed.
- line-of-sight estimation using the inward camera image 320 is performed (S209).
- the center direction of the outward camera image 310 may be regarded as the line-of-sight direction.
- the visual target position estimation based on the estimated viewer position and line of sight is performed (S211).
- the information on the viewing target position estimated by the viewer terminal 400 and the outward camera image 310 acquired by the camera 300 are uploaded to the server 500 (S213).
- the processing of S201 to S211 may be executed by the server 500.
- the outward camera image 310 is uploaded from the camera 300 to the server 500. It may just be done.
- the spatial audience rating estimation unit 520 calculates the density of the visual target position in the grid or cluster in the space based on the uploaded information (S215).
- the density of the visual target position corresponds to the spatial audience rating of the grid or cluster.
- the point cloud generation unit 540 determines whether or not the density exceeds the threshold value in the loop processing (S217) for each grid or cluster in which the density of the target position is calculated (S219), and the density If the value exceeds the threshold value, point cloud data for the grid or cluster is generated (S221).
- the point cloud data generated by the above processing is distributed to the user (S223), and a free viewpoint image is generated according to the viewpoint selection (S225) by the user (S227).
- FIG. 10 is a block diagram illustrating a hardware configuration example of the information processing apparatus according to the embodiment of the present disclosure.
- the illustrated information processing apparatus 900 can realize, for example, a viewer terminal, a server apparatus, and / or a user terminal in the above-described embodiment.
- the information processing apparatus 900 includes a CPU (Central Processing unit) 901, a ROM (Read Only Memory) 903, and a RAM (Random Access Memory) 905.
- the information processing apparatus 900 may include a host bus 907, a bridge 909, an external bus 911, an interface 913, an input device 915, an output device 917, a storage device 919, a drive 921, a connection port 923, and a communication device 925.
- the information processing apparatus 900 may include an imaging device 933 and a sensor 935 as necessary.
- the information processing apparatus 900 may include a processing circuit such as a DSP (Digital Signal Processor), an ASIC (Application Specific Integrated Circuit), or an FPGA (Field-Programmable Gate Array) instead of or in addition to the CPU 901.
- DSP Digital Signal Processor
- ASIC Application Specific Integrated Circuit
- FPGA Field-Programmable Gate Array
- the CPU 901 functions as an arithmetic processing device and a control device, and controls all or a part of the operation in the information processing device 900 according to various programs recorded in the ROM 903, the RAM 905, the storage device 919, or the removable recording medium 927.
- the ROM 903 stores programs and calculation parameters used by the CPU 901.
- the RAM 905 primarily stores programs used in the execution of the CPU 901, parameters that change as appropriate during the execution, and the like.
- the CPU 901, the ROM 903, and the RAM 905 are connected to each other by a host bus 907 configured by an internal bus such as a CPU bus. Further, the host bus 907 is connected to an external bus 911 such as a PCI (Peripheral Component Interconnect / Interface) bus via a bridge 909.
- PCI Peripheral Component Interconnect / Interface
- the input device 915 is a device operated by the user, such as a mouse, a keyboard, a touch panel, a button, a switch, and a lever.
- the input device 915 may be, for example, a remote control device that uses infrared rays or other radio waves, or may be an external connection device 929 such as a mobile phone that supports the operation of the information processing device 900.
- the input device 915 includes an input control circuit that generates an input signal based on information input by the user and outputs the input signal to the CPU 901. The user operates the input device 915 to input various data and instruct processing operations to the information processing device 900.
- the output device 917 is configured by a device capable of notifying the acquired information to the user using a sense such as vision, hearing, or touch.
- the output device 917 can be, for example, a display device such as an LCD (Liquid Crystal Display) or an organic EL (Electro-Luminescence) display, an audio output device such as a speaker or headphones, or a vibrator.
- the output device 917 outputs the result obtained by the processing of the information processing device 900 as video such as text or image, sound such as sound or sound, or vibration.
- the storage device 919 is a data storage device configured as an example of a storage unit of the information processing device 900.
- the storage device 919 includes, for example, a magnetic storage device such as an HDD (Hard Disk Drive), a semiconductor storage device, an optical storage device, or a magneto-optical storage device.
- the storage device 919 stores, for example, programs executed by the CPU 901 and various data, and various data acquired from the outside.
- the drive 921 is a reader / writer for a removable recording medium 927 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory, and is built in or externally attached to the information processing apparatus 900.
- the drive 921 reads information recorded on the attached removable recording medium 927 and outputs the information to the RAM 905.
- the drive 921 writes a record in the attached removable recording medium 927.
- the connection port 923 is a port for connecting a device to the information processing apparatus 900.
- the connection port 923 can be, for example, a USB (Universal Serial Bus) port, an IEEE 1394 port, a SCSI (Small Computer System Interface) port, or the like.
- the connection port 923 may be an RS-232C port, an optical audio terminal, an HDMI (registered trademark) (High-Definition Multimedia Interface) port, or the like.
- the communication device 925 is a communication interface configured with, for example, a communication device for connecting to the communication network 931.
- the communication device 925 can be, for example, a communication card for LAN (Local Area Network), Bluetooth (registered trademark), Wi-Fi, or WUSB (Wireless USB).
- the communication device 925 may be a router for optical communication, a router for ADSL (Asymmetric Digital Subscriber Line), or a modem for various communication.
- the communication device 925 transmits and receives signals and the like using a predetermined protocol such as TCP / IP with the Internet and other communication devices, for example.
- the communication network 931 connected to the communication device 925 is a network connected by wire or wireless, and may include, for example, the Internet, a home LAN, infrared communication, radio wave communication, satellite communication, or the like.
- the imaging device 933 uses various members such as an imaging element such as a CMOS (Complementary Metal Oxide Semiconductor) or a CCD (Charge Coupled Device), and a lens for controlling the formation of a subject image on the imaging element. It is an apparatus that images a real space and generates a captured image.
- the imaging device 933 may capture a still image or may capture a moving image.
- the sensor 935 is various sensors such as an acceleration sensor, an angular velocity sensor, a geomagnetic sensor, an illuminance sensor, a temperature sensor, an atmospheric pressure sensor, or a sound sensor (microphone).
- the sensor 935 acquires information about the state of the information processing apparatus 900 itself, such as the posture of the information processing apparatus 900, and information about the surrounding environment of the information processing apparatus 900, such as brightness and noise around the information processing apparatus 900, for example. To do.
- the sensor 935 may include a GNSS (Global Navigation Satellite Systems) receiver that receives a signal from an artificial satellite and measures the latitude, longitude, and altitude of the apparatus.
- GNSS Global Navigation Satellite Systems
- Each component described above may be configured using a general-purpose member, or may be configured by hardware specialized for the function of each component. Such a configuration can be appropriately changed according to the technical level at the time of implementation.
- an information processing apparatus for example, an information processing apparatus, a system, an information processing method executed by the information processing apparatus or system, a program for causing the information processing apparatus to function, and a program are recorded. It may include tangible media that is not temporary.
- the evaluation result (spatial audience rating) of the position in the space according to the density of the viewing target position is used for generating point cloud data for providing a free viewpoint image.
- the embodiment of the present disclosure is not limited to such an example.
- a position with a high spatial audience rating is estimated to be a position where a user who is viewing a captured image is highly interested, and therefore various processes for recording an image according to such a user's interest are performed. It may be carried out according to the spatial audience rating.
- an image taken at a position with a high spatial audience rate is stored at a high image quality (large image size and / or high bit rate), and an image taken at a position other than that is low image quality (small).
- a configuration may be employed in which resources for recording data are effectively used by recording at a small image size and / or low bit rate.
- a sufficient image cannot be acquired to generate a free viewpoint image, or the user viewing the image is free because the spatial audience rating is low.
- a position where data for generating a viewpoint image for example, point cloud data
- a plurality of viewpoint images obtained by capturing the position are simply switched and viewed using available images. May be possible.
- the following configurations also belong to the technical scope of the present disclosure.
- An information processing apparatus comprising: a processor that realizes a function of evaluating a position in the space according to a density of the visual target position.
- the processor A function of acquiring image data obtained by photographing the space;
- the information processing apparatus according to (1) further realizing a function of executing editing of the image data based on a result of evaluating a position in the space.
- the information processing apparatus according to (2), wherein the editing of the image data includes generation of data for providing a free viewpoint image.
- the information processing apparatus includes generating data for selectively providing the free viewpoint image according to a density of the viewing target position.
- the editing of the image data includes changing the quality of the image data recorded according to the density of the viewing target position, according to any one of (2) to (4).
- Information processing device (6) The information processing according to any one of (2) to (5), wherein the editing of the image data includes selecting the image data to be recorded according to a density of the viewing target position. apparatus. (7) The information processing apparatus according to any one of (2) to (6), wherein the viewing target position is estimated based on the image data.
- the processor A function of acquiring image data obtained by photographing the space; The information processing apparatus according to any one of (1) to (7), further realizing a function of executing control of acquisition of the image data based on a result of evaluating a position in the space. (9) The information processing apparatus according to (8), wherein the acquisition control of the image data includes control of a position to be imaged. (10) The information processing apparatus according to (8) or (9), wherein the acquisition control of the image data includes control of a range to be imaged. (11) The information processing apparatus according to (1), wherein the viewing target position is estimated based on image data obtained by photographing the space from the plurality of viewpoints.
- the processor The information processing apparatus according to (11), further realizing a function of estimating the viewing target position based on information indicating the viewing target position.
- the information indicating the viewing target position further includes information indicating a line of sight from the plurality of viewpoints.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
- Studio Devices (AREA)
- Processing Or Creating Images (AREA)
Abstract
Description
1.空間視聴率の推定
2.画像の収集および自由視点画像の生成
3.システム構成例
4.処理フローの例
5.ハードウェア構成
6.補足
本開示の一実施形態では、空間内に存在する複数の視点からの視対象位置(visual target position)の密度に基づいて空間視聴率(spatial rating)が推定される。空間視聴率は、空間内のそれぞれの位置がどの程度の注目を集めているかを示す。例えば競技場のような空間を想定する場合、複数の視点は、例えば競技場にいる観客の視点を含みうる。また、複数の視点は、例えば競技場に設置されたカメラの視点を含んでもよい。これらの視点からの視対象位置は、例えば視線上の位置である。観客やカメラからの視線が集中している位置、つまり空間視聴率が高い位置は、競技場の中で特に注目されていると推定される。本実施形態では、このような空間視聴率に基づいて、自由視点画像を提供するための3次元点群(ポイントクラウド)データを生成する。なお、本明細書において、画像は動画像(映像)および静止画像を含む。
図4は、本開示の一実施形態における画像の収集の例について概念的に説明するための図である。図4に示されるように、本実施形態では、空間SP(図示された例では競技場)内に配置された視点210のそれぞれにおいて撮影された画像のうち、空間視聴率に基づいて選定された位置220を含む画像が、自由視点画像の生成のために収集される。視点210は、例えばウェアラブルカメラを装用したり、モバイルカメラ(通常のデジタルカメラまたはスマートフォンなどに搭載されるカメラなど)で撮影を実行したりしている観客(プロフェッショナルのカメラマンであってもよい)の位置でありうる。また、視点210は、中継などのために固定設置されたカメラの位置を含んでもよい。
図8は、本開示の一実施形態に係るシステムの構成例を示す図である。図8を参照すると、システム10は、カメラ300と、視聴者端末400と、サーバ500と、ユーザ端末600とを含む。なお、本明細書において、視聴者は空間内に存在する視点を有する主体を意味し、ユーザは視聴者の視対象位置の密度に応じて推定された空間視聴率に基づいて提供される自由視点画像を閲覧する主体を意味する。以下、それぞれの
図9は、本開示の一実施形態における処理の例を示すフローチャートである。図9では、上記で図8を参照して説明したシステム10において実行される処理の例が示されている。
次に、図10を参照して、本開示の実施形態に係る情報処理装置のハードウェア構成について説明する。図10は、本開示の実施形態に係る情報処理装置のハードウェア構成例を示すブロック図である。図示された情報処理装置900は、例えば、上記の実施形態における視聴者端末、サーバ装置、および/またはユーザ端末を実現しうる。
本開示の実施形態は、例えば、上記で説明したような情報処理装置、システム、情報処理装置またはシステムで実行される情報処理方法、情報処理装置を機能させるためのプログラム、およびプログラムが記録された一時的でない有形の媒体を含みうる。
(1)空間内に存在する複数の視点からの視対象位置を示す情報を収集する機能と、
前記視対象位置の密度に応じて前記空間内の位置を評価する機能と
を実現するプロセッサを備える情報処理装置。
(2)前記プロセッサは、
前記空間内を撮影することによって得られる画像データを取得する機能と、
前記空間内の位置を評価した結果に基づいて、前記画像データの編集を実行する機能と
をさらに実現する、前記(1)に記載の情報処理装置。
(3)前記画像データの編集は、自由視点画像を提供するためのデータの生成を含む、前記(2)に記載の情報処理装置。
(4)前記画像データの編集は、前記視対象位置の密度に応じて選択的に前記自由視点画像を提供するためのデータを生成することを含む、前記(3)に記載の情報処理装置。
(5)前記画像データの編集は、前記視対象位置の密度に応じて記録される前記画像データの品質を変化させることを含む、前記(2)~(4)のいずれか1項に記載の情報処理装置。
(6)前記画像データの編集は、前記視対象位置の密度に応じて記録される前記画像データを選別することを含む、前記(2)~(5)のいずれか1項に記載の情報処理装置。
(7)前記視対象位置は、前記画像データに基づいて推定される、前記(2)~(6)のいずれか1項に記載の情報処理装置。
(8)前記プロセッサは、
前記空間内を撮影することによって得られる画像データを取得する機能と、
前記空間内の位置を評価した結果に基づいて、前記画像データの取得の制御を実行する機能と
をさらに実現する、前記(1)~(7)のいずれか1項に記載の情報処理装置。
(9)前記画像データの取得の制御は、前記撮影の対象になる位置の制御を含む、前記(8)に記載の情報処理装置。
(10)前記画像データの取得の制御は、前記撮影の対象になる範囲の制御を含む、前記(8)または(9)に記載の情報処理装置。
(11)前記視対象位置は、前記空間内を前記複数の視点から撮影することによって得られる画像データに基づいて推定される、前記(1)に記載の情報処理装置。
(12)前記プロセッサは、
前記視対象位置を示す情報に基づいて前記視対象位置を推定する機能
をさらに実現する、前記(11)に記載の情報処理装置。
(13)前記視対象位置を示す情報は、前記画像データを含む、前記(12)に記載の情報処理装置。
(14)前記視対象位置を示す情報は、前記画像データに基づいて推定された前記複数の視点の位置を示す情報を含む、前記(12)に記載の情報処理装置。
(15)前記視対象位置を示す情報は、前記複数の視点からの視線を示す情報をさらに含む、前記(14)に記載の情報処理装置。
(16)前記視対象位置の密度は、前記空間内に設定されるグリッドごとに評価される、前記(1)~(15)のいずれか1項に記載の情報処理装置。
(17)前記視対象位置の密度は、互いに近接する前記視対象位置を含むクラスタごとに評価される、前記(1)~(15)のいずれか1項に記載の情報処理装置。
(18)空間内に存在する複数の視点からの視対象位置を示す情報を収集することと、
プロセッサが、前記視対象位置の密度に応じて前記空間内の位置を評価することと
を含む情報処理方法。
(19)空間内に存在する複数の視点からの視対象位置を示す情報を収集する機能と、
前記視対象位置の密度に応じて前記空間内の位置を評価する機能と
をプロセッサに実現させるためのプログラム。
300 カメラ
400 視聴者端末
500 サーバ
510 視対象位置データ
520 空間視聴率推定部
530 画像データ
540 ポイントクラウド生成部
600 ユーザ端末
610 自由視点画像生成部
Claims (19)
- 空間内に存在する複数の視点からの視対象位置を示す情報を収集する機能と、
前記視対象位置の密度に応じて前記空間内の位置を評価する機能と
を実現するプロセッサを備える情報処理装置。 - 前記プロセッサは、
前記空間内を撮影することによって得られる画像データを取得する機能と、
前記空間内の位置を評価した結果に基づいて、前記画像データの編集を実行する機能と
をさらに実現する、請求項1に記載の情報処理装置。 - 前記画像データの編集は、自由視点画像を提供するためのデータの生成を含む、請求項2に記載の情報処理装置。
- 前記画像データの編集は、前記視対象位置の密度に応じて選択的に前記自由視点画像を提供するためのデータを生成することを含む、請求項3に記載の情報処理装置。
- 前記画像データの編集は、前記視対象位置の密度に応じて記録される前記画像データの品質を変化させることを含む、請求項2に記載の情報処理装置。
- 前記画像データの編集は、前記視対象位置の密度に応じて記録される前記画像データを選別することを含む、請求項2に記載の情報処理装置。
- 前記視対象位置は、前記画像データに基づいて推定される、請求項2に記載の情報処理装置。
- 前記プロセッサは、
前記空間内を撮影することによって得られる画像データを取得する機能と、
前記空間内の位置を評価した結果に基づいて、前記画像データの取得の制御を実行する機能と
をさらに実現する、請求項1に記載の情報処理装置。 - 前記画像データの取得の制御は、前記撮影の対象になる位置の制御を含む、請求項8に記載の情報処理装置。
- 前記画像データの取得の制御は、前記撮影の対象になる範囲の制御を含む、請求項8に記載の情報処理装置。
- 前記視対象位置は、前記空間内を前記複数の視点から撮影することによって得られる画像データに基づいて推定される、請求項1に記載の情報処理装置。
- 前記プロセッサは、
前記視対象位置を示す情報に基づいて前記視対象位置を推定する機能
をさらに実現する、請求項11に記載の情報処理装置。 - 前記視対象位置を示す情報は、前記画像データを含む、請求項12に記載の情報処理装置。
- 前記視対象位置を示す情報は、前記画像データに基づいて推定された前記複数の視点の位置を示す情報を含む、請求項12に記載の情報処理装置。
- 前記視対象位置を示す情報は、前記複数の視点からの視線を示す情報をさらに含む、請求項14に記載の情報処理装置。
- 前記視対象位置の密度は、前記空間内に設定されるグリッドごとに評価される、請求項1に記載の情報処理装置。
- 前記視対象位置の密度は、互いに近接する前記視対象位置を含むクラスタごとに評価される、請求項1に記載の情報処理装置。
- 空間内に存在する複数の視点からの視対象位置を示す情報を収集することと、
プロセッサが、前記視対象位置の密度に応じて前記空間内の位置を評価することと
を含む情報処理方法。 - 空間内に存在する複数の視点からの視対象位置を示す情報を収集する機能と、
前記視対象位置の密度に応じて前記空間内の位置を評価する機能と
をプロセッサに実現させるためのプログラム。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/520,073 US10190869B2 (en) | 2014-12-04 | 2015-09-25 | Information processing device and information processing method |
EP15865510.0A EP3229459B1 (en) | 2014-12-04 | 2015-09-25 | Information processing device, information processing method and program |
JP2016562332A JP6702196B2 (ja) | 2014-12-04 | 2015-09-25 | 情報処理装置、情報処理方法およびプログラム |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2014245901 | 2014-12-04 | ||
JP2014-245901 | 2014-12-04 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2016088437A1 true WO2016088437A1 (ja) | 2016-06-09 |
Family
ID=56091394
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2015/077216 WO2016088437A1 (ja) | 2014-12-04 | 2015-09-25 | 情報処理装置、情報処理方法およびプログラム |
Country Status (4)
Country | Link |
---|---|
US (1) | US10190869B2 (ja) |
EP (1) | EP3229459B1 (ja) |
JP (1) | JP6702196B2 (ja) |
WO (1) | WO2016088437A1 (ja) |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018088037A1 (ja) * | 2016-11-14 | 2018-05-17 | ソニー株式会社 | 可動型撮像装置の制御装置、可動型撮像装置の制御方法及びプログラム |
JP6415675B1 (ja) * | 2017-12-04 | 2018-10-31 | キヤノン株式会社 | 生成装置、生成方法及びプログラム |
WO2019078248A1 (ja) * | 2017-10-17 | 2019-04-25 | 株式会社ニコン | 制御装置、制御システム、および制御プログラム |
JP2019075740A (ja) * | 2017-10-18 | 2019-05-16 | キヤノン株式会社 | 画像処理システム、画像処理装置、画像伝送方法、及び、プログラム |
WO2019092790A1 (ja) * | 2017-11-07 | 2019-05-16 | ヤマハ株式会社 | 撮像システム、制御装置、撮像装置、撮像装置の制御方法およびプログラム |
JP2019193209A (ja) * | 2018-04-27 | 2019-10-31 | エスゼット ディージェイアイ テクノロジー カンパニー リミテッドSz Dji Technology Co.,Ltd | 制御装置及び撮影方法 |
JP2019537135A (ja) * | 2016-11-04 | 2019-12-19 | ディープマインド テクノロジーズ リミテッド | ニューラルネットワークを使用したシーンの理解および生成 |
WO2020256120A1 (ja) * | 2019-06-21 | 2020-12-24 | キヤノン株式会社 | 画像提供システム及びその制御方法 |
JP2021002825A (ja) * | 2019-06-21 | 2021-01-07 | キヤノン株式会社 | 画像提供システム、画像提供装置及びその制御方法 |
JP2021010088A (ja) * | 2019-06-28 | 2021-01-28 | キヤノン株式会社 | 画像提供システム、画像提供装置、その制御方法、及びプログラム |
WO2021131991A1 (ja) | 2019-12-23 | 2021-07-01 | ソニーグループ株式会社 | 映像生成装置、映像生成方法および映像生成プログラム |
WO2022018836A1 (ja) * | 2020-07-21 | 2022-01-27 | ソニーグループ株式会社 | 情報処理装置、情報処理方法及びプログラム |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106454069B (zh) * | 2016-08-31 | 2019-11-08 | 歌尔股份有限公司 | 一种控制无人机拍摄的方法、装置和可穿戴设备 |
JP6434947B2 (ja) * | 2016-09-30 | 2018-12-05 | キヤノン株式会社 | 撮像システム、画像処理装置、画像処理方法、及びプログラム |
US20180232574A1 (en) * | 2017-02-14 | 2018-08-16 | The Board Of Trustees Of The Leland Junior University | Systems and Methods for Transmitting and Receiving Data Using Machine Learning Classification |
JP6433559B1 (ja) * | 2017-09-19 | 2018-12-05 | キヤノン株式会社 | 提供装置および提供方法、プログラム |
JP7193938B2 (ja) * | 2018-07-04 | 2022-12-21 | キヤノン株式会社 | 情報処理装置及びその制御方法、プログラム |
JP2020136776A (ja) * | 2019-02-14 | 2020-08-31 | キヤノン株式会社 | 画像生成システムおよび画像生成方法、制御装置および制御方法 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005100367A (ja) * | 2003-09-02 | 2005-04-14 | Fuji Photo Film Co Ltd | 画像生成装置、画像生成方法、及び画像生成プログラム |
JP2005109757A (ja) * | 2003-09-29 | 2005-04-21 | Fuji Photo Film Co Ltd | 画像撮像装置、画像処理装置、画像撮像方法、及びプログラム |
JP2007235399A (ja) * | 2006-02-28 | 2007-09-13 | Matsushita Electric Ind Co Ltd | 自動撮影装置 |
JP2010081260A (ja) * | 2008-09-25 | 2010-04-08 | Casio Computer Co Ltd | 撮像装置及びそのプログラム |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003179800A (ja) | 2001-12-13 | 2003-06-27 | Sony Corp | 多視点画像生成装置、画像処理装置、および方法、並びにコンピュータ・プログラム |
-
2015
- 2015-09-25 JP JP2016562332A patent/JP6702196B2/ja not_active Expired - Fee Related
- 2015-09-25 WO PCT/JP2015/077216 patent/WO2016088437A1/ja active Application Filing
- 2015-09-25 EP EP15865510.0A patent/EP3229459B1/en active Active
- 2015-09-25 US US15/520,073 patent/US10190869B2/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005100367A (ja) * | 2003-09-02 | 2005-04-14 | Fuji Photo Film Co Ltd | 画像生成装置、画像生成方法、及び画像生成プログラム |
JP2005109757A (ja) * | 2003-09-29 | 2005-04-21 | Fuji Photo Film Co Ltd | 画像撮像装置、画像処理装置、画像撮像方法、及びプログラム |
JP2007235399A (ja) * | 2006-02-28 | 2007-09-13 | Matsushita Electric Ind Co Ltd | 自動撮影装置 |
JP2010081260A (ja) * | 2008-09-25 | 2010-04-08 | Casio Computer Co Ltd | 撮像装置及びそのプログラム |
Non-Patent Citations (1)
Title |
---|
See also references of EP3229459A4 * |
Cited By (29)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2019537135A (ja) * | 2016-11-04 | 2019-12-19 | ディープマインド テクノロジーズ リミテッド | ニューラルネットワークを使用したシーンの理解および生成 |
JP7295906B2 (ja) | 2016-11-04 | 2023-06-21 | ディープマインド テクノロジーズ リミテッド | ニューラルネットワークを使用したシーンの理解および生成 |
US11587344B2 (en) | 2016-11-04 | 2023-02-21 | Deepmind Technologies Limited | Scene understanding and generation using neural networks |
JP7161992B2 (ja) | 2016-11-04 | 2022-10-27 | ディープマインド テクノロジーズ リミテッド | ニューラルネットワークを使用したシーンの理解および生成 |
JP2021152948A (ja) * | 2016-11-04 | 2021-09-30 | ディープマインド テクノロジーズ リミテッド | ニューラルネットワークを使用したシーンの理解および生成 |
JPWO2018088037A1 (ja) * | 2016-11-14 | 2019-09-26 | ソニー株式会社 | 可動型撮像装置の制御装置、可動型撮像装置の制御方法及びプログラム |
WO2018088037A1 (ja) * | 2016-11-14 | 2018-05-17 | ソニー株式会社 | 可動型撮像装置の制御装置、可動型撮像装置の制御方法及びプログラム |
JP7059937B2 (ja) | 2016-11-14 | 2022-04-26 | ソニーグループ株式会社 | 可動型撮像装置の制御装置、可動型撮像装置の制御方法及びプログラム |
US10924691B2 (en) | 2016-11-14 | 2021-02-16 | Sony Corporation | Control device of movable type imaging device and control method of movable type imaging device |
WO2019078248A1 (ja) * | 2017-10-17 | 2019-04-25 | 株式会社ニコン | 制御装置、制御システム、および制御プログラム |
JPWO2019078248A1 (ja) * | 2017-10-17 | 2020-11-19 | 株式会社ニコン | 制御装置、制御システム、および制御プログラム |
JP7375542B2 (ja) | 2017-10-17 | 2023-11-08 | 株式会社ニコン | 制御装置、制御システム、および制御プログラム |
JP7104504B2 (ja) | 2017-10-18 | 2022-07-21 | キヤノン株式会社 | 画像処理システム、画像処理装置、画像伝送方法、及び、プログラム |
JP2019075740A (ja) * | 2017-10-18 | 2019-05-16 | キヤノン株式会社 | 画像処理システム、画像処理装置、画像伝送方法、及び、プログラム |
WO2019092790A1 (ja) * | 2017-11-07 | 2019-05-16 | ヤマハ株式会社 | 撮像システム、制御装置、撮像装置、撮像装置の制御方法およびプログラム |
WO2019111817A1 (ja) * | 2017-12-04 | 2019-06-13 | キヤノン株式会社 | 生成装置、生成方法及びプログラム |
US11012679B2 (en) | 2017-12-04 | 2021-05-18 | Canon Kabushiki Kaisha | Generating apparatus, generating method, and storage medium |
JP6415675B1 (ja) * | 2017-12-04 | 2018-10-31 | キヤノン株式会社 | 生成装置、生成方法及びプログラム |
JP2019101795A (ja) * | 2017-12-04 | 2019-06-24 | キヤノン株式会社 | 生成装置、生成方法及びプログラム |
JP2019193209A (ja) * | 2018-04-27 | 2019-10-31 | エスゼット ディージェイアイ テクノロジー カンパニー リミテッドSz Dji Technology Co.,Ltd | 制御装置及び撮影方法 |
WO2020256120A1 (ja) * | 2019-06-21 | 2020-12-24 | キヤノン株式会社 | 画像提供システム及びその制御方法 |
JP2021002825A (ja) * | 2019-06-21 | 2021-01-07 | キヤノン株式会社 | 画像提供システム、画像提供装置及びその制御方法 |
JP7504634B2 (ja) | 2019-06-21 | 2024-06-24 | キヤノン株式会社 | 画像提供システム、その制御方法、プログラムおよび記憶媒体 |
US12035066B2 (en) | 2019-06-21 | 2024-07-09 | Canon Kabushiki Kaisha | Image providing system and method for controlling same |
JP2021010088A (ja) * | 2019-06-28 | 2021-01-28 | キヤノン株式会社 | 画像提供システム、画像提供装置、その制御方法、及びプログラム |
JP7305461B2 (ja) | 2019-06-28 | 2023-07-10 | キヤノン株式会社 | 画像提供システム、画像提供装置、その制御方法、及びプログラム |
WO2021131991A1 (ja) | 2019-12-23 | 2021-07-01 | ソニーグループ株式会社 | 映像生成装置、映像生成方法および映像生成プログラム |
WO2022018836A1 (ja) * | 2020-07-21 | 2022-01-27 | ソニーグループ株式会社 | 情報処理装置、情報処理方法及びプログラム |
US12106439B2 (en) | 2020-07-21 | 2024-10-01 | Sony Group Corporation | Device and associated methodology for suppressing interaction delay of interacting with a field of view of a mobile terminal on a different display |
Also Published As
Publication number | Publication date |
---|---|
EP3229459A1 (en) | 2017-10-11 |
EP3229459A4 (en) | 2018-08-15 |
US20170322017A1 (en) | 2017-11-09 |
JPWO2016088437A1 (ja) | 2017-09-14 |
EP3229459B1 (en) | 2022-08-24 |
US10190869B2 (en) | 2019-01-29 |
JP6702196B2 (ja) | 2020-05-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6702196B2 (ja) | 情報処理装置、情報処理方法およびプログラム | |
US10679676B2 (en) | Automatic generation of video and directional audio from spherical content | |
US9729788B2 (en) | Image generation apparatus and image generation method | |
RU2683262C2 (ru) | Устройство обработки информации, способ обработки информации и программа | |
US9600723B1 (en) | Systems and methods for attention localization using a first-person point-of-view device | |
JP6621063B2 (ja) | カメラ選択方法及び映像配信システム | |
EP2779620B1 (en) | Image generation device, and image generation method | |
US9894272B2 (en) | Image generation apparatus and image generation method | |
US10284776B2 (en) | Image generation apparatus and image generation method | |
JP6897728B2 (ja) | 画像処理装置、画像処理方法およびプログラム | |
EP2998935B1 (en) | Image processing device, image processing method, and program | |
JP6332037B2 (ja) | 画像処理装置および方法、並びにプログラム | |
US11195295B2 (en) | Control system, method of performing analysis and storage medium | |
JP6359704B2 (ja) | イベントに関連付けられる情報を人に供給するための方法 | |
CN112839165A (zh) | 人脸跟踪摄像的实现方法、装置、计算机设备和存储介质 | |
WO2021220892A1 (ja) | 画像処理装置、画像処理方法、及びプログラム | |
JP2015111371A (ja) | 情報処理装置、情報処理方法およびプログラム | |
Matsumoto et al. | Image processing device and method to obtain a 360 image without remapping | |
JP2015056727A (ja) | 装置、方法、プログラム | |
Nishimura et al. | Visualization of multiple people's lifelog: collecting" Ant's-eye view" to generate" Bird's-eye view" | |
CN109918466A (zh) | 一种基于视频分析的实时地图信息全局呈现方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15865510 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2016562332 Country of ref document: JP Kind code of ref document: A |
|
REEP | Request for entry into the european phase |
Ref document number: 2015865510 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2015865510 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15520073 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |