WO2014171474A1 - 情報処理装置、コンテンツ要求方法およびコンピュータプログラム - Google Patents
情報処理装置、コンテンツ要求方法およびコンピュータプログラム Download PDFInfo
- Publication number
- WO2014171474A1 WO2014171474A1 PCT/JP2014/060804 JP2014060804W WO2014171474A1 WO 2014171474 A1 WO2014171474 A1 WO 2014171474A1 JP 2014060804 W JP2014060804 W JP 2014060804W WO 2014171474 A1 WO2014171474 A1 WO 2014171474A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- unit
- data
- image
- content
- encoding
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 56
- 230000010365 information processing Effects 0.000 title claims abstract description 10
- 238000004590 computer program Methods 0.000 title claims description 9
- 238000004891 communication Methods 0.000 claims abstract description 72
- 238000003860 storage Methods 0.000 claims abstract description 62
- 238000012423 maintenance Methods 0.000 abstract description 6
- 206010000060 Abdominal distension Diseases 0.000 abstract 1
- 208000024330 bloating Diseases 0.000 abstract 1
- 238000012545 processing Methods 0.000 description 106
- 230000006870 function Effects 0.000 description 44
- 230000015654 memory Effects 0.000 description 40
- 239000000872 buffer Substances 0.000 description 38
- 230000005540 biological transmission Effects 0.000 description 28
- 238000010586 diagram Methods 0.000 description 24
- 238000003384 imaging method Methods 0.000 description 24
- 230000008569 process Effects 0.000 description 22
- 230000009467 reduction Effects 0.000 description 22
- 238000006243 chemical reaction Methods 0.000 description 20
- 238000013500 data storage Methods 0.000 description 17
- 230000003287 optical effect Effects 0.000 description 16
- 230000005236 sound signal Effects 0.000 description 15
- 238000009826 distribution Methods 0.000 description 13
- 238000005516 engineering process Methods 0.000 description 9
- 238000013507 mapping Methods 0.000 description 9
- 230000004044 response Effects 0.000 description 9
- 230000003321 amplification Effects 0.000 description 7
- 238000003199 nucleic acid amplification method Methods 0.000 description 7
- 239000004065 semiconductor Substances 0.000 description 7
- 239000000284 extract Substances 0.000 description 6
- 230000002123 temporal effect Effects 0.000 description 5
- 230000001413 cellular effect Effects 0.000 description 4
- 238000007726 management method Methods 0.000 description 4
- 238000012544 monitoring process Methods 0.000 description 4
- 230000006835 compression Effects 0.000 description 3
- 238000007906 compression Methods 0.000 description 3
- 238000012937 correction Methods 0.000 description 3
- 238000001514 detection method Methods 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 230000008859 change Effects 0.000 description 2
- 239000000470 constituent Substances 0.000 description 2
- 238000003780 insertion Methods 0.000 description 2
- 230000037431 insertion Effects 0.000 description 2
- 230000010354 integration Effects 0.000 description 2
- 239000007787 solid Substances 0.000 description 2
- 230000001133 acceleration Effects 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- 230000003044 adaptive effect Effects 0.000 description 1
- 239000003990 capacitor Substances 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 230000001427 coherent effect Effects 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000007306 functionalization reaction Methods 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 210000003127 knee Anatomy 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- AWSBQWZZLBPUQH-UHFFFAOYSA-N mdat Chemical compound C1=C2CC(N)CCC2=CC2=C1OCO2 AWSBQWZZLBPUQH-UHFFFAOYSA-N 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 238000007639 printing Methods 0.000 description 1
- 230000001172 regenerating effect Effects 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 230000002194 synthesizing effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
- H04N21/858—Linking data to content, e.g. by linking an URL to a video object, by creating a hotspot
- H04N21/8586—Linking data to content, e.g. by linking an URL to a video object, by creating a hotspot by using a URL
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/21—Server components or server architectures
- H04N21/218—Source of audio or video content, e.g. local disk arrays
- H04N21/2187—Live feed
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/2343—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
- H04N21/23439—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements for generating different versions
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/25—Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
- H04N21/262—Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists
- H04N21/26258—Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists for generating a list of items to be played back in a given order, e.g. playlist, or scheduling item distribution according to such list
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/25—Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
- H04N21/266—Channel or content management, e.g. generation and management of keys and entitlement messages in a conditional access system, merging a VOD unicast channel into a multicast channel
- H04N21/2662—Controlling the complexity of the video stream, e.g. by scaling the resolution or bitrate of the video stream based on the client capabilities
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/437—Interfacing the upstream path of the transmission network, e.g. for transmitting client requests to a VOD server
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/4402—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
- H04N21/44029—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display for generating different versions
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/61—Network physical structure; Signal processing
- H04N21/6106—Network physical structure; Signal processing specially adapted to the downstream path of the transmission network
- H04N21/6125—Network physical structure; Signal processing specially adapted to the downstream path of the transmission network involving transmission via Internet
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/65—Transmission of management data between client and server
- H04N21/658—Transmission by the client directed to the server
- H04N21/6587—Control parameters, e.g. trick play commands, viewpoint selection
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/845—Structuring of content, e.g. decomposing content into time segments
- H04N21/8456—Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
Definitions
- This disclosure relates to an information processing apparatus, a content request method, and a computer program.
- HTTP HyperText Transfer Protocol
- MP4 ISO / IEC-14296-12, 14
- MP4 can be used not only as a storage format but also as a transmission format such as download or streaming.
- ABS Adaptive BitStreaming
- the ABS technology stores a plurality of encoded data in which the same content is expressed at different bit rates in a content server, and a client selects one of the plurality of encoded data according to a network band. It is a technology to reproduce.
- Non-Patent Document 2 a technique for performing remote control of a network camera from the client side is proposed as in Non-Patent Document 2 below.
- Non-Patent Document 3 a parameter is added on the server side, and the server stores the parameter, and performs redirection that appropriately changes the server of the delivery destination. Technology has also been proposed.
- MPEG-DASH Dynamic-Adaptive-Streaming-over-HTTP
- URL http://mpeg.chiariglione.org/standards/mpeg-dash/media-presentation-description-and-segment-formats/text-isoiec-23009-12012-dam -1)
- m28017 DASH Device / Server Specific Representation in MPD (CE-URLPARAM), MPEG # 103, Geneva, 2013 m28354 Core Experiment on Parameters insertion in media segment URL, MPEG # 103, Geneva, 2013
- the information described in the meta information becomes larger as the control becomes finer. End up. If the information described in the meta information is enlarged, the ease of maintenance is also deteriorated.
- the present disclosure provides a new and improved information processing apparatus, content request method, and computer capable of performing maintenance while avoiding the enlargement of meta information when a device is remotely controlled from a client using ABS technology. Provide a program.
- a request to the server for accessing each of a plurality of sub-segments constituting each encoded data obtained by encoding the same content at different bit rates or described in the MPD Based on the predetermined definition stored in the storage unit at the address defined in the access information, and a storage unit that stores a predetermined definition for adding a parameter to the access information for accessing any element acquired by A communication unit that specifies and accesses a parameter, and the information defined by the access information is an address for controlling a device that captures the content.
- an address defined by the access information is an address for controlling an apparatus that captures the content.
- the computer accesses each of a plurality of sub-segments constituting each encoded data obtained by encoding the same content at different bit rates, or is described in the MPD. Storing a predetermined definition for adding a parameter to access information for accessing any element acquired by a request to the server, and a predetermined storage stored in the storage unit at an address defined by the access information. There is provided a computer program in which access is performed by specifying a parameter based on the definition, and the address defined in the access information is an address for controlling a device that captures the content.
- a new and improved information processing apparatus capable of performing maintenance while avoiding the enlargement of meta information when remotely controlling a device from a client using the ABS technology.
- a content requesting method and a computer program can be provided.
- FIG. 10 is a sequence diagram illustrating an operation example of the content reproduction system according to the embodiment of the present disclosure.
- FIG. 10 is a sequence diagram illustrating an operation example of the content reproduction system according to the embodiment of the present disclosure.
- It is a block diagram which shows the structural example of the hardware of a computer.
- It is a figure which shows the example of a multiview image encoding system.
- It is a figure which shows the structural example of the multiview image coding apparatus to which this indication is applied.
- It is a figure which shows the example of a hierarchy image coding system.
- It is a figure explaining the example of spatial scalable encoding.
- It is a figure explaining the example of temporal scalable encoding.
- 2 illustrates an example of a schematic configuration of a video set to which the present disclosure is applied. 2 illustrates an example of a schematic configuration of a video processor to which the present disclosure is applied.
- a plurality of constituent elements having substantially the same functional configuration may be distinguished by adding different alphabets after the same reference numeral.
- a plurality of configurations having substantially the same functional configuration or logical significance are distinguished as content playback devices 20A, 20B, and 20C as necessary.
- only the same reference numerals are given.
- the content reproduction devices 20A, 20B, and 20C they are simply referred to as the content reproduction device 20.
- FIG. 1 is an explanatory diagram showing a configuration of a content reproduction system according to an embodiment of the present disclosure.
- the content reproduction system according to an embodiment of the present disclosure includes content servers 10 and 13, a network camera 11, a network 12, and a content reproduction device 20 (client device).
- the content server 10 and the content reproduction apparatus 20 are connected via the network 12.
- the network 12 is a wired or wireless transmission path for information transmitted from a device connected to the network 12.
- the network 12 may include a public line network such as the Internet, a telephone line network, a satellite communication network, various LANs including the Ethernet (Registered Trademark), a WAN (Wide Area Network), and the like. Further, the network 12 may include a dedicated line network such as an IP-VPN (Internet Protocol-Virtual Private Network).
- a public line network such as the Internet
- a telephone line network such as the Internet
- a satellite communication network such as the Internet
- WAN Wide Area Network
- the network 12 may include a dedicated line network such as an IP-VPN (Internet Protocol-Virtual Private Network).
- IP-VPN Internet Protocol-Virtual Private Network
- the content server 10 encodes the content data, generates a data file including the encoded data and the meta information of the encoded data, and stores the data file.
- the encoded data corresponds to “mdat” and the meta information corresponds to “moov”.
- the content data may be music data such as music, lectures and radio programs, video data such as movies, television programs, video programs, photographs, documents, pictures and charts, games and software.
- the content data may be a video imaged by the network camera 11.
- the content server 10 can control the network camera 11 in response to a request from the content reproduction device 20.
- the content server 10 generates a plurality of data files at different bit rates for the same content. Further, in response to a content reproduction request from the content reproduction device 20, the content server 13 according to the present embodiment includes the URL information of the content server 10 including information on parameters to be added to the URL by the content reproduction device 20. Transmit to the playback device 20.
- the matter will be specifically described with reference to FIG.
- FIG. 2 is an explanatory diagram showing a data flow in the content reproduction system according to the present embodiment.
- the content server 10 encodes the same content data at different bit rates, and generates, for example, a file A of 2 Mbps, a file B of 1.5 Mbps, and a file C of 1 Mbps as shown in FIG. In comparison, file A has a high bit rate, file B has a standard bit rate, and file C has a low bit rate.
- the encoded data of each file is divided into a plurality of segments.
- the encoded data of file A is divided into segments “A1”, “A2”, “A3”,... “An”
- the encoded data of file B is “B1”, “B2”, “B3”,... “Bn” is segmented
- the encoded data of file C is segmented as “C1”, “C2”, “C3”,. .
- Each segment is composed of one or more video encoded data and audio encoded data that can be reproduced independently, starting with an MP4 sync sample (for example, IDR-picture in the case of AVC / H.264 video encoding). It may be constituted by. For example, when video data of 30 frames per second is encoded with a GOP (Group of Pictures) with a fixed length of 15 frames, each segment is encoded video and audio data for 2 seconds corresponding to 4 GOP. Alternatively, it may be 10 seconds worth of video and audio encoded data corresponding to 20 GOP.
- an MP4 sync sample for example, IDR-picture in the case of AVC / H.264 video encoding.
- GOP Group of Pictures
- each segment is encoded video and audio data for 2 seconds corresponding to 4 GOP.
- the playback range (the range of the time position from the beginning of the content) by the segment with the same arrangement order in each file is the same.
- the playback ranges of the segment “A2”, the segment “B2”, and the segment “C2” are the same and each segment is encoded data for 2 seconds
- the segment “A2”, the segment “B2”, and The playback range of the segment “C2” is 2 to 4 seconds for the content.
- the content server 10 When the content server 10 generates the file A to the file C composed of such a plurality of segments, the content server 10 stores the file A to the file C. Then, as shown in FIG. 2, the content server 10 sequentially transmits segments constituting different files to the content reproduction device 20, and the content reproduction device 20 performs streaming reproduction of the received segments.
- the content server 10 transmits a playlist file (hereinafter referred to as MPD: Media Presentation Description) including the bit rate information and access information of each encoded data to the content playback device 20, and the content playback device 20. Selects one of a plurality of bit rates based on the MPD, and requests the content server 10 to transmit a segment corresponding to the selected bit rate.
- MPD Media Presentation Description
- FIG. 1 only one content server 10 is illustrated, but it goes without saying that the present disclosure is not limited to such an example.
- FIG. 3 is an explanatory diagram showing a specific example of MPD.
- MPD includes access information regarding a plurality of encoded data having different bit rates (BANDWIDTH).
- BANDWIDTH bit rates
- the MPD shown in FIG. 3 indicates that each encoded data of 256 Kbps, 1.024 Mbps, 1.384 Mbps, 1.536 Mbps, and 2.048 Mbps exists, and includes access information regarding each encoded data.
- the content playback apparatus 20 can dynamically change the bit rate of encoded data to be streamed based on such MPD.
- FIG. 1 shows a mobile terminal as an example of the content playback apparatus 20, the content playback apparatus 20 is not limited to such an example.
- the content playback device 20 is an information processing device such as a PC (Personal Computer), a home video processing device (DVD recorder, VCR, etc.), a PDA (Personal Digital Assistant), a home game device, or a home appliance. Also good.
- the content playback device 20 may be an information processing device such as a mobile phone, a PHS (Personal Handyphone System), a portable music playback device, a portable video processing device, or a portable game device.
- FIG. 4 is a functional block diagram showing the configuration of the content server 10 according to the present embodiment.
- the content server 10 according to the present embodiment includes a file generation unit 120, a storage unit 130, and a communication unit 140.
- the file generation unit 120 includes an encoder 122 that encodes content data, and generates a plurality of encoded data having the same content and different bit rates, and the MPD described above. For example, when the encoded data of 256 Kbps, 1.024 Mbps, 1.384 Mbps, 1.536 Mbps, and 2.048 Mbps is generated, the file generation unit 120 generates MPD as shown in FIG.
- the storage unit 130 stores a plurality of encoded data and MPD having different bit rates generated by the file generation unit 120.
- the storage unit 130 may be a storage medium such as a non-volatile memory, a magnetic disk, an optical disk, and an MO (Magneto Optical) disk.
- Non-volatile memories include, for example, EEPROM (Electrically Erasable Programmable Read-Only Memory) and EPROM (Erasable Programmable ROM).
- Examples of the magnetic disk include a hard disk and a disk type magnetic disk.
- Examples of the optical disc include a CD (Compact Disc, a DVD-R (Digital Versatile Disc Recordable), a BD (Blu-Ray Disc (registered trademark)), and the like.
- the communication unit 140 is an interface with the content reproduction apparatus 20 and communicates with the content reproduction apparatus 20 via the network 12. More specifically, the communication unit 140 has a function as an HTTP server that communicates with the content reproduction apparatus 20 according to HTTP. For example, the communication unit 140 transmits the MPD to the content reproduction device 20, extracts encoded data requested from the content reproduction device 20 based on the MPD according to HTTP from the storage unit 130, and transmits it to the content reproduction device 20 as an HTTP response. Transmit encoded data.
- Configuration of Content Playback Device 20 The configuration of the content server 10 according to the present embodiment has been described above. Next, the configuration of the content reproduction apparatus 20 according to the present embodiment will be described with reference to FIG.
- FIG. 5 is a functional block diagram showing the configuration of the content reproduction apparatus 20 according to the present embodiment.
- the content reproduction apparatus 20 according to the present embodiment includes a communication unit 220, a storage unit 230, a reproduction unit 240, and a selection unit 250.
- the communication unit 220 is an interface with the content server 10, requests data from the content server 10, and acquires data from the content server 10. More specifically, the communication unit 220 has a function as an HTTP client that communicates with the content reproduction apparatus 20 according to HTTP. For example, the communication unit 220 can selectively acquire a segment of MPD or encoded data from the content server 10 by using HTTP Range.
- the storage unit 230 stores various information related to content reproduction. For example, the segments acquired from the content server 10 by the communication unit 220 are sequentially buffered. The segments of the encoded data buffered in the storage unit 230 are sequentially supplied to the reproduction unit 240 using a FIFO (First In First Out).
- FIFO First In First Out
- the storage unit 230 instructs the network camera 11 on the direction and magnification, and holds definition information for acquiring the video captured by the network camera 11 from the content server 10.
- the definition information held by the storage unit 230 will be described in detail later.
- the playback unit 240 sequentially plays back the segments supplied from the storage unit 230. Specifically, the playback unit 240 performs segment decoding, DA conversion, rendering, and the like.
- the selection unit 250 sequentially selects within the same content which segment of the encoded data corresponding to which bit rate included in the MPD is to be acquired. For example, when the selection unit 250 sequentially selects the segments “A1”, “B2”, and “A3” according to the bandwidth of the network 12, the communication unit 220 receives the segment “A1” from the content server 10 as illustrated in FIG. ”,“ B2 ”, and“ A3 ”are acquired sequentially.
- the content reproduction apparatus 20 controls the network camera 11 to acquire the video captured by the network camera 11 from the content server 10 from the designated direction and at the designated magnification.
- the content playback apparatus 20 uses the ABS technique disclosed in Non-Patent Document 1 for controlling the network camera 11.
- FIG. 6 is an explanatory diagram illustrating a configuration example of the content server 13.
- the content server 13 includes a storage unit 310 and a communication unit 320.
- Storage unit 310 stores MPD URL information.
- the MPD URL information is transmitted from the content server 13 to the content reproduction device 20 in response to a request from the content reproduction device 20 that requests content reproduction.
- the storage unit 310 stores definition information used when the content reproduction apparatus 20 adds a parameter to the URL described in the MPD when providing the MPD URL information to the content reproduction apparatus 20.
- the communication unit 320 is an interface with the content reproduction apparatus 20 and communicates with the content reproduction apparatus 20 via the network 12. That is, the communication unit 320 receives a request for MPD URL information from the content playback apparatus 20 that requests content playback, and transmits the MPD URL information to the content playback apparatus 20.
- the MPD URL transmitted from the communication unit 320 includes information for adding parameters by the content reproduction apparatus 20.
- the parameters to be added to the MPD URL by the content playback apparatus 20 can be variously set by definition information shared by the content server 13 and the content playback apparatus 20. For example, information such as the current position of the content playback device 20, the user ID of the user who uses the content playback device 20, the memory size of the content playback device 20, and the storage capacity of the content playback device 20 may be used. Can be added to the MPD URL.
- the SegmentTemplate element whose contents are disclosed in section 5.3.9.4.2 of the Non-Patent Document 1 includes the pan angle attribute “pan_range”, the tilt angle A method for designating an attribute “tilt_range” and a zoom magnification attribute “zoom_range” has been proposed.
- the attribute “pd” is an attribute for designating one or more URIs of definition information for controlling the network camera 11.
- the definition information specified by the attribute “pd” of the SegmentTemplate element is also referred to as Parameter Description.
- Parameter Description is, for example, WADL (Web Application Description Language, http://www.w3.org/Submission/wadl/), WSDL (Web Service Description Language, http://www.ibm.com/developerworks/webservices/ library / ws-restwsdl /) may be described based on other Web API description languages.
- This definition information can be stored in the storage unit 310 or the storage unit 230.
- Parameter Description is stored in both storage unit 310 and storage unit 230, it is assumed that both contents are synchronized, that is, both contents are kept the same.
- the URL for acquiring video captured by the network camera 11 is described in the MPD as “http://cdn1.example.com/camera1/”, and the EssentialProperty element of the ⁇ AdaptationSet> tag of the MPD
- FIG. 7 is an explanatory diagram showing the contents of “urn: PanAngleDef”, which is the above Parameter Description.
- This Parameter Description is definition information for designating the pan angle of the network camera 11.
- the ⁇ resource> tag is a tag that designates the location of the image captured by the network camera 11, and the location is described in the element “path”. In the example of FIG. 6, “pan- ⁇ degree ⁇ ” is described. By setting a value for this “degree” and making a request to the content server 10, the content playback apparatus 20 can specify the pan angle of the network camera 11.
- the “name” element specifies the name of a parameter for controlling the network camera 11 from the content playback apparatus 20.
- a parameter “degree” for designating the pan angle of the network camera 11 is designated in the “name” element.
- the “required” element specifies whether or not the description of the query parameter in the content playback device 20 is mandatory. In the example of FIG. 7, “true” is specified in the “required” element, which indicates that the description of this parameter is essential.
- the “style” element defines the format of parameters to be added to the MPD URL by the content playback apparatus 20.
- the parameter format includes a query parameter and a template.
- template is designated in the “style” element, and this parameter indicates that the content reproduction apparatus 20 is described in the template format.
- FIG. 7 shows details about the parameter “degree”, and indicates that the maximum value, minimum value, and increase amount of the pan angle are designated.
- the content playback device 20 adds parameters based on the contents shown in FIG. For example, when the content reproduction apparatus 20 requests a video imaged from a direction with a pan angle of 40 degrees, the content reproduction apparatus 20 designates the URL as “http://cdn1.example.com/camera1/pan-40” and sets the communication unit 220. To access the URL.
- the content server 10 can designate the pan angle of the network camera 11 as 40 degrees based on a request from the content playback device 20 and transmit the video captured by the network camera 11 to the content playback device 20.
- FIG. 8 is an explanatory diagram showing the contents of “urn: ZoomScaleDef”, which is the above Parameter Description.
- the Parameter Description is definition information for designating the zoom magnification of the network camera 11.
- the content playback apparatus 20 when requesting an image with a zoom magnification of 3 times, designates the URL as “http://cdn1.example.com/camera1/zoom-3” and transmits the URL from the communication unit 220.
- the content server 10 can specify the zoom magnification of the network camera 11 as 3 times based on a request from the content playback device 20 and transmit the video captured by the network camera 11 to the content playback device 20.
- FIG. 9 is an explanatory diagram showing an example of an MPD configured to refer to the Parameter Description described above.
- the Parameter element is introduced, the name of the parameter used as a template (“angle” in the example of FIG. 9) is specified in the id attribute of the Parameter element, and the EssentialProperty (target property) is specified from the descriptorId attribute. It is proposed to refer to elements that specify descriptors that should be supported by MPD. However, the above non-patent document 2 does not propose a definition beyond that.
- the MPD shown in FIG. 9 uses the above-described Parameter Description for the definition ahead of it.
- PanAngleDef specified by the schemeIdUri element is the Parameter Description shown in FIG. Since the MPD is defined in this way, the content reproduction apparatus 20 can access the address http://cdn1.example.com/camera1/pan-40 and acquire the content.
- a standard description method for Parameter Description it is possible to use an existing standard framework, and it is easy to add functions such as a URL signature.
- 10A and 10B are sequence diagrams showing in detail the series of processes described above.
- the MPD URL information is sent from the content server 13 to the content playback device 20, and the content playback device 20 acquires the content from the network camera 11 based on the MPD URL.
- video which the camera 11 imaged is shown.
- the content playback device 20 When the content playback device 20 intends to play back content on the playback unit 240, the content playback device 20 first acquires the MPD URL from the content server 13 (step S101). Acquisition of step S101 is performed by the reproducing unit 240 through the communication unit 220, for example. After acquiring the MPD URL from the content server 13, the content reproduction apparatus 20 subsequently determines whether the MPD includes a URL parameter reference (step S102). The playback unit 240 can execute the determination in step S102.
- the MPD including the reference of the URL parameter refers to an MPD including a parameter such as “urn: PanAngleDef” illustrated in FIG.
- step S102 if the MPD URL acquired in step S101 is not an MPD that includes a URL parameter reference, the content playback apparatus 20 requests the MPD URL from the content server 13 via http (step S103). ).
- the request in step S103 is executed by the playback unit 240 through the communication unit 220, for example. For example, when the content reproduction device 20 acquires the URL “http://a.com/x.mpd”, the content reproduction device 20 sets “http://a.com/x.mpd” by http-request. Requests to the content server 13.
- the content server 13 that has received the MPD URL request from the content playback apparatus 20 determines state information desired to be acquired from the content playback apparatus 20, for example, the angle information of the network camera 11 in the above example, and stores the corresponding Parameter Description in the storage unit. 310 is inquired (step S104). The storage unit 310 responds with the URI of the Parameter Description (step S105). If applied to the above-described example, the URI “urn: PanAngleDef” is returned from the storage unit 310 in response to the inquiry from the content server 10.
- the content server 10 that has received the response from the storage unit 310 stores the URI (urn: PanAngleDef) of Parameter Description in the MPD. Further, the content server 10 adds the attribute pd of the SegmentTemplate element to update the MPD (for example, y.mpd), and the MPD URL (http://a.com/y.mpd) indicating the updated MPD In response to the content reproduction apparatus 20 by http-response-redirect (step S106).
- URI Urn: PanAngleDef
- the content server 10 adds the attribute pd of the SegmentTemplate element to update the MPD (for example, y.mpd), and the MPD URL (http://a.com/y.mpd) indicating the updated MPD
- http-response-redirect step S106.
- the content reproduction device 20 that has received the response from the content server 13 requests the updated MPD URL from the content server 10 via http (step S107). If applied to the above example, the content reproduction apparatus 20 requests the MPD URL “http://a.com/y.mpd” to the content server 10 by http-request.
- the content server 10 that has received the request from the content playback device 20 responds the main body of the MPD to the content playback device 20 (step S108).
- the MPD body is sent from the content server 10 (or as a result of the determination in step S102, if the MPD includes the URL parameter reference acquired in step S101), the MPD body is sent from the content server 10.
- the received content reproduction apparatus 20 interprets the sent MPD (step S109).
- the playback unit 240 can execute the MPD interpretation in step S109.
- the content playback apparatus 20 interprets the MPD sent from the content server 10 and determines an AdaptationSet or Representation / SubRepresentation to be played back.
- the content reproduction apparatus 20 executes a detection process for the reproduction target (Step S110).
- the reproducing unit 240 can execute the detection process in step S110.
- the content playback apparatus 20 has an EssentialProperty in the AdaptationSet or Representation / SubRepresentation, the value of the schemeIdUri attribute is specified in the EssentialProperty (for example, urn: PanAngleDef), and the URI specified in the schemeIdUri attribute It is detected that the interpretation of the Parameter Description that is referred to is essential for the playback process. Further, the content reproduction apparatus 20 detects that the URI is a URI meaning insertion of a URL parameter.
- the content reproduction device 20 executes the detection process for the reproduction target, the content reproduction device 20 subsequently stores the URI (step S111). Specifically, the content playback apparatus 20 stores a URI (urn: PanAngleDef) specified by the attribute pd of the SegmentTemplate element. The storage process of the URI is executed by the playback unit 240, for example.
- URI Urn: PanAngleDef
- the content reproduction device 20 inquires of the storage unit 310 for a parameter description using a URI (for example, urn: PanAngleDef) specified by the URL parameter (step S112).
- a URI for example, urn: PanAngleDef
- the playback unit 240 executes the inquiry process in step S112 through the communication unit 220.
- the storage unit 310 stores the parameter described in, for example, WADL.
- the main body of Description is responded to the content reproduction apparatus 20 (step S113).
- the content playback apparatus 20 determines the parameter configuration method and acquires the content stored in the parameter (step S114). For example, the playback unit 240 can execute the process of step S114. For example, when the angle parameter of the network camera 11 is stored in the parameter, the content reproduction apparatus 20 stores the angle value of the network camera 11 in the parameter through interactive processing with the user or the like.
- the content playback device 20 When the content playback device 20 acquires the content to be stored in the parameter, the content playback device 20 generates a URL of the segment to which the parameter is added, and makes a request to the content server 10 via http (step S115). Specifically, the content reproduction device 20 generates the URL of the segment to which the URL parameter is added in accordance with the rules described in the media attribute of the SegmentTemplate element. For example, if there is a rule for inserting the parameter in the substitution parameter portion “angle” surrounded by “$” in FIG. 9, the content playback apparatus 20 follows the rule and the URL of the segment to which the parameter is added. Is generated. The content reproduction apparatus 20 generates a URL of the segment “http://cdn1.example.com/camera1/pan-40” by the process of step S115.
- the content server 10 Upon receipt of the request from the content playback apparatus 20, the content server 10 interprets the URL parameter and responds to the content playback apparatus 20 with the optimum segment body (step S116). The content reproduction device 20 reproduces the segment received from the content server 10 (step S117).
- the content reproduction apparatus 20 adds a parameter to the URL described in the MPD by executing the above-described series of processing with the content server 10, and http://cdn1.example.com/camera1/ By accessing the address pan-40, the video captured by the network camera 11 whose angle is 40 degrees can be acquired, and the acquired video can be reproduced.
- FIG. 11 is a block diagram showing an example of the hardware configuration of a computer that executes the above-described series of processing by a program.
- a CPU Central Processing Unit
- ROM Read Only Memory
- RAM Random Access Memory
- An input / output interface 205 is further connected to the bus 204.
- An input unit 206, an output unit 207, a storage unit 208, a communication unit 209, and a drive 210 are connected to the input / output interface 205.
- the input unit 206 includes a keyboard, a mouse, a microphone, and the like.
- the output unit 207 includes a display, a speaker, and the like.
- the storage unit 208 includes a hard disk, a nonvolatile memory, and the like.
- the communication unit 209 includes a network interface and the like.
- the drive 210 drives a removable medium 211 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory.
- the CPU 201 loads, for example, the program stored in the storage unit 208 to the RAM 203 via the input / output interface 205 and the bus 204 and executes the program. Is performed.
- the program executed by the computer (CPU 201) can be provided by being recorded in the removable medium 211 as a package medium or the like, for example.
- the program can be provided via a wired or wireless transmission medium such as a local area network, the Internet, or digital satellite broadcasting.
- the program can be installed in the storage unit 208 via the input / output interface 205 by attaching the removable medium 211 to the drive 210.
- the program can be received by the communication unit 209 via a wired or wireless transmission medium and installed in the storage unit 208.
- the program can be installed in advance in the ROM 202 or the storage unit 208.
- the program executed by the computer may be a program that is processed in time series in the order described in this specification, or in parallel or at a necessary timing such as when a call is made. It may be a program for processing.
- FIG. 12 shows an example of the multi-view image encoding method.
- the multi-viewpoint image includes images of a plurality of viewpoints (views). Multiple views of this multi-viewpoint image are encoded using the base view that encodes and decodes using only the image of its own view without using the image of the other view, and the image of the other view.
- -It consists of a non-base view that performs decoding. For the non-base view, an image of the base view may be used, or an image of another non-base view may be used.
- FIG. 13 is a diagram illustrating a multi-view image encoding apparatus that performs the above-described multi-view image encoding.
- the multi-view image encoding device 600 includes an encoding unit 601, an encoding unit 602, and a multiplexing unit 603.
- the encoding unit 601 encodes the base view image and generates a base view image encoded stream.
- the encoding unit 602 encodes the non-base view image and generates a non-base view image encoded stream.
- the multiplexing unit 603 multiplexes the base view image encoded stream generated by the encoding unit 601 and the non-base view image encoded stream generated by the encoding unit 602 to generate a multi-view image encoded stream. To do.
- the encoding device 10 (FIG. 19) can be applied to the encoding unit 601 and the encoding unit 602 of the multi-view image encoding device 600. That is, in the encoding for each view, the default mapping between the enhancement layer and the resolution information can be improved. Also, the encoding unit 601 and the encoding unit 602 can perform encoding using the same flags and parameters (for example, syntax elements related to processing between images) (that is, share the flags and parameters). Therefore, it is possible to suppress a reduction in encoding efficiency.
- flags and parameters for example, syntax elements related to processing between images
- FIG. 14 is a diagram illustrating a multi-view image decoding apparatus that performs the above-described multi-view image decoding.
- the multi-view image decoding device 610 includes a demultiplexing unit 611, a decoding unit 612, and a decoding unit 613.
- the demultiplexing unit 611 demultiplexes the multi-view image encoded stream in which the base view image encoded stream and the non-base view image encoded stream are multiplexed, and the base view image encoded stream and the non-base view image The encoded stream is extracted.
- the decoding unit 612 decodes the base view image encoded stream extracted by the demultiplexing unit 611 to obtain a base view image.
- the decoding unit 613 decodes the non-base view image encoded stream extracted by the demultiplexing unit 611 to obtain a non-base view image.
- the decoding device 110 (FIG. 25) can be applied to the decoding unit 612 and the decoding unit 613 of the multi-view image decoding device 610. That is, the default mapping between the enhancement layer and the resolution information can be improved.
- the decoding unit 612 and the decoding unit 613 can perform decoding using the same flags and parameters (for example, syntax elements related to processing between images) (that is, the flags and parameters can be shared). Therefore, it is possible to suppress a reduction in encoding efficiency.
- FIG. 15 shows an example of a hierarchical image encoding method.
- Hierarchical image coding is a method in which image data is divided into a plurality of layers (hierarchized) so as to have a scalable function for a predetermined parameter, and is encoded for each layer.
- Hierarchical image decoding is decoding corresponding to the hierarchical image encoding.
- the hierarchized image includes images of a plurality of hierarchies (layers) having different predetermined parameter values.
- a plurality of layers of this hierarchical image are encoded / decoded using only the image of the own layer without using the image of the other layer, and encoded / decoded using the image of the other layer.
- It consists of a non-base layer (also called enhancement layer) that performs decoding.
- the non-base layer an image of the base layer may be used, or an image of another non-base layer may be used.
- the non-base layer is composed of difference image data (difference data) between its own image and an image of another layer so that redundancy is reduced.
- difference image data difference data
- an image with lower quality than the original image can be obtained using only the base layer data.
- an original image that is, a high-quality image
- image compression information of only the base layer (base layer) is transmitted, and a moving image with low spatiotemporal resolution or poor image quality is played
- base layer For terminals with high processing power, such as televisions and personal computers, in addition to the base layer (base layer), image enhancement information of the enhancement layer (enhancement layer) is transmitted.
- Image compression information corresponding to the capabilities of the terminal and the network can be transmitted from the server without performing transcoding processing, such as playing a moving image with high image quality.
- parameters having a scalable function are arbitrary.
- the spatial resolution as shown in FIG. 16 may be used as the parameter (spatial scalability).
- the resolution of the image is different for each layer. That is, in this case, as shown in FIG. 16, each picture has two layers of a base layer having a spatially lower resolution than the original image and an enhancement layer from which the original spatial resolution can be obtained by combining with the base layer. Is layered.
- this number of hierarchies is an example, and the number of hierarchies can be hierarchized.
- a temporal resolution as shown in FIG. 17 may be applied (temporal scalability).
- the frame rate is different for each layer. That is, in this case, as shown in FIG. 17, each picture is divided into two layers of a base layer having a lower frame rate than the original moving image and an enhancement layer in which the original frame rate can be obtained by combining with the base layer. Layered.
- this number of hierarchies is an example, and the number of hierarchies can be hierarchized.
- a signal-to-noise ratio (SNR (Signal-to-Noise-ratio)) may be applied (SNR--scalability) as a parameter for providing such scalability.
- SNR Signal-to-noise ratio
- the SN ratio is different for each layer. That is, in this case, as shown in FIG. 18, each picture is hierarchized into two layers: a base layer having a lower SNR than the original image and an enhancement layer from which the original SNR can be obtained by combining with the base layer.
- this number of hierarchies is an example, and the number of hierarchies can be hierarchized.
- bit depth can also be used as a parameter for providing scalability (bit-depth scalability).
- bit-depth scalability bit depth scalability
- the bit depth differs for each layer.
- the base layer is composed of an 8-bit image
- an enhancement layer is added to the base layer, whereby a 10-bit image can be obtained.
- chroma format can be used as a parameter for providing scalability (chroma scalability).
- the chroma format differs for each layer.
- the base layer is composed of component images in 4: 2: 0 format, and by adding an enhancement layer (enhancement layer) to this, a component image in 4: 2: 2 format can be obtained. Can be.
- FIG. 19 is a diagram illustrating a hierarchical image encoding apparatus that performs the above-described hierarchical image encoding.
- the hierarchical image encoding device 620 includes an encoding unit 621, an encoding unit 622, and a multiplexing unit 623.
- the encoding unit 621 encodes the base layer image and generates a base layer image encoded stream.
- the encoding unit 622 encodes the non-base layer image and generates a non-base layer image encoded stream.
- the multiplexing unit 623 multiplexes the base layer image encoded stream generated by the encoding unit 621 and the non-base layer image encoded stream generated by the encoding unit 622 to generate a hierarchical image encoded stream. .
- the encoding device 10 (FIG. 19) can be applied to the encoding unit 621 and the encoding unit 622 of the hierarchical image encoding device 620. That is, the default mapping between the enhancement layer and the resolution information can be improved. Also, the encoding unit 621 and the encoding unit 622 can perform control of intra prediction filter processing using the same flags and parameters (for example, syntax elements related to processing between images) (that is, the intra prediction processing). Therefore, it is possible to share a flag and a parameter), and it is possible to suppress a reduction in encoding efficiency.
- flags and parameters for example, syntax elements related to processing between images
- FIG. 20 is a diagram illustrating a hierarchical image decoding apparatus that performs the hierarchical image decoding described above.
- the hierarchical image decoding device 630 includes a demultiplexing unit 631, a decoding unit 632, and a decoding unit 633.
- the demultiplexing unit 631 demultiplexes the hierarchical image encoded stream in which the base layer image encoded stream and the non-base layer image encoded stream are multiplexed, and the base layer image encoded stream and the non-base layer image code Stream.
- the decoding unit 632 decodes the base layer image encoded stream extracted by the demultiplexing unit 631 to obtain a base layer image.
- the decoding unit 633 decodes the non-base layer image encoded stream extracted by the demultiplexing unit 631 to obtain a non-base layer image.
- the decoding device 110 (FIG. 25) can be applied to the decoding unit 632 and the decoding unit 633 of the hierarchical image decoding device 630. That is, the default mapping between the enhancement layer and the resolution information can be improved.
- the decoding unit 612 and the decoding unit 613 can perform decoding using the same flags and parameters (for example, syntax elements related to processing between images) (that is, the flags and parameters can be shared). Therefore, it is possible to suppress a reduction in encoding efficiency.
- FIG. 21 illustrates a schematic configuration of a television apparatus to which the present disclosure is applied.
- the television apparatus 900 includes an antenna 901, a tuner 902, a demultiplexer 903, a decoder 904, a video signal processing unit 905, a display unit 906, an audio signal processing unit 907, a speaker 908, and an external interface unit 909. Furthermore, the television apparatus 900 includes a control unit 910, a user interface unit 911, and the like.
- the tuner 902 selects a desired channel from the broadcast wave signal received by the antenna 901, demodulates it, and outputs the obtained encoded bit stream to the demultiplexer 903.
- the demultiplexer 903 extracts video and audio packets of the program to be viewed from the encoded bit stream, and outputs the extracted packet data to the decoder 904. Further, the demultiplexer 903 supplies a packet of data such as EPG (Electronic Program Guide) to the control unit 910. If scrambling is being performed, descrambling is performed by a demultiplexer or the like.
- EPG Electronic Program Guide
- the decoder 904 performs packet decoding processing, and outputs video data generated by the decoding processing to the video signal processing unit 905 and audio data to the audio signal processing unit 907.
- the video signal processing unit 905 performs noise removal, video processing according to user settings, and the like on the video data.
- the video signal processing unit 905 generates video data of a program to be displayed on the display unit 906, image data by processing based on an application supplied via a network, and the like.
- the video signal processing unit 905 generates video data for displaying a menu screen for selecting an item and the like, and superimposes the video data on the video data of the program.
- the video signal processing unit 905 generates a drive signal based on the video data generated in this way, and drives the display unit 906.
- the display unit 906 drives a display device (for example, a liquid crystal display element or the like) based on a drive signal from the video signal processing unit 905 to display a program video or the like.
- a display device for example, a liquid crystal display element or the like
- the audio signal processing unit 907 performs predetermined processing such as noise removal on the audio data, performs D / A conversion processing and amplification processing on the processed audio data, and outputs the audio data to the speaker 908.
- the external interface unit 909 is an interface for connecting to an external device or a network, and transmits and receives data such as video data and audio data.
- a user interface unit 911 is connected to the control unit 910.
- the user interface unit 911 includes an operation switch, a remote control signal receiving unit, and the like, and supplies an operation signal corresponding to a user operation to the control unit 910.
- the control unit 910 is configured using a CPU (Central Processing Unit), a memory, and the like.
- the memory stores a program executed by the CPU, various data necessary for the CPU to perform processing, EPG data, data acquired via a network, and the like.
- the program stored in the memory is read and executed by the CPU at a predetermined timing such as when the television device 900 is activated.
- the CPU executes each program to control each unit so that the television device 900 operates in accordance with the user operation.
- the television device 900 includes a bus 912 for connecting the tuner 902, the demultiplexer 903, the video signal processing unit 905, the audio signal processing unit 907, the external interface unit 909, and the control unit 910.
- the decoder 904 is provided with the function of the decoding apparatus (decoding method) of the present application. For this reason, in the decoding process of the encoded stream, the default mapping between the enhancement layer and the resolution information can be improved.
- FIG. 22 illustrates a schematic configuration of a mobile phone to which the present disclosure is applied.
- the cellular phone 920 includes a communication unit 922, an audio codec 923, a camera unit 926, an image processing unit 927, a demultiplexing unit 928, a recording / reproducing unit 929, a display unit 930, and a control unit 931. These are connected to each other via a bus 933.
- an antenna 921 is connected to the communication unit 922, and a speaker 924 and a microphone 925 are connected to the audio codec 923. Further, an operation unit 932 is connected to the control unit 931.
- the mobile phone 920 performs various operations such as transmission / reception of voice signals, transmission / reception of e-mail and image data, image shooting, and data recording in various modes such as a voice call mode and a data communication mode.
- the voice signal generated by the microphone 925 is converted into voice data and compressed by the voice codec 923 and supplied to the communication unit 922.
- the communication unit 922 performs audio data modulation processing, frequency conversion processing, and the like to generate a transmission signal.
- the communication unit 922 supplies a transmission signal to the antenna 921 and transmits it to a base station (not shown).
- the communication unit 922 performs amplification, frequency conversion processing, demodulation processing, and the like of the reception signal received by the antenna 921, and supplies the obtained audio data to the audio codec 923.
- the audio codec 923 performs data expansion of the audio data and conversion to an analog audio signal and outputs the result to the speaker 924.
- the control unit 931 receives character data input by operating the operation unit 932 and displays the input characters on the display unit 930.
- the control unit 931 generates mail data based on a user instruction or the like in the operation unit 932 and supplies the mail data to the communication unit 922.
- the communication unit 922 performs mail data modulation processing, frequency conversion processing, and the like, and transmits the obtained transmission signal from the antenna 921.
- the communication unit 922 performs amplification, frequency conversion processing, demodulation processing, and the like of the reception signal received by the antenna 921, and restores mail data. This mail data is supplied to the display unit 930 to display the mail contents.
- the mobile phone 920 can also store the received mail data in a storage medium by the recording / playback unit 929.
- the storage medium is any rewritable storage medium.
- the storage medium is a removable memory such as a RAM, a semiconductor memory such as a built-in flash memory, a hard disk, a magnetic disk, a magneto-optical disk, an optical disk, a USB (Universal Serial Bus) memory, or a memory card.
- the image data generated by the camera unit 926 is supplied to the image processing unit 927.
- the image processing unit 927 performs encoding processing of image data and generates encoded data.
- the demultiplexing unit 928 multiplexes the encoded data generated by the image processing unit 927 and the audio data supplied from the audio codec 923 by a predetermined method, and supplies the multiplexed data to the communication unit 922.
- the communication unit 922 performs modulation processing and frequency conversion processing of multiplexed data, and transmits the obtained transmission signal from the antenna 921.
- the communication unit 922 performs amplification, frequency conversion processing, demodulation processing, and the like of the reception signal received by the antenna 921, and restores multiplexed data. This multiplexed data is supplied to the demultiplexing unit 928.
- the demultiplexing unit 928 performs demultiplexing of the multiplexed data, and supplies the encoded data to the image processing unit 927 and the audio data to the audio codec 923.
- the image processing unit 927 performs a decoding process on the encoded data to generate image data.
- the image data is supplied to the display unit 930 and the received image is displayed.
- the audio codec 923 converts the audio data into an analog audio signal, supplies the analog audio signal to the speaker 924, and outputs the received audio.
- the image processing unit 927 is provided with the functions of the encoding device and the decoding device (encoding method and decoding method) of the present application. For this reason, the default mapping between the enhancement layer and the resolution information can be improved.
- FIG. 23 illustrates a schematic configuration of a recording / reproducing apparatus to which the present disclosure is applied.
- the recording / reproducing apparatus 940 records, for example, audio data and video data of a received broadcast program on a recording medium, and provides the recorded data to the user at a timing according to a user instruction.
- the recording / reproducing device 940 can also acquire audio data and video data from another device, for example, and record them on a recording medium. Further, the recording / reproducing apparatus 940 decodes and outputs the audio data and video data recorded on the recording medium, thereby enabling image display and audio output on the monitor apparatus or the like.
- the recording / reproducing apparatus 940 includes a tuner 941, an external interface unit 942, an encoder 943, an HDD (Hard Disk Drive) unit 944, a disk drive 945, a selector 946, a decoder 947, an OSD (On-Screen Display) unit 948, a control unit 949, A user interface unit 950 is included.
- Tuner 941 selects a desired channel from a broadcast signal received by an antenna (not shown).
- the tuner 941 outputs an encoded bit stream obtained by demodulating the received signal of a desired channel to the selector 946.
- the external interface unit 942 includes at least one of an IEEE 1394 interface, a network interface unit, a USB interface, a flash memory interface, and the like.
- the external interface unit 942 is an interface for connecting to an external device, a network, a memory card, and the like, and receives data such as video data and audio data to be recorded.
- the encoder 943 performs encoding by a predetermined method when the video data and audio data supplied from the external interface unit 942 are not encoded, and outputs an encoded bit stream to the selector 946.
- the HDD unit 944 records content data such as video and audio, various programs, and other data on a built-in hard disk, and reads them from the hard disk during playback.
- the disk drive 945 records and reproduces signals with respect to the mounted optical disk.
- An optical disk such as a DVD disk (DVD-Video, DVD-RAM, DVD-R, DVD-RW, DVD + R, DVD + RW, etc.), a Blu-ray (registered trademark) disk, or the like.
- the selector 946 selects one of the encoded bit streams from the tuner 941 or the encoder 943 and supplies it to either the HDD unit 944 or the disk drive 945 when recording video or audio. Further, the selector 946 supplies the encoded bit stream output from the HDD unit 944 or the disk drive 945 to the decoder 947 at the time of reproduction of video and audio.
- the decoder 947 performs a decoding process on the encoded bit stream.
- the decoder 947 supplies the video data generated by performing the decoding process to the OSD unit 948.
- the decoder 947 outputs audio data generated by performing the decoding process.
- the OSD unit 948 generates video data for displaying a menu screen for selecting an item and the like, and superimposes it on the video data output from the decoder 947 and outputs the video data.
- a user interface unit 950 is connected to the control unit 949.
- the user interface unit 950 includes an operation switch, a remote control signal receiving unit, and the like, and supplies an operation signal corresponding to a user operation to the control unit 949.
- the control unit 949 is configured using a CPU, a memory, and the like.
- the memory stores programs executed by the CPU and various data necessary for the CPU to perform processing.
- the program stored in the memory is read and executed by the CPU at a predetermined timing such as when the recording / reproducing apparatus 940 is activated.
- the CPU executes the program to control each unit so that the recording / reproducing device 940 operates according to the user operation.
- the encoder 943 is provided with the function of the encoding apparatus (encoding method) of the present application. For this reason, in the encoding of the encoded stream, the default mapping between the enhancement layer and the resolution information can be improved.
- the decoder 947 is provided with the function of the decoding device (decoding method) of the present application. For this reason, in the decoding of the encoded stream, the default mapping between the enhancement layer and the resolution information can be improved.
- FIG. 24 illustrates a schematic configuration of an imaging apparatus to which the present disclosure is applied.
- the imaging device 960 images a subject, displays an image of the subject on a display unit, and records it on a recording medium as image data.
- the imaging device 960 includes an optical block 961, an imaging unit 962, a camera signal processing unit 963, an image data processing unit 964, a display unit 965, an external interface unit 966, a memory unit 967, a media drive 968, an OSD unit 969, and a control unit 970. Have. In addition, a user interface unit 971 is connected to the control unit 970. Furthermore, the image data processing unit 964, the external interface unit 966, the memory unit 967, the media drive 968, the OSD unit 969, the control unit 970, and the like are connected via a bus 972.
- the optical block 961 is configured using a focus lens, a diaphragm mechanism, and the like.
- the optical block 961 forms an optical image of the subject on the imaging surface of the imaging unit 962.
- the imaging unit 962 is configured using a CCD or CMOS image sensor, generates an electrical signal corresponding to the optical image by photoelectric conversion, and supplies the electrical signal to the camera signal processing unit 963.
- the camera signal processing unit 963 performs various camera signal processing such as knee correction, gamma correction, and color correction on the electrical signal supplied from the imaging unit 962.
- the camera signal processing unit 963 supplies the image data after the camera signal processing to the image data processing unit 964.
- the image data processing unit 964 performs an encoding process on the image data supplied from the camera signal processing unit 963.
- the image data processing unit 964 supplies the encoded data generated by performing the encoding process to the external interface unit 966 and the media drive 968. Further, the image data processing unit 964 performs a decoding process on the encoded data supplied from the external interface unit 966 and the media drive 968.
- the image data processing unit 964 supplies the image data generated by performing the decoding process to the display unit 965. Further, the image data processing unit 964 superimposes the processing for supplying the image data supplied from the camera signal processing unit 963 to the display unit 965 and the display data acquired from the OSD unit 969 on the image data. To supply.
- the OSD unit 969 generates display data such as a menu screen and icons made up of symbols, characters, or figures and outputs them to the image data processing unit 964.
- the external interface unit 966 includes, for example, a USB input / output terminal, and is connected to a printer when printing an image.
- a drive is connected to the external interface unit 966 as necessary, a removable medium such as a magnetic disk or an optical disk is appropriately mounted, and a computer program read from them is installed as necessary.
- the external interface unit 966 has a network interface connected to a predetermined network such as a LAN or the Internet.
- the control unit 970 reads the encoded data from the media drive 968 in accordance with an instruction from the user interface unit 971, and supplies the encoded data to the other device connected via the network from the external interface unit 966. it can.
- the control unit 970 may acquire encoded data and image data supplied from another device via the network via the external interface unit 966 and supply the acquired data to the image data processing unit 964. it can.
- any readable / writable removable medium such as a magnetic disk, a magneto-optical disk, an optical disk, or a semiconductor memory is used.
- the recording medium may be any type of removable medium, and may be a tape device, a disk, or a memory card. Of course, a non-contact IC (Integrated Circuit) card may be used.
- media drive 968 and the recording medium may be integrated and configured by a non-portable storage medium such as a built-in hard disk drive or an SSD (Solid State Drive).
- a non-portable storage medium such as a built-in hard disk drive or an SSD (Solid State Drive).
- the control unit 970 is configured using a CPU.
- the memory unit 967 stores a program executed by the control unit 970, various data necessary for the control unit 970 to perform processing, and the like.
- the program stored in the memory unit 967 is read and executed by the control unit 970 at a predetermined timing such as when the imaging device 960 is activated.
- the control unit 970 controls each unit so that the imaging device 960 performs an operation according to a user operation by executing a program.
- the image data processing unit 964 is provided with the functions of the encoding apparatus and decoding apparatus (encoding method and decoding method) of the present application. For this reason, it is possible to improve the default mapping between the enhancement layer and the resolution information in encoding or decoding of the encoded stream.
- Scalable encoding is used for selection of data to be transmitted, for example, as in the example shown in FIG.
- the distribution server 1002 reads the scalable encoded data stored in the scalable encoded data storage unit 1001, and via the network 1003, the personal computer 1004, the AV device 1005, the tablet This is distributed to the terminal device such as the device 1006 and the mobile phone 1007. *
- the distribution server 1002 selects and transmits encoded data of appropriate quality according to the capability of the terminal device, the communication environment, and the like. Even if the distribution server 1002 transmits unnecessarily high-quality data, the terminal device does not always obtain a high-quality image, and may cause a delay or an overflow. Moreover, there is a possibility that the communication band is unnecessarily occupied or the load on the terminal device is unnecessarily increased. On the other hand, even if the distribution server 1002 transmits unnecessarily low quality data, there is a possibility that an image with sufficient image quality cannot be obtained in the terminal device. Therefore, the distribution server 1002 appropriately reads and transmits the scalable encoded data stored in the scalable encoded data storage unit 1001 as encoded data having an appropriate quality with respect to the capability and communication environment of the terminal device. .
- the scalable encoded data storage unit 1001 stores scalable encoded data (BL + EL) 1011 encoded in a scalable manner.
- the scalable encoded data (BL + EL) 1011 is encoded data including both a base layer and an enhancement layer, and is a data that can be decoded to obtain both a base layer image and an enhancement layer image. It is.
- the distribution server 1002 selects an appropriate layer according to the capability of the terminal device that transmits data, the communication environment, and the like, and reads the data of the layer. For example, the distribution server 1002 reads high-quality scalable encoded data (BL + EL) 1011 from the scalable encoded data storage unit 1001 and transmits it to the personal computer 1004 and the tablet device 1006 with high processing capability as they are. . On the other hand, for example, the distribution server 1002 extracts base layer data from the scalable encoded data (BL + EL) 1011 for the AV device 1005 and the cellular phone 1007 having a low processing capability, and performs scalable encoding. Although it is data of the same content as the data (BL + EL) 1011, it is transmitted as scalable encoded data (BL) 1012 having a lower quality than the scalable encoded data (BL + EL) 1011.
- BL scalable encoded data
- scalable encoded data By using scalable encoded data in this way, the amount of data can be easily adjusted, so that the occurrence of delay and overflow can be suppressed, and the unnecessary increase in the load on the terminal device and communication medium can be suppressed. be able to.
- scalable encoded data (BL + EL) 1011 since scalable encoded data (BL + EL) 1011 has reduced redundancy between layers, the amount of data can be reduced as compared with the case where encoded data of each layer is used as individual data. . Therefore, the storage area of the scalable encoded data storage unit 1001 can be used more efficiently.
- the hardware performance of the terminal device varies depending on the device.
- the application which a terminal device performs is also various, the capability of the software is also various.
- the network 1003 serving as a communication medium can be applied to any communication network including wired, wireless, or both, such as the Internet and a LAN (Local Area Network), and has various data transmission capabilities. Furthermore, there is a risk of change due to other communications.
- the distribution server 1002 communicates with the terminal device that is the data transmission destination before starting data transmission, and the hardware performance of the terminal device, the performance of the application (software) executed by the terminal device, etc. Information regarding the capability of the terminal device and information regarding the communication environment such as the available bandwidth of the network 1003 may be obtained. The distribution server 1002 may select an appropriate layer based on the information obtained here.
- the layer extraction may be performed by the terminal device.
- the personal computer 1004 may decode the transmitted scalable encoded data (BL + EL) 1011 and display a base layer image or an enhancement layer image. Further, for example, the personal computer 1004 extracts the base layer scalable encoded data (BL) 1012 from the transmitted scalable encoded data (BL + EL) 1011 and stores it or transfers it to another device. The base layer image may be displayed after decoding.
- the numbers of the scalable encoded data storage unit 1001, the distribution server 1002, the network 1003, and the terminal devices are arbitrary.
- the example in which the distribution server 1002 transmits data to the terminal device has been described, but the usage example is not limited to this.
- the data transmission system 1000 may be any system as long as it transmits a scalable encoded data to a terminal device by selecting an appropriate layer according to the capability of the terminal device or a communication environment. Can be applied to the system.
- scalable coding is used for transmission via a plurality of communication media, for example, as in the example shown in FIG.
- a broadcasting station 1101 transmits base layer scalable encoded data (BL) 1121 by terrestrial broadcasting 1111. Also, the broadcast station 1101 transmits enhancement layer scalable encoded data (EL) 1122 via an arbitrary network 1112 including a wired or wireless communication network or both (for example, packetized transmission).
- BL base layer scalable encoded data
- EL enhancement layer scalable encoded data
- the terminal apparatus 1102 has a reception function of the terrestrial broadcast 1111 broadcast by the broadcast station 1101 and receives base layer scalable encoded data (BL) 1121 transmitted via the terrestrial broadcast 1111.
- the terminal apparatus 1102 further has a communication function for performing communication via the network 1112, and receives enhancement layer scalable encoded data (EL) 1122 transmitted via the network 1112.
- BL base layer scalable encoded data
- EL enhancement layer scalable encoded data
- the terminal device 1102 decodes the base layer scalable encoded data (BL) 1121 acquired via the terrestrial broadcast 1111 according to, for example, a user instruction, and obtains or stores a base layer image. Or transmit to other devices.
- BL base layer scalable encoded data
- the terminal device 1102 for example, in response to a user instruction, the base layer scalable encoded data (BL) 1121 acquired via the terrestrial broadcast 1111 and the enhancement layer scalable encoded acquired via the network 1112 Data (EL) 1122 is combined to obtain scalable encoded data (BL + EL), or decoded to obtain an enhancement layer image, stored, or transmitted to another device.
- BL base layer scalable encoded data
- EL enhancement layer scalable encoded acquired via the network 1112 Data
- the scalable encoded data can be transmitted via a communication medium that is different for each layer, for example. Therefore, the load can be distributed, and the occurrence of delay and overflow can be suppressed.
- the communication medium used for transmission may be selected for each layer. For example, scalable encoded data (BL) 1121 of a base layer having a relatively large amount of data is transmitted via a communication medium having a wide bandwidth, and scalable encoded data (EL) 1122 having a relatively small amount of data is transmitted. You may make it transmit via a communication medium with a narrow bandwidth. Further, for example, the communication medium for transmitting the enhancement layer scalable encoded data (EL) 1122 is switched between the network 1112 and the terrestrial broadcast 1111 according to the available bandwidth of the network 1112. May be. Of course, the same applies to data of an arbitrary layer.
- the number of layers is arbitrary, and the number of communication media used for transmission is also arbitrary.
- the number of terminal devices 1102 serving as data distribution destinations is also arbitrary.
- broadcasting from the broadcasting station 1101 has been described as an example, but the usage example is not limited to this.
- the data transmission system 1100 can be applied to any system as long as it is a system that divides scalable encoded data into a plurality of layers and transmits them through a plurality of lines.
- scalable encoding is used for storing encoded data as in the example shown in FIG. 27, for example.
- the imaging device 1201 performs scalable coding on image data obtained by imaging the subject 1211, and as scalable coded data (BL + EL) 1221, a scalable coded data storage device 1202. To supply.
- the scalable encoded data storage device 1202 stores the scalable encoded data (BL + EL) 1221 supplied from the imaging device 1201 with quality according to the situation. For example, in the normal case, the scalable encoded data storage device 1202 extracts base layer data from the scalable encoded data (BL + EL) 1221, and the base layer scalable encoded data ( BL) 1222. On the other hand, for example, in the case of attention, the scalable encoded data storage device 1202 stores scalable encoded data (BL + EL) 1221 with high quality and a large amount of data.
- the scalable encoded data storage device 1202 can store an image with high image quality only when necessary, so that an increase in the amount of data can be achieved while suppressing a reduction in the value of the image due to image quality degradation. And the use efficiency of the storage area can be improved.
- the imaging device 1201 is a surveillance camera.
- the monitoring target for example, an intruder
- the content of the captured image is likely to be unimportant, so reduction of the data amount is given priority, and the image data (scalable coding) Data) is stored in low quality.
- the image quality is given priority and the image data (scalable) (Encoded data) is stored with high quality.
- whether it is normal time or attention time may be determined by the scalable encoded data storage device 1202 analyzing an image, for example.
- the imaging apparatus 1201 may make a determination, and the determination result may be transmitted to the scalable encoded data storage device 1202.
- the criterion for determining whether the time is normal or noting is arbitrary, and the content of the image as the criterion is arbitrary. Of course, conditions other than the contents of the image can also be used as the criterion. For example, it may be switched according to the volume or waveform of the recorded sound, may be switched at every predetermined time, or may be switched by an external instruction such as a user instruction.
- the number of states is arbitrary, for example, normal, slightly attention, attention, very attention, etc.
- three or more states may be switched.
- the upper limit number of states to be switched depends on the number of layers of scalable encoded data.
- the imaging apparatus 1201 may determine the number of layers for scalable coding according to the state. For example, in a normal case, the imaging apparatus 1201 may generate base layer scalable encoded data (BL) 1222 with low quality and a small amount of data, and supply the scalable encoded data storage apparatus 1202 to the scalable encoded data storage apparatus 1202. For example, when attention is paid, the imaging device 1201 generates scalable encoded data (BL + EL) 1221 having a high quality and a large amount of data, and supplies the scalable encoded data storage device 1202 to the scalable encoded data storage device 1202. May be.
- BL base layer scalable encoded data
- BL + EL scalable encoded data
- the monitoring camera has been described as an example.
- the use of the imaging system 1200 is arbitrary and is not limited to the monitoring camera.
- FIG. 28 illustrates an example of a schematic configuration of a video set to which the present disclosure is applied.
- the video set 1300 shown in FIG. 28 has such a multi-functional configuration, and a device having a function relating to image encoding and decoding (either or both of them) can be used for the function. It is a combination of devices having other related functions.
- the video set 1300 includes a module group such as a video module 1311, an external memory 1312, a power management module 1313, and a front-end module 1314, and an associated module 1321, a camera 1322, a sensor 1323, and the like. And a device having a function.
- a module is a component that has several functions that are related to each other and that has a coherent function.
- the specific physical configuration is arbitrary. For example, a plurality of processors each having a function, electronic circuit elements such as resistors and capacitors, and other devices arranged on a wiring board or the like can be considered. . It is also possible to combine the module with another module, a processor, or the like to form a new module.
- the video module 1311 is a combination of configurations having functions related to image processing, and includes an application processor, a video processor, a broadband modem 1333, and an RF module 1334.
- a processor is a configuration in which a configuration having a predetermined function is integrated on a semiconductor chip by a SoC (System On a Chip), and for example, there is a system LSI (Large Scale Integration).
- the configuration having the predetermined function may be a logic circuit (hardware configuration), a CPU, a ROM, a RAM, and the like, and a program (software configuration) executed using them. , Or a combination of both.
- a processor has a logic circuit and a CPU, ROM, RAM, etc., a part of the function is realized by a logic circuit (hardware configuration), and other functions are executed by the CPU (software configuration) It may be realized by.
- the 28 is a processor that executes an application related to image processing.
- the application executed in the application processor 1331 not only performs arithmetic processing to realize a predetermined function, but also can control the internal and external configurations of the video module 1311 such as the video processor 1332 as necessary. .
- the video processor 1332 is a processor having a function related to image encoding / decoding (one or both of them).
- the broadband modem 1333 is a processor (or module) that performs processing related to wired or wireless (or both) broadband communication performed via a broadband line such as the Internet or a public telephone line network.
- the broadband modem 1333 digitally modulates data to be transmitted (digital signal) to convert it into an analog signal, or demodulates the received analog signal to convert it into data (digital signal).
- the broadband modem 1333 can digitally modulate and demodulate arbitrary information such as image data processed by the video processor 1332, a stream obtained by encoding the image data, an application program, setting data, and the like.
- the RF module 1334 is a module that performs frequency conversion, modulation / demodulation, amplification, filter processing, and the like on an RF (Radio Frequency) signal transmitted / received via an antenna. For example, the RF module 1334 generates an RF signal by performing frequency conversion or the like on the baseband signal generated by the broadband modem 1333. Further, for example, the RF module 1334 generates a baseband signal by performing frequency conversion or the like on the RF signal received via the front end module 1314.
- RF Radio Frequency
- the application processor 1331 and the video processor 1332 may be integrated into a single processor.
- the external memory 1312 is a module that is provided outside the video module 1311 and has a storage device used by the video module 1311.
- the storage device of the external memory 1312 may be realized by any physical configuration, but is generally used for storing a large amount of data such as image data in units of frames. For example, it is desirable to realize it with a relatively inexpensive and large-capacity semiconductor memory such as DRAM (Dynamic Random Access Memory).
- the power management module 1313 manages and controls power supply to the video module 1311 (each component in the video module 1311).
- the front-end module 1314 is a module that provides the RF module 1334 with a front-end function (circuit on the transmitting / receiving end on the antenna side). As illustrated in FIG. 20, the front end module 1314 includes, for example, an antenna unit 1351, a filter 1352, and an amplification unit 1353.
- the antenna unit 1351 has an antenna for transmitting and receiving a radio signal and its peripheral configuration.
- the antenna unit 1351 transmits the signal supplied from the amplification unit 1353 as a radio signal, and supplies the received radio signal to the filter 1352 as an electric signal (RF signal).
- the filter 1352 performs a filtering process on the RF signal received via the antenna unit 1351 and supplies the processed RF signal to the RF module 1334.
- the amplifying unit 1353 amplifies the RF signal supplied from the RF module 1334 and supplies the amplified RF signal to the antenna unit 1351.
- Connectivity 1321 is a module having a function related to connection with the outside.
- the physical configuration of the connectivity 1321 is arbitrary.
- the connectivity 1321 has a configuration having a communication function other than the communication standard supported by the broadband modem 1333, an external input / output terminal, and the like.
- the communication 1321 is compliant with wireless communication standards such as Bluetooth (registered trademark), IEEE 802.11 (for example, Wi-Fi (Wireless Fidelity, registered trademark)), NFC (Near Field Communication), IrDA (InfraRed Data Association), etc. You may make it have a module which has a function, an antenna etc. which transmit / receive the signal based on the standard.
- the connectivity 1321 has a module having a communication function compliant with a wired communication standard such as USB (Universal Serial Bus), HDMI (registered trademark) (High-Definition Multimedia Interface), or a terminal compliant with the standard. You may do it.
- the connectivity 1321 may have other data (signal) transmission functions such as analog input / output terminals.
- the connectivity 1321 may include a data (signal) transmission destination device.
- the drive 1321 reads and writes data to and from a recording medium such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory (not only a removable medium drive, but also a hard disk, SSD (Solid State Drive) NAS (including Network Attached Storage) and the like.
- the connectivity 1321 may include an image or audio output device (a monitor, a speaker, or the like).
- the camera 1322 is a module having a function of capturing a subject and obtaining image data of the subject.
- Image data obtained by imaging by the camera 1322 is supplied to, for example, a video processor 1332 and encoded.
- the sensor 1323 includes, for example, a voice sensor, an ultrasonic sensor, an optical sensor, an illuminance sensor, an infrared sensor, an image sensor, a rotation sensor, an angle sensor, an angular velocity sensor, a velocity sensor, an acceleration sensor, an inclination sensor, a magnetic identification sensor, an impact sensor, It is a module having an arbitrary sensor function such as a temperature sensor.
- the data detected by the sensor 1323 is supplied to the application processor 1331 and used by an application or the like.
- the configuration described as a module in the above may be realized as a processor, or conversely, the configuration described as a processor may be realized as a module.
- the present disclosure can be applied to the video processor 1332 as described later. Accordingly, the video set 1300 can be implemented as a set to which the present disclosure is applied.
- FIG. 29 illustrates an example of a schematic configuration of a video processor 1332 (FIG. 28) to which the present disclosure is applied.
- the video processor 1332 receives the video signal and the audio signal, encodes them in a predetermined method, decodes the encoded video data and audio data, A function of reproducing and outputting an audio signal.
- the video processor 1332 includes a video input processing unit 1401, a first image enlargement / reduction unit 1402, a second image enlargement / reduction unit 1403, a video output processing unit 1404, a frame memory 1405, and a memory control unit 1406.
- the video processor 1332 includes an encoding / decoding engine 1407, video ES (ElementaryElementStream) buffers 1408A and 1408B, and audio ES buffers 1409A and 1409B.
- the video processor 1332 includes an audio encoder 1410, an audio decoder 1411, a multiplexing unit (MUX (Multiplexer)) 1412, a demultiplexing unit (DMUX (Demultiplexer)) 1413, and a stream buffer 1414.
- MUX Multiplexing unit
- DMUX demultiplexing unit
- the video input processing unit 1401 acquires, for example, a video signal input from the connectivity 1321 (FIG. 28) and converts it into digital image data.
- the first image enlargement / reduction unit 1402 performs format conversion, image enlargement / reduction processing, and the like on the image data.
- the second image enlargement / reduction unit 1403 performs image enlargement / reduction processing on the image data in accordance with the format of the output destination via the video output processing unit 1404, or is the same as the first image enlargement / reduction unit 1402. Format conversion and image enlargement / reduction processing.
- the video output processing unit 1404 performs format conversion, conversion to an analog signal, and the like on the image data, and outputs the reproduced video signal to, for example, the connectivity 1321 (FIG. 28).
- the frame memory 1405 is a memory for image data shared by the video input processing unit 1401, the first image scaling unit 1402, the second image scaling unit 1403, the video output processing unit 1404, and the encoding / decoding engine 1407. .
- the frame memory 1405 is realized as a semiconductor memory such as a DRAM, for example.
- the memory control unit 1406 receives the synchronization signal from the encoding / decoding engine 1407, and controls the write / read access to the frame memory 1405 according to the access schedule to the frame memory 1405 written in the access management table 1406A.
- the access management table 1406A is updated by the memory control unit 1406 in accordance with processing executed by the encoding / decoding engine 1407, the first image enlargement / reduction unit 1402, the second image enlargement / reduction unit 1403, and the like.
- the encoding / decoding engine 1407 performs encoding processing of image data and decoding processing of a video stream that is data obtained by encoding the image data. For example, the encoding / decoding engine 1407 encodes the image data read from the frame memory 1405 and sequentially writes the data as a video stream in the video ES buffer 1408A. Further, for example, the video stream is sequentially read from the video ES buffer 1408B, decoded, and sequentially written in the frame memory 1405 as image data.
- the encoding / decoding engine 1407 uses the frame memory 1405 as a work area in the encoding and decoding. Also, the encoding / decoding engine 1407 outputs a synchronization signal to the memory control unit 1406, for example, at a timing at which processing for each macroblock is started.
- the video ES buffer 1408A buffers the video stream generated by the encoding / decoding engine 1407 and supplies the buffered video stream to the multiplexing unit (MUX) 1412.
- the video ES buffer 1408B buffers the video stream supplied from the demultiplexer (DMUX) 1413 and supplies the buffered video stream to the encoding / decoding engine 1407.
- the audio ES buffer 1409A buffers the audio stream generated by the audio encoder 1410 and supplies the buffered audio stream to the multiplexing unit (MUX) 1412.
- the audio ES buffer 1409B buffers the audio stream supplied from the demultiplexer (DMUX) 1413 and supplies the buffered audio stream to the audio decoder 1411.
- the audio encoder 1410 converts, for example, an audio signal input from the connectivity 1321 (FIG. 28), for example, into a digital format, and encodes the audio signal according to a predetermined format such as an MPEG audio format or an AC3 (Audio Code number 3) format.
- the audio encoder 1410 sequentially writes an audio stream, which is data obtained by encoding an audio signal, in the audio ES buffer 1409A.
- the audio decoder 1411 decodes the audio stream supplied from the audio ES buffer 1409B, converts it into an analog signal, for example, and supplies it as a reproduced audio signal to, for example, the connectivity 1321 (FIG. 28).
- the multiplexing unit (MUX) 1412 multiplexes the video stream and the audio stream.
- the multiplexing method (that is, the format of the bit stream generated by multiplexing) is arbitrary.
- the multiplexing unit (MUX) 1412 can also add predetermined header information or the like to the bit stream. That is, the multiplexing unit (MUX) 1412 can convert the stream format by multiplexing. For example, the multiplexing unit (MUX) 1412 multiplexes the video stream and the audio stream to convert it into a transport stream that is a bit stream in a transfer format. Further, for example, the multiplexing unit (MUX) 1412 multiplexes the video stream and the audio stream, thereby converting the data into file format data (file data) for recording.
- the demultiplexing unit (DMUX) 1413 demultiplexes the bit stream in which the video stream and the audio stream are multiplexed by a method corresponding to the multiplexing by the multiplexing unit (MUX) 1412. That is, the demultiplexer (DMUX) 1413 extracts the video stream and the audio stream from the bit stream read from the stream buffer 1414 (separates the video stream and the audio stream). That is, the demultiplexer (DMUX) 1413 can convert the stream format by demultiplexing (inverse conversion of the conversion by the multiplexer (MUX) 1412). For example, the demultiplexer (DMUX) 1413 obtains a transport stream supplied from, for example, the connectivity 1321 and the broadband modem 1333 (both in FIG.
- the demultiplexer (DMUX) 1413 obtains file data read from various recording media by, for example, the connectivity 1321 (FIG. 28) via the stream buffer 1414, and demultiplexes it. It can be converted into a video stream and an audio stream.
- DMUX demultiplexer
- Stream buffer 1414 buffers the bit stream.
- the stream buffer 1414 buffers the transport stream supplied from the multiplexing unit (MUX) 1412 and, for example, at the predetermined timing or based on a request from the outside, for example, the connectivity 1321 or the broadband modem 1333 (whichever Are also supplied to FIG.
- MUX multiplexing unit
- the stream buffer 1414 buffers the file data supplied from the multiplexing unit (MUX) 1412, and, for example, in the connectivity 1321 (FIG. 28) or the like at a predetermined timing or based on a request from the outside. To be supplied and recorded on various recording media.
- MUX multiplexing unit
- the stream buffer 1414 buffers the transport stream acquired through, for example, the connectivity 1321 and the broadband modem 1333 (both of which are shown in FIG. 28), and performs reverse processing at a predetermined timing or based on an external request or the like.
- the data is supplied to a multiplexing unit (DMUX) 1413.
- DMUX multiplexing unit
- the stream buffer 1414 buffers file data read from various recording media, for example, in the connectivity 1321 (FIG. 28), and the demultiplexing unit at a predetermined timing or based on an external request or the like. (DMUX) 1413.
- a video signal input from the connectivity 1321 (FIG. 28) or the like to the video processor 1332 is converted into digital image data of a predetermined format such as 4: 2: 2Y / Cb / Cr format by the video input processing unit 1401.
- the data is sequentially written into the frame memory 1405.
- This digital image data is read by the first image enlargement / reduction unit 1402 or the second image enlargement / reduction unit 1403, and format conversion to a predetermined method such as 4: 2: 0Y / Cb / Cr method and enlargement / reduction processing are performed. Is written again in the frame memory 1405.
- This image data is encoded by the encoding / decoding engine 1407 and written as a video stream in the video ES buffer 1408A.
- an audio signal input from the connectivity 1321 (FIG. 28) or the like to the video processor 1332 is encoded by the audio encoder 1410 and written as an audio stream in the audio ES buffer 1409A.
- the video stream of the video ES buffer 1408A and the audio stream of the audio ES buffer 1409A are read and multiplexed by the multiplexing unit (MUX) 1412 and converted into a transport stream, file data, or the like.
- the transport stream generated by the multiplexing unit (MUX) 1412 is buffered in the stream buffer 1414 and then output to the external network via, for example, the connectivity 1321 or the broadband modem 1333 (both of which are shown in FIG. 28).
- the file data generated by the multiplexing unit (MUX) 1412 is buffered in the stream buffer 1414, and then output to, for example, the connectivity 1321 (FIG. 28) or the like and recorded on various recording media.
- a transport stream input from an external network to the video processor 1332 via the connectivity 1321 or the broadband modem 1333 (both in FIG. 28) is buffered in the stream buffer 1414 and then demultiplexed (DMUX) 1413 is demultiplexed.
- DMUX demultiplexed
- file data read from various recording media in the connectivity 1321 (FIG. 28) and input to the video processor 1332 is buffered in the stream buffer 1414 and then demultiplexed by the demultiplexer (DMUX) 1413. It becomes. That is, the transport stream or file data input to the video processor 1332 is separated into a video stream and an audio stream by the demultiplexer (DMUX) 1413.
- the audio stream is supplied to the audio decoder 1411 via the audio ES buffer 1409B and decoded to reproduce the audio signal.
- the video stream is written to the video ES buffer 1408B, and then sequentially read and decoded by the encoding / decoding engine 1407, and written to the frame memory 1405.
- the decoded image data is enlarged / reduced by the second image enlargement / reduction unit 1403 and written to the frame memory 1405.
- the decoded image data is read out to the video output processing unit 1404, format-converted to a predetermined system such as 4: 2: 2Y / Cb / Cr system, and further converted into an analog signal to be converted into a video signal. Is played out.
- the present disclosure (that is, the functions of the image encoding device and the image decoding device according to each embodiment described above) may be realized by hardware such as a logic circuit, It may be realized by software such as an embedded program, or may be realized by both of them.
- an attribute for designating one or more URIs of definition information for remotely controlling the network camera 11 is added to the SegmentTemplate element disclosed in Non-Patent Document 1. to add.
- the segment request URL can be dynamically specified by the content reproduction apparatus 20 as a client.
- the embodiment of the present disclosure can compress MPD as compared with the method proposed in Non-Patent Document 2 above.
- the definition information (Parameter ⁇ ⁇ Description) from the MPD
- one embodiment of the present disclosure has an MPD schema according to the application requirements as compared with the method proposed in Non-Patent Document 2 above. Maintenance load at the time of individual expansion can be reduced.
- a storage medium storing the computer program is also provided.
- this technique can also take the following structures. (1) For every element acquired by a request to the server for accessing each of a plurality of sub-segments constituting each encoded data obtained by encoding the same content at different bit rates or described in MPD A storage unit for storing a predetermined definition for adding a parameter to access information for access; A communication unit that specifies and accesses a parameter based on a predetermined definition stored in the storage unit at an address defined in the access information; With The address defined in the access information is an information processing apparatus that is an address for controlling an apparatus that captures the content.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Databases & Information Systems (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Information Transfer Between Computers (AREA)
- Telephonic Communication Services (AREA)
- Facsimiles In General (AREA)
Abstract
Description
<1.コンテンツ再生システムの概要>
<2.コンテンツサーバ10の構成>
<3.コンテンツ再生装置20の構成>
<4.コンテンツサーバ13の構成>
<5.MPDの構成>
<6.まとめ>
まず、図1~図3を参照し、本開示の実施形態によるコンテンツ再生システムについて概略的に説明する。
以上、図1~図3を参照し、本開示の実施形態によるコンテンツ再生システムの概要を説明した。続いて、図4を参照し、本実施形態によるコンテンツサーバ10の構成を説明する。
以上、本実施形態によるコンテンツサーバ10の構成を説明した。続いて、図5を参照し、本実施形態によるコンテンツ再生装置20の構成を説明する。
図6は、コンテンツサーバ13の構成例を示す説明図である。図6に示したように、本実施形態によるコンテンツサーバ13は、記憶部310と、通信部320と、を備える。
続いて、本実施形態によるMPDの構成について説明する。上記非特許文献2では、ネットワークカメラをクライアントから制御するために、上記非特許文献1のsection 5.3.9.4.2で内容が公開されているSegmentTemplate要素に、パン角度の属性「pan_range」、チルト角度の属性「tilt_range」、ズーム倍率の属性「zoom_range」を指定する方法が提案されている。
Descriptionの本体を、コンテンツ再生装置20に応答する(ステップS113)。
(本開示を適用したコンピュータの説明)
上述した一連の処理は、ハードウエアにより実行することもできるし、ソフトウエアにより実行することもできる。一連の処理をソフトウエアにより実行する場合には、そのフトウエアを構成するプログラムが、コンピュータにインストールされる。ここで、コンピュータには、専用のハードウエアに組み込まれているコンピュータや、各種のプログラムをインストールすることで、各種の機能を実行することが可能な、例えば汎用のパーソナルコンピュータなどが含まれる。
(多視点画像符号化・多視点画像復号への適用)
上述した一連の処理は、多視点画像符号化・多視点画像復号に適用することができる。図12は、多視点画像符号化方式の一例を示す。
図13は、上述した多視点画像符号化を行う多視点画像符号化装置を示す図である。図13に示されるように、多視点画像符号化装置600は、符号化部601、符号化部602、および多重化部603を有する。
図14は、上述した多視点画像復号を行う多視点画像復号装置を示す図である。図14に示されるように、多視点画像復号装置610は、逆多重化部611、復号部612、および復号部613を有する。
(階層画像符号化・階層画像復号への適用)
上述した一連の処理は、階層画像符号化・階層画像復号(スケーラブル符号化・スケーラブル復号)に適用することができる。図15は、階層画像符号化方式の一例を示す。
このような階層画像符号化・階層画像復号(スケーラブル符号化・スケーラブル復号)において、スケーラブル(scalable)機能を有するパラメータは、任意である。例えば、図16に示されるような空間解像度をそのパラメータとしてもよい(spatial scalability)。このスペーシャルスケーラビリティ(spatial scalability)の場合、レイヤ毎に画像の解像度が異なる。つまり、この場合、図16に示されるように、各ピクチャが、元の画像より空間的に低解像度のベースレイヤと、ベースレイヤと合成することにより元の空間解像度が得られるエンハンスメントレイヤの2階層に階層化される。もちろん、この階層数は一例であり、任意の階層数に階層化することができる。
図19は、上述した階層画像符号化を行う階層画像符号化装置を示す図である。図19に示されるように、階層画像符号化装置620は、符号化部621、符号化部622、および多重化部623を有する。
図20は、上述した階層画像復号を行う階層画像復号装置を示す図である。図20に示されるように、階層画像復号装置630は、逆多重化部631、復号部632、および復号部633を有する。
(テレビジョン装置の構成例)
図21は、本開示を適用したテレビジョン装置の概略構成を例示している。テレビジョン装置900は、アンテナ901、チューナ902、デマルチプレクサ903、デコーダ904、映像信号処理部905、表示部906、音声信号処理部907、スピーカ908、外部インタフェース部909を有している。さらに、テレビジョン装置900は、制御部910、ユーザインタフェース部911等を有している。
(携帯電話機の構成例)
図22は、本開示を適用した携帯電話機の概略構成を例示している。携帯電話機920は、通信部922、音声コーデック923、カメラ部926、画像処理部927、多重分離部928、記録再生部929、表示部930、制御部931を有している。これらは、バス933を介して互いに接続されている。
(記録再生装置の構成例)
図23は、本開示を適用した記録再生装置の概略構成を例示している。記録再生装置940は、例えば受信した放送番組のオーディオデータとビデオデータを、記録媒体に記録して、その記録されたデータをユーザの指示に応じたタイミングでユーザに提供する。また、記録再生装置940は、例えば他の装置からオーディオデータやビデオデータを取得し、それらを記録媒体に記録させることもできる。さらに、記録再生装置940は、記録媒体に記録されているオーディオデータやビデオデータを復号して出力することで、モニタ装置等において画像表示や音声出力を行うことができるようにする。
(撮像装置の構成例)
図24は、本開示を適用した撮像装置の概略構成を例示している。撮像装置960は、被写体を撮像し、被写体の画像を表示部に表示させたり、それを画像データとして、記録媒体に記録する。
(第1のシステム)
次に、スケーラブル符号化(階層符号化)されたスケーラブル符号化データの具体的な利用例について説明する。スケーラブル符号化は、例えば、図25に示される例のように、伝送するデータの選択のために利用される。
また、スケーラブル符号化は、例えば、図26に示される例のように、複数の通信媒体を介する伝送のために利用される。
また、スケーラブル符号化は、例えば、図27に示される例のように、符号化データの記憶に利用される。
(実施のその他の例)
以上において本開示を適用する装置やシステム等の例を説明したが、本開示は、これに限らず、このような装置またはシステムを構成する装置に搭載するあらゆる構成、例えば、システムLSI(Large Scale Integration)等としてのプロセッサ、複数のプロセッサ等を用いるモジュール、複数のモジュール等を用いるユニット、ユニットにさらにその他の機能を付加したセット等(すなわち、装置の一部の構成)として実施することもできる。
本開示をセットとして実施する場合の例について、図28を参照して説明する。図28は、本開示を適用したビデオセットの概略的な構成の一例を示している。
図29は、本開示を適用したビデオプロセッサ1332(図28)の概略的な構成の一例を示している。
以上説明したように本開示の一実施形態では、非特許文献1で公開されているSegmentTemplate要素に、ネットワークカメラ11を遠隔制御するための定義情報のURIを1つまたは複数指定するための属性を追加する。このようにSegmentTemplate要素に属性を追加することで、クライアントであるコンテンツ再生装置20で動的にセグメントのリクエストURLを指定できる。
(1)
同一のコンテンツを異なるビットレートで符号化して得られる各符号化データを構成する複数のサブセグメントの各々にアクセスするための、またはMPDに記載されている、サーバへのリクエストにより取得するあらゆる要素にアクセスするためのアクセス情報にパラメータを付加するための所定の定義を記憶する記憶部と、
前記アクセス情報で定義されているアドレスに前記記憶部が記憶する所定の定義に基づきパラメータを指定してアクセスする通信部と、
を備え、
前記アクセス情報で定義されているアドレスは、前記コンテンツを撮像する装置を制御するためのアドレスである、情報処理装置。
(2)
同一のコンテンツを異なるビットレートで符号化して得られる各符号化データを構成する複数のサブセグメントの各々にアクセスするための、またはMPDに記載されている、サーバへのリクエストにより取得するあらゆる要素にアクセスするためのアクセス情報にパラメータを付加するための所定の定義を記憶することと、
前記アクセス情報で定義されているアドレスに前記記憶部が記憶する所定の定義に基づきパラメータを指定してアクセスすることと、
を備え、
前記アクセス情報で定義されているアドレスは、前記コンテンツを撮像する装置を制御するためのアドレスである、コンテンツ要求方法。
(3)
コンピュータに、
同一のコンテンツを異なるビットレートで符号化して得られる各符号化データを構成する複数のサブセグメントの各々にアクセスするための、またはMPDに記載されている、サーバへのリクエストにより取得するあらゆる要素にアクセスするためのアクセス情報にパラメータを付加するための所定の定義を記憶することと、
前記アクセス情報で定義されているアドレスに前記記憶部が記憶する所定の定義に基づきパラメータを指定してアクセスすることと、
を実行させ、
前記アクセス情報で定義されているアドレスは、前記コンテンツを撮像する装置を制御するためのアドレスである、コンピュータプログラム。
11 ネットワークカメラ
12 ネットワーク
20 コンテンツ再生装置
120 ファイル生成部
122 エンコーダ
130 記憶部
140 通信部
220 通信部
230 記憶部
240 再生部
250 選択部
Claims (3)
- 同一のコンテンツを異なるビットレートで符号化して得られる各符号化データを構成する複数のサブセグメントの各々にアクセスするための、またはMPDに記載されている、サーバへのリクエストにより取得するあらゆる要素にアクセスするためのアクセス情報にパラメータを付加するための所定の定義を記憶する記憶部と、
前記アクセス情報で定義されているアドレスに前記記憶部が記憶する所定の定義に基づきパラメータを指定してアクセスする通信部と、
を備え、
前記アクセス情報で定義されているアドレスは、前記コンテンツを撮像する装置を制御するためのアドレスである、情報処理装置。 - 同一のコンテンツを異なるビットレートで符号化して得られる各符号化データを構成する複数のサブセグメントの各々にアクセスするための、またはMPDに記載されている、サーバへのリクエストにより取得するあらゆる要素にアクセスするためのアクセス情報にパラメータを付加するための所定の定義を記憶することと、
前記アクセス情報で定義されているアドレスに前記記憶部が記憶する所定の定義に基づきパラメータを指定してアクセスすることと、
を備え、
前記アクセス情報で定義されているアドレスは、前記コンテンツを撮像する装置を制御するためのアドレスである、コンテンツ要求方法。 - コンピュータに、
同一のコンテンツを異なるビットレートで符号化して得られる各符号化データを構成する複数のサブセグメントの各々にアクセスするための、またはMPDに記載されている、サーバへのリクエストにより取得するあらゆる要素にアクセスするためのアクセス情報にパラメータを付加するための所定の定義を記憶することと、
前記アクセス情報で定義されているアドレスに前記記憶部が記憶する所定の定義に基づきパラメータを指定してアクセスすることと、
を実行させ、
前記アクセス情報で定義されているアドレスは、前記コンテンツを撮像する装置を制御するためのアドレスである、コンピュータプログラム。
Priority Applications (14)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AU2014254809A AU2014254809B2 (en) | 2013-04-19 | 2014-04-16 | Information processing apparatus, content requesting method, and computer program |
KR1020157026702A KR102177605B1 (ko) | 2013-04-19 | 2014-04-16 | 정보 처리 장치, 콘텐츠 요구 방법 및 컴퓨터 판독가능 저장 매체 |
MX2015014186A MX354254B (es) | 2013-04-19 | 2014-04-16 | Dispositivo de procesamiento de la informacion, metodo de solicitar el contenido y programa de computadora. |
JP2015512498A JP6481206B2 (ja) | 2013-04-19 | 2014-04-16 | 情報処理装置、コンテンツ要求方法およびコンピュータプログラム |
RU2015143204A RU2652789C2 (ru) | 2013-04-19 | 2014-04-16 | Устройство обработки информации, способ запроса содержания и компьютерная программа |
EP14784676.0A EP2988521A4 (en) | 2013-04-19 | 2014-04-16 | INFORMATION PROCESSING DEVICE, CONTENT PROCESSING AND COMPUTER PROGRAM |
CN201480020470.2A CN105122828B (zh) | 2013-04-19 | 2014-04-16 | 信息处理装置、内容请求方法以及计算机程序 |
US14/780,054 US20160057509A1 (en) | 2013-04-19 | 2014-04-16 | Information processing device, content requesting method, and computer program |
SG11201508375VA SG11201508375VA (en) | 2013-04-19 | 2014-04-16 | Information processing apparatus, content requesting method, and computer program |
BR112015025883A BR112015025883A8 (pt) | 2013-04-19 | 2014-04-16 | dispositivo de processamento de informação, método de processamento de informação, e, mídia legível por computador não-transitória |
EP18155847.9A EP3349476A1 (en) | 2013-04-19 | 2014-04-16 | Information processing device, content requesting method, and computer program |
CA2908163A CA2908163A1 (en) | 2013-04-19 | 2014-04-16 | Information processing device, content requesting method, and computer program |
AU2017228638A AU2017228638B2 (en) | 2013-04-19 | 2017-09-14 | Information processing device, content requesting method, and computer program |
US15/938,319 US20180220204A1 (en) | 2013-04-19 | 2018-03-28 | Information processing device, content requesting method, and computer program |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2013-088875 | 2013-04-19 | ||
JP2013088875 | 2013-04-19 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/780,054 A-371-Of-International US20160057509A1 (en) | 2013-04-19 | 2014-04-16 | Information processing device, content requesting method, and computer program |
US15/938,319 Continuation US20180220204A1 (en) | 2013-04-19 | 2018-03-28 | Information processing device, content requesting method, and computer program |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014171474A1 true WO2014171474A1 (ja) | 2014-10-23 |
Family
ID=51731411
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2014/060804 WO2014171474A1 (ja) | 2013-04-19 | 2014-04-16 | 情報処理装置、コンテンツ要求方法およびコンピュータプログラム |
Country Status (13)
Country | Link |
---|---|
US (2) | US20160057509A1 (ja) |
EP (2) | EP2988521A4 (ja) |
JP (2) | JP6481206B2 (ja) |
KR (1) | KR102177605B1 (ja) |
CN (2) | CN110087093A (ja) |
AU (2) | AU2014254809B2 (ja) |
BR (1) | BR112015025883A8 (ja) |
CA (1) | CA2908163A1 (ja) |
MX (1) | MX354254B (ja) |
MY (1) | MY177461A (ja) |
RU (2) | RU2755145C2 (ja) |
SG (3) | SG10201801143XA (ja) |
WO (1) | WO2014171474A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2017139628A (ja) * | 2016-02-03 | 2017-08-10 | キヤノン株式会社 | 通信装置、通信システム、通信制御方法およびコンピュータプログラム |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6565922B2 (ja) | 2014-10-10 | 2019-08-28 | ソニー株式会社 | 符号化装置および方法、再生装置および方法、並びにプログラム |
EP3220649A4 (en) * | 2014-11-13 | 2018-06-20 | LG Electronics Inc. | Broadcasting signal transmission device, broadcasting signal reception device, broadcasting signal transmission method, and broadcasting signal reception method |
US11442760B2 (en) * | 2016-07-01 | 2022-09-13 | Intel Corporation | Aperture access processors, methods, systems, and instructions |
CN108668179B (zh) * | 2017-03-27 | 2021-05-14 | 华为技术有限公司 | 媒体索引文件的传输方法及相关设备 |
CN111294658A (zh) * | 2018-12-07 | 2020-06-16 | 北京字节跳动网络技术有限公司 | 信息处理方法和装置 |
US11470136B2 (en) * | 2020-10-07 | 2022-10-11 | Tencent America LLC | URL customization using the session-based dash operations |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2012009454A1 (en) * | 2010-07-13 | 2012-01-19 | Qualcomm Incorporated | Video switching for streaming video data |
WO2012018951A1 (en) * | 2010-08-05 | 2012-02-09 | Qualcomm Incorporated | Signaling attributes for network-streamed video data |
Family Cites Families (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7941554B2 (en) * | 2003-08-01 | 2011-05-10 | Microsoft Corporation | Sparse caching for streaming media |
US9380096B2 (en) * | 2006-06-09 | 2016-06-28 | Qualcomm Incorporated | Enhanced block-request streaming system for handling low-latency streaming |
US8176525B2 (en) * | 2006-09-29 | 2012-05-08 | Rockstar Bidco, L.P. | Method and system for trusted contextual communications |
US8219711B2 (en) * | 2008-11-24 | 2012-07-10 | Juniper Networks, Inc. | Dynamic variable rate media delivery system |
CN102055773B (zh) * | 2009-11-09 | 2013-10-09 | 华为技术有限公司 | 实现基于http的流媒体业务的方法、系统和网络设备 |
CN102055717B (zh) * | 2009-11-09 | 2014-08-13 | 华为技术有限公司 | 快速播放的方法、终端及服务器 |
WO2011070552A1 (en) * | 2009-12-11 | 2011-06-16 | Nokia Corporation | Apparatus and methods for describing and timing representations in streaming media files |
CN102714662B (zh) * | 2010-01-18 | 2017-06-09 | 瑞典爱立信有限公司 | 用于http媒体流分发的方法和装置 |
KR20120034550A (ko) * | 2010-07-20 | 2012-04-12 | 한국전자통신연구원 | 스트리밍 컨텐츠 제공 장치 및 방법 |
EP3742741A1 (en) * | 2010-09-01 | 2020-11-25 | Ideahub Inc. | Method and device for providing content via http adaptive streaming using a media presentation description including a group element |
US8849950B2 (en) * | 2011-04-07 | 2014-09-30 | Qualcomm Incorporated | Network streaming of video data using byte range requests |
WO2013089437A1 (ko) * | 2011-12-12 | 2013-06-20 | 엘지전자 주식회사 | 미디어 컨텐트를 수신하는 장치 및 방법 |
WO2014008637A1 (en) * | 2012-07-11 | 2014-01-16 | Nokia Corporation | Method and apparatus for interacting with a media presentation description that describes a summary media presentation and an original media presentation |
US9143543B2 (en) * | 2012-11-30 | 2015-09-22 | Google Technology Holdings LLC | Method and system for multi-streaming multimedia data |
US9396126B2 (en) * | 2013-01-30 | 2016-07-19 | Google Inc. | Clearing an application cache |
US8869218B2 (en) * | 2013-03-15 | 2014-10-21 | Wowza Media Systems, LLC | On the fly transcoding of video on demand content for adaptive streaming |
-
2014
- 2014-04-16 RU RU2018109258A patent/RU2755145C2/ru active
- 2014-04-16 CN CN201910207940.0A patent/CN110087093A/zh active Pending
- 2014-04-16 SG SG10201801143XA patent/SG10201801143XA/en unknown
- 2014-04-16 EP EP14784676.0A patent/EP2988521A4/en not_active Ceased
- 2014-04-16 SG SG11201508375VA patent/SG11201508375VA/en unknown
- 2014-04-16 BR BR112015025883A patent/BR112015025883A8/pt not_active Application Discontinuation
- 2014-04-16 MY MYPI2015703576A patent/MY177461A/en unknown
- 2014-04-16 SG SG10201909965R patent/SG10201909965RA/en unknown
- 2014-04-16 US US14/780,054 patent/US20160057509A1/en not_active Abandoned
- 2014-04-16 AU AU2014254809A patent/AU2014254809B2/en not_active Ceased
- 2014-04-16 RU RU2015143204A patent/RU2652789C2/ru active
- 2014-04-16 MX MX2015014186A patent/MX354254B/es active IP Right Grant
- 2014-04-16 EP EP18155847.9A patent/EP3349476A1/en not_active Ceased
- 2014-04-16 KR KR1020157026702A patent/KR102177605B1/ko active IP Right Grant
- 2014-04-16 CN CN201480020470.2A patent/CN105122828B/zh not_active Expired - Fee Related
- 2014-04-16 WO PCT/JP2014/060804 patent/WO2014171474A1/ja active Application Filing
- 2014-04-16 CA CA2908163A patent/CA2908163A1/en not_active Abandoned
- 2014-04-16 JP JP2015512498A patent/JP6481206B2/ja not_active Expired - Fee Related
-
2017
- 2017-09-14 AU AU2017228638A patent/AU2017228638B2/en not_active Ceased
-
2018
- 2018-03-28 US US15/938,319 patent/US20180220204A1/en not_active Abandoned
-
2019
- 2019-01-18 JP JP2019006720A patent/JP6708271B2/ja not_active Expired - Fee Related
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2012009454A1 (en) * | 2010-07-13 | 2012-01-19 | Qualcomm Incorporated | Video switching for streaming video data |
WO2012018951A1 (en) * | 2010-08-05 | 2012-02-09 | Qualcomm Incorporated | Signaling attributes for network-streamed video data |
Non-Patent Citations (5)
Title |
---|
"ISO/IEC 23009-1 Information technology - Dynamic adaptive streaming over HTTP (DASH", PART 1: MEDIA PRESENTATION DESCRIPTION AND SEGMENT FORMATS, pages 67 - 69,16-19, 114-115, XP055286909 * |
CHRISTOPHOR RICK: "MPEG DASH GAINS INDUSTRY SUPPORT AS "STREAMING MEDIA STANDARD"", REELSTO, XP055294106, Retrieved from the Internet <URL:http://www.reelseo.com/mpeg-dash-streaming-media-standard> * |
DLNA GUIDELINES, 2006 |
I.SODAGAR, IEEE MULTIMEDIA, vol. 18, no. ISSUE, 14 November 2011 (2011-11-14), pages 62 - 67, XP002717752 * |
OPEN IPTV FORUM, 2009 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2017139628A (ja) * | 2016-02-03 | 2017-08-10 | キヤノン株式会社 | 通信装置、通信システム、通信制御方法およびコンピュータプログラム |
Also Published As
Publication number | Publication date |
---|---|
EP2988521A4 (en) | 2016-12-21 |
RU2018109258A3 (ja) | 2021-07-06 |
JP6481206B2 (ja) | 2019-03-13 |
RU2015143204A (ru) | 2017-04-13 |
BR112015025883A8 (pt) | 2020-01-14 |
US20160057509A1 (en) | 2016-02-25 |
KR20150145228A (ko) | 2015-12-29 |
SG11201508375VA (en) | 2015-11-27 |
US20180220204A1 (en) | 2018-08-02 |
EP3349476A1 (en) | 2018-07-18 |
RU2652789C2 (ru) | 2018-05-03 |
AU2017228638B2 (en) | 2018-11-08 |
MX2015014186A (es) | 2015-12-11 |
AU2014254809A1 (en) | 2015-10-08 |
BR112015025883A2 (pt) | 2017-07-25 |
MX354254B (es) | 2018-02-20 |
RU2018109258A (ru) | 2019-02-26 |
CN105122828B (zh) | 2019-05-10 |
SG10201909965RA (en) | 2019-11-28 |
JP2019083555A (ja) | 2019-05-30 |
JPWO2014171474A1 (ja) | 2017-02-23 |
AU2014254809B2 (en) | 2017-07-13 |
AU2017228638A1 (en) | 2017-10-05 |
JP6708271B2 (ja) | 2020-06-10 |
SG10201801143XA (en) | 2018-04-27 |
RU2755145C2 (ru) | 2021-09-13 |
CA2908163A1 (en) | 2014-10-23 |
MY177461A (en) | 2020-09-16 |
CN110087093A (zh) | 2019-08-02 |
KR102177605B1 (ko) | 2020-11-11 |
EP2988521A1 (en) | 2016-02-24 |
CN105122828A (zh) | 2015-12-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6658931B2 (ja) | 情報処理装置および方法 | |
JP6708271B2 (ja) | 情報処理装置、コンテンツ要求方法およびコンピュータプログラム | |
RU2645284C2 (ru) | Устройство для обработки изображений и способ | |
JP6481205B2 (ja) | サーバ装置、クライアント装置、コンテンツ配信方法およびコンピュータプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14784676 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2015512498 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2908163 Country of ref document: CA Ref document number: 20157026702 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14780054 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2014784676 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2014254809 Country of ref document: AU Date of ref document: 20140416 Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: MX/A/2015/014186 Country of ref document: MX |
|
ENP | Entry into the national phase |
Ref document number: 2015143204 Country of ref document: RU Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112015025883 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 112015025883 Country of ref document: BR Kind code of ref document: A2 Effective date: 20151009 |