WO2018123078A1 - Monitoring camera system - Google Patents

Monitoring camera system Download PDF

Info

Publication number
WO2018123078A1
WO2018123078A1 PCT/JP2017/006251 JP2017006251W WO2018123078A1 WO 2018123078 A1 WO2018123078 A1 WO 2018123078A1 JP 2017006251 W JP2017006251 W JP 2017006251W WO 2018123078 A1 WO2018123078 A1 WO 2018123078A1
Authority
WO
WIPO (PCT)
Prior art keywords
server
video data
data
time
video
Prior art date
Application number
PCT/JP2017/006251
Other languages
French (fr)
Japanese (ja)
Inventor
昭弘 佐久間
Original Assignee
株式会社Nexpoint
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社Nexpoint filed Critical 株式会社Nexpoint
Publication of WO2018123078A1 publication Critical patent/WO2018123078A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/266Channel or content management, e.g. generation and management of keys and entitlement messages in a conditional access system, merging a VOD unicast channel into a multicast channel
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/91Television signal processing therefor
    • H04N5/915Television signal processing therefor for field- or frame-skip recording or reproducing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/91Television signal processing therefor
    • H04N5/92Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast

Definitions

  • the present invention relates to a surveillance camera system.
  • a surveillance camera For the purpose of crime prevention, investigation, or management, points to be monitored such as entrances and exits of buildings such as apartment buildings, storefronts, streets, factories, distribution centers, etc. are timely imaged by a surveillance camera, and the camera image is connected to an electric communication line such as the Internet.
  • a surveillance camera system that sends and monitors to a monitor of a user terminal (see, for example, JP-A-2002-77882).
  • JP-A-2002-77882 In the surveillance camera industry, since it is an essential condition to display real-time video, the analog configuration is still dominant even when the technology of the IP camera (network camera) has been developed.
  • the stream output from the IP camera is H.264. Limited to H.264 baseline profile. This is because the surveillance camera system does not allow a high-profile stream with a delay due to the nature of viewing in real time.
  • “high profile” means moving image data including bi-directional prediction inter frames (also expressed as future prediction frames, B frames, etc.). The high profile is optimized by referring to both the past frame and the future frame.
  • a surveillance camera system includes: A surveillance camera, A server connected to the surveillance camera via a network; With The surveillance camera is An imaging unit that captures the received video and generates raw video data; An encoding unit that encodes the raw video data into video data including a bi-directional prediction interframe created by referring to both past and future frames, and transmits the encoded video data to the server; Creating a real-time video data from the raw video data and transmitting it to the server; Have
  • FIG. 1 is a diagram illustrating a schematic configuration of a surveillance camera system according to an embodiment.
  • FIG. 2 is a diagram showing a schematic configuration of the surveillance camera in the surveillance camera system shown in FIG.
  • FIG. 3 is a block diagram showing a schematic internal configuration of the encoding unit of the surveillance camera shown in FIG.
  • FIG. 4 is a diagram for explaining output frame delay and rearrangement in the encoding unit shown in FIG. 3.
  • FIG. 5 is a diagram for explaining the principle operation of the encoding unit shown in FIG.
  • FIG. 6 is a block diagram illustrating an example of a hardware configuration of the encoding unit illustrated in FIG. 3.
  • FIG. 7 is a block diagram showing a modification of the hardware configuration of the encoding unit shown in FIG. FIG.
  • FIG. 8 is a diagram showing a schematic configuration of a server in the monitoring camera system shown in FIG.
  • FIG. 9 is a diagram showing a schematic configuration of a terminal device (mobile terminal) in the surveillance camera system shown in FIG.
  • FIG. 10 is a diagram showing a schematic configuration of a terminal device (browsing PC) in the surveillance camera system shown in FIG.
  • FIG. 11 is a schematic diagram showing an example of the operation of the surveillance camera system shown in FIG.
  • FIG. 12 is a diagram illustrating a schematic configuration of a monitoring camera according to a first modification of the embodiment.
  • FIG. 13 is a diagram illustrating a schematic configuration of a monitoring camera according to a second modification of the embodiment.
  • FIG. 14 is a diagram illustrating a schematic configuration of a surveillance camera according to a third modification of the embodiment.
  • the optimized data can be saved, but the conversion is performed using the CPU resource of the server serving as a recorder. Such costs tend to increase.
  • the embodiment described below has been made in consideration of the above points, and the purpose thereof is a surveillance camera system capable of acquiring real-time video, and is optimal without imposing a load on the server.
  • Another object of the present invention is to provide a surveillance camera system that can store data in an organized manner.
  • the surveillance camera system includes: A surveillance camera, A server connected to the surveillance camera via a network; With The surveillance camera is An imaging unit that captures the received video and generates raw video data; An encoding unit that encodes the raw video data into video data including a bi-directional prediction interframe created by referring to both past and future frames, and transmits the encoded video data to the server; Creating a real-time video data from the raw video data and transmitting it to the server; Have
  • the encoding unit of the surveillance camera encodes the raw video data into moving image data including a bi-directional prediction inter frame (B frame) and transmits the encoded video data to the server, which places a load on the server.
  • B frame bi-directional prediction inter frame
  • optimized data storage is possible.
  • the amount of data flowing on the network is significantly reduced, and network traffic can be reduced.
  • the load on the network design and equipment is reduced, and the life of the network equipment can be extended.
  • a plurality of cameras can be connected to the network.
  • encoding into moving image data including a bi-predictive inter frame (B frame) is not performed using the CPU resource of the server, but is performed inside the surveillance camera. Therefore, the load on the server is reduced and the cost for the system can be reduced. Moreover, since the load on the server is reduced, the number of surveillance cameras that can be accommodated per server can be increased.
  • the moving image data transmitted from the surveillance camera includes a bidirectional prediction inter frame (B frame), and thus is inevitably delayed, but the live video generation unit performs the delay. Since real-time video data is created from raw video data without data and transmitted to the server, the server can acquire real-time video, and the essential requirement of the surveillance camera system for real-time browsing is lost. It will not be.
  • B frame bidirectional prediction inter frame
  • a surveillance camera system is the surveillance camera system according to the first aspect,
  • the encoding unit transmits the moving image data to the server by a stream method.
  • a surveillance camera system is the surveillance camera system according to the first aspect,
  • the encoding unit creates divided data in order of shooting time from the moving image data in GOP units, stores the divided data in a memory, and transfers the divided data from the memory to the server one by one each time a divided data request is received from the server.
  • the server creates moving image data by combining divided data received from the monitoring camera in order of photographing time.
  • a surveillance camera system is the surveillance camera system according to any one of the first to third aspects, Each time the live video generation unit receives a live image request from the server, the live video generation unit cuts out real-time still image data from the raw video data and transmits it to the server.
  • a surveillance camera system is the surveillance camera system according to any one of the first to third aspects,
  • the live video generation unit encodes the raw video data into real-time video data not including the bidirectional prediction interframe, and transmits the encoded video data to the server in a stream format.
  • a surveillance camera system is the surveillance camera system according to any one of the first to third aspects,
  • the live video generation unit repeats cutting out real-time still image data from the raw video data and transmitting it to the server at a predetermined snapshot time interval.
  • a surveillance camera system is the surveillance camera system according to the fourth or sixth aspect,
  • the server A primary video storage unit for storing still image data received from the surveillance camera in a storage device;
  • a live image transmission unit that transmits the still image data as a live image to a terminal device connected via a network;
  • the terminal device acquires still image data at a time point that is a certain time past from the previously transmitted still image data from the storage device, and the terminal device as a rewind live image.
  • a rewind live image transmitter to transmit to When a fast-forward request from a rewind live image is received from the terminal device, still image data at a time point that is a certain time later than the previously transmitted still image data is acquired from the storage device until the current still image data is reached.
  • a fast-forward live image transmission unit for transmitting to the terminal device as a fast-forward live image;
  • the terminal device transmits a rewind request from the terminal device to the server. It is possible to check back in the past as it is, and by sending a fast-forward request from the terminal device to the server, it is possible to check from the past to the current time by fast-forward.
  • a surveillance camera system is the surveillance camera system according to any one of the first to seventh aspects,
  • the server includes a secondary video storage unit that stores the moving image data received from the monitoring camera in a storage device by adjusting the time according to the delay associated with the encoding by the encoding unit.
  • the moving image data transmitted from the surveillance camera is inevitably delayed by encoding in the encoding unit, but is stored in the storage device in a state adjusted in time according to the delay time. Therefore, when the server specifies the time from the terminal device and receives a reproduction request for past moving image data, the server can accurately extract the moving image data at the requested time from the storage device and transmit it to the terminal device. it can.
  • the surveillance camera is A surveillance camera connected to a server via a network, An imaging unit that captures the received video and generates raw video data; An encoding unit that encodes the raw video data into video data including a bi-directional prediction interframe created by referring to both past and future frames, and transmits the encoded video data to the server; Creating a real-time video data from the raw video data and transmitting it to the server; Have
  • the server is An imaging unit that captures the received video and generates raw video data;
  • An encoding unit that encodes the raw video data into video data including a bi-directional prediction interframe created by referring to both past and future frames, and transmits the encoded video data to the server; Creating a real-time video data from the raw video data and transmitting it to the server;
  • a server connected to a surveillance camera via a network,
  • a primary video storage unit that stores real-time video data received from the surveillance camera in a storage device;
  • a secondary video storage unit that saves the video data received from the monitoring camera in the storage device by adjusting the time according to the delay related to the encoding in the encoding unit; Is provided.
  • FIG. 1 is a diagram illustrating a schematic configuration of a surveillance camera system 10 according to an embodiment.
  • the surveillance camera system 10 includes surveillance cameras 400A and 400B, a server 100, and terminal devices 200A, 200B, and 300A to 300C.
  • the monitoring cameras 400A and 400B, the server 100, and the terminal devices 200A, 200B, and 300A to 300C are connected to each other via the network 500.
  • the network 500 may be either a wired line or a wireless line, and the type and form of the line are not limited.
  • the network 500 includes a communication network such as a LAN, the Internet, a Wi-Fi line, or a 3G / LTE line.
  • the monitoring cameras 400A and 400B every time the monitoring cameras 400A and 400B accept a live image request from the server 100, the monitoring cameras 400A and 400B create a real-time still image from the camera video and transmit it to the server 100.
  • the server 100 acquires and accumulates still images from the monitoring camera 400, and each time a browsing request is received from the terminal devices 200A, 200B, 300A to 300C, the still image for browsing is stored in the terminal devices 200A, 200B, 300A to 300C. Send.
  • the monitoring cameras 400A and 400B exist in the same LAN as the server 100.
  • the communication between the monitoring cameras 400A and 400B and the server 100 is a TCP / IP system, can reliably transmit and receive camera images, and has high reliability.
  • the terminal devices 200A, 200B, 300A to 300C are used by the user, and are mobile terminals such as smartphones and tablets, notebook computers or desktop computers.
  • the terminal devices 200A, 200B, and 300A to 300C may exist in the same LAN as the server 100. If the terminal devices 200A, 200B, and 300A to 300C are connected to the network 500 via the router 600, they are connected to the server 100 via the Internet. May be.
  • the monitoring cameras 400A and 400B may be collectively expressed by adding a reference numeral 400.
  • the terminal devices 200A and 200B as mobile terminals are denoted by reference numeral 200
  • the terminal devices 300A to 300C as viewing PCs are denoted by reference numeral 300 and collectively represented.
  • FIG. 2 is a diagram illustrating a schematic configuration of the monitoring camera 400.
  • the monitoring camera 400 includes an imaging unit 401, an encoding unit 402, a live video generation unit 403, and a communication control unit such as a network interface 404.
  • a communication control unit such as a network interface 404.
  • Each of these parts 401 to 404 is housed in the same housing 405.
  • the imaging unit 401 includes a lens 4011, a video data receiving unit 4012 that captures an image received through the lens 4011, and raw video data (also expressed as RAW data) based on the imaging result. And a RAW data generation unit 4013 for generating.
  • the configuration of the imaging unit 401 is the same as that of a conventional surveillance camera, and detailed description thereof is omitted.
  • the encoding unit 402 is connected to the imaging unit 401, and both of the raw video data (RAW data) generated by the imaging unit 401 are created with reference to both past frames and future frames. It encodes into the moving image data containing a direction prediction inter frame (B frame), and transmits to the server 100 via the network interface 404. In the present embodiment, the encoding unit 402 transmits moving image data to the server 100 by a stream method (also referred to as a streaming method).
  • a stream method also referred to as a streaming method
  • FIG. 3 is a block diagram showing a schematic internal configuration of the encoding unit 402.
  • the encoding unit 402 includes a control circuit 4021, an intra-frame compression block 4022, an inter-frame compression block 4023, a variable length code compression / frame rearrangement block 4024, and a memory 4025. ing.
  • the control circuit 4021 controls each block such as setting compression parameters.
  • the intra-frame compression block 4022 performs information compression (data compression within one screen) based on the spatial correlation within the frame.
  • the inter-frame compression block 4023 performs information compression based on temporal correlation with previous and subsequent frames.
  • the variable-length code compression / frame rearrangement block 4024 compresses information by assigning a short code to frequently occurring data (entropy coding) and rearranges the output order of the encoded frame data.
  • the memory 4025 accumulates a plurality of (5 frames in the illustrated example) frame data in order to generate a bidirectional prediction inter frame (B frame).
  • FIG. 4 is a diagram for explaining output frame delay and rearrangement in the encoding unit 402.
  • the encoding unit 402 includes a plurality of pieces of RAW data between an intra frame (I frame) that can be decoded independently and created at regular intervals, and an intra frame with reference to a past frame. It is encoded into moving image data including a prediction inter frame (P frame) to be created and a bidirectional prediction inter frame (B frame) created by referring to both the past frame and the future frame.
  • a bidirectional prediction inter frame (B frame) cannot be generated unless a prediction inter frame (P frame) is determined. Therefore, a high profile output including a bi-predictive interframe (B frame) has a longer delay time to output and a rearrangement of frames than a baseline output not including a bi-predictive interframe (B frame). is required. Since the delay time is long, the encoding unit 402 is provided with a memory 4025 that holds a frame for that period.
  • FIG. 5 is a diagram for explaining the principle operation of the encoding unit 402.
  • the encoding unit 402 performs intra-frame compression (spatial information compression) for the first frame data among the sequentially input RAW data (step S41).
  • the encoding unit 402 first decompresses the frames before and after the frame (step S42), and performs inter-frame compression (temporal information compression) with reference to them.
  • intra-frame compression (spatial information compression) is performed (step S41).
  • the encoding unit 402 performs compression (entropy encoding) and rearrangement on the frame data compressed in the frame based on the appearance frequency of the code and outputs the result (step S44).
  • FIG. 6 is a block diagram illustrating an example of a hardware configuration of the encoding unit 402.
  • the encoding unit 402 includes an encoding LSI 410 and interface blocks 411A and 411B.
  • the encoding LSI 410 is hardware that performs high-profile encoding, and includes a control circuit 4021, an intra-frame compression block 4022, an inter-frame compression block 4023, and a variable-length code compression / frame rearrangement block 4024 (see FIG. 3). Is included.
  • the encoding LSI 410 includes a memory 4025 inside.
  • the interface blocks 411A and 411B perform conversion when data input / output and LSI input / output are different.
  • the encode LSI 410 specifically, for example, H.S. H.264 translation compatible memory built-in transcoder LSI MB86M0x is used.
  • a stationary device equipped with a digital broadcast recording function such as a television, a hard disk recorder, or a personal computer, or It is used for mobile-related devices for smartphones and tablet products, and as far as the inventors know, there is no example used for surveillance cameras.
  • the memory 4025 is built in the encode LSI 410, but the present invention is not limited to this, and the memory 4025 may be externally attached to the encode LSI 410 as shown in FIG.
  • the hardware configuration of the encoding unit 402 is not limited to an aspect implemented using the encoding LSI 410 as shown in FIGS. 6 and 7, and may be an aspect implemented using a microcomputer and software.
  • the live video generation unit 403 is connected to the imaging unit 401, creates real-time video data from raw video data (RAW data) without delay, and transmits the real-time video data to the server 100.
  • the live video generation unit 403 every time a live image request is received from the server 100, creates real-time still image data from raw video data (RAW data) without delay (for example, as it is).
  • the live image is cut out and transmitted to the server 100 via the network interface 404.
  • “live image” means one still image representing a real-time camera image.
  • FIG. 8 is a diagram illustrating a schematic configuration of the server 100.
  • the server 100 includes a control / arithmetic unit having a CPU 101 and a device driver with a memory 102 as a cache memory, and a storage device 110 having a main storage device such as a DRAM and an auxiliary storage device such as a hard disk. And a communication control device such as a network interface 104, a display 103 as a display device, and an input device including a keyboard 105, a mouse 106, and the like.
  • a control / arithmetic unit having a CPU 101 and a device driver with a memory 102 as a cache memory
  • a storage device 110 having a main storage device such as a DRAM and an auxiliary storage device such as a hard disk.
  • a communication control device such as a network interface 104, a display 103 as a display device, and an input device including a keyboard 105, a mouse 106, and the like.
  • the storage device 110 stores a primary video folder 111, a secondary video folder 112, a program 113, an operating system 114, an authentication database, an environment setting folder (not shown), and the like.
  • real-time video data received from the monitoring camera 400 is stored.
  • real-time still image data received from the monitoring camera 400 is sequentially stored in the primary video folder 111.
  • moving image data received from the monitoring camera 400 is stored as one file for each predetermined storage unit (for example, 10 minutes).
  • the program 113 is normally stored in the auxiliary storage device of the storage device 110, and is loaded into the main storage device at the time of execution.
  • the authentication database stores IDs, passwords, port numbers and IP addresses of the respective monitoring cameras 400 and terminal devices 200 and 300, and individual identification numbers (UIDs) in the mobile terminal 200 having no IP address.
  • IIDs individual identification numbers
  • the server 100 since the server and the terminal device are integrated with each other in the present embodiment, the server 100 also serves as a terminal device that displays the camera video by itself, and for maintenance and management, the display 103 as a display device. And a keyboard 105 and a mouse 106 as input devices. If the server 109 does not require display of camera video, the terminal function as a display device may not be provided.
  • the server 100 implements a computer function that allows the CPU 101 to load the program 113 into the memory 102 and execute it, thereby acquiring the camera video from the monitoring camera 400 and sending the camera video to the terminal devices 200 and 300. To do.
  • the CPU 101 is an arithmetic processing device mounted on a normal computer, executes various programs, and performs various controls.
  • the server 100 may be a single server or a server group including a plurality of servers.
  • the secondary video folder 112 provided in a server different from the server that acquires the video data from the monitoring camera 400 as the storage destination of the video data after a certain period (for example, 24 hours) has elapsed. It is good.
  • the load on the server 100 can be reduced, and a larger number of cameras can be monitored on the same network.
  • the program 113 causes the computer to realize a secondary video storage function that saves the video data received from the monitoring camera 400 in the secondary video folder 112 by adjusting the time according to the delay related to the encoding in the encoding unit 402. It is a program.
  • the CPU 101 loads the program 113 to the memory 102 and executes it so that the moving image data received from the monitoring camera 400 is time-adjusted according to the delay related to the encoding in the encoding unit 402 and the secondary video. It functions as a secondary video storage unit stored in the folder 112.
  • the program 113 is a program for causing a computer to realize a primary video storage function for storing still image data received from the surveillance camera 400 in the primary video folder 111, for example, for a predetermined browsing period (for example, 24 hours). is there.
  • a primary video storage unit that stores still image data received from the monitoring camera 400 in the primary video folder 111, for example, for a predetermined viewing period, by the CPU 101 loading and executing the program 113 in the memory 102.
  • the still image data browsing period (storage period in the primary video folder 111) is not limited to a predetermined mode, and is set as appropriate according to the free capacity of the storage device 110 at that time. May be.
  • the program 113 has a camera video playback request receiving function for receiving camera video playback requests from the terminal devices 200 and 300 in the computer, and the latest still image data stored in the primary video folder 111 as a live image as a terminal.
  • This is a program for realizing a live image transmission function to be transmitted to the devices 200 and 300.
  • the CPU 101 receives a camera video playback request from the terminal devices 200 and 300 by the CPU 101 loading and executing the program 113 in the memory 102
  • the latest still image data stored in the primary video folder 111 is It functions as a live image transmission unit that transmits to the terminal devices 200 and 300 as a live image.
  • the program 113 when the program 113 receives a rewind request receiving function from the terminal devices 200 and 300 for receiving a rewind request from a live image, or when receiving a rewind request from a live image, the program 113 transmits the still image data transmitted last time.
  • a rewind live image transmission function for acquiring still image data of a certain time in the past from the primary video folder 111 and transmitting the still image data to the terminal devices 200 and 300 as a rewind live image.
  • a fast-forwarding request acceptance function that accepts a fast-forward request from a return live image, and when a fast-forward request from a rewind live image is accepted, the still image at a future point in time from the transmitted still image data until the current still image data is reached
  • Acquire image data from the primary video folder 111 as a fast-forward live image Is a program for realizing the fast-forward live image transmission function of transmitting to the terminal device 200 and 300.
  • the server 100 accepts a rewind request from the live image from the terminal devices 200 and 300, and then the past time from the previously transmitted still image data.
  • Still image data from the primary video folder 111 is transmitted from the primary video folder 111 and transmitted to the terminal devices 200 and 300 as a rewind live image. From the terminal devices 200 and 300, the rewind live image is transmitted from the rewind live image.
  • the fast-forwarding request is received, still image data at a time point that is a certain time later than the previously transmitted still image data is acquired from the primary video folder 111 until the current still image data is reached, and the terminal device 200 as a fast-forwarding live image is acquired. It functions as a fast-forward live image transmission unit that transmits to 300.
  • the program 113 extracts a past video reproduction reception function for accepting reproduction of past video data from the terminal devices 200 and 300 to the computer, extracts the received past video data from the secondary video folder 112, and extracts the extracted video data.
  • the program 113 causes the computer to realize a monitoring camera connection function for connecting to one or a plurality of monitoring cameras 400, and a terminal device connection function for connecting to one or a plurality of terminal devices 200 and 300. It is a program.
  • the connection between the monitoring camera 400 and the server 100 is performed using the TCP / IP method, and authentication is performed using the user ID and password set on the server 100 side in order to identify the monitoring camera 400 to be connected.
  • the connection between the terminal devices 200 and 300 and the server 100 is also performed by the TCP / IP method, and authentication is performed with a user ID and a password, and it is confirmed that the terminal devices 200 and 300 are terminal devices registered in the server 100. Then send the camera video.
  • the authentication is preferably performed by an authentication database on the server 100.
  • the IP address of the router 600 to be connected and the port number assigned to each are used. Identification is done. When connected via the Internet, the port number of the router 600 between the two is made “valid”. The mobile terminal 200 having no IP address uses a UID.
  • transmission of moving image data from the monitoring camera 400 to the server 100 is performed by TCP / IP-based RTSP (Real Streaming Protocol).
  • the moving image data is H.264.
  • H.264 is compressed with high profile and transmitted.
  • a high profile of a next generation codec such as H.265 may be used.
  • FIG. 9 is a diagram illustrating a schematic configuration of the terminal device 200.
  • the terminal device 200 includes a control / arithmetic apparatus including a CPU 201 with a memory 202 and a device driver, a storage device 210, a communication control device (not shown) that transmits and receives data, a display, and the like.
  • a display 204 as an apparatus and an output device (not shown) such as operation buttons or a touch panel are provided.
  • the storage device 210 stores an image display program 213 and an operating system 214.
  • the terminal device 200 is, for example, a mobile phone such as a smartphone, and the CPU 201 loads the program 213 into the memory 202 and executes it, thereby realizing a computer function capable of displaying the camera video transmitted from the server 100.
  • the CPU 201 is an arithmetic processing device mounted on a normal mobile terminal, executes various programs, and performs various controls.
  • FIG. 10 is a diagram illustrating a schematic configuration of the terminal device 300.
  • a terminal device 300 includes a control / arithmetic unit having a CPU 301 and a device driver with a memory 302, a main storage device such as a DRAM, a storage device 310 having an auxiliary storage device such as a hard disk, and a network.
  • a communication control device such as an interface 304, a display 303 as a display device, and output devices such as a keyboard 305 and a mouse 306 are included.
  • the storage device 310 stores an image display program 313 and an operating system 314.
  • the terminal device 300 is, for example, a desktop computer, a notebook computer, a tablet terminal, or the like.
  • the CPU 301 loads a program 313 into the memory 302 and executes it, so that the camera image transmitted from the server 100 can be displayed. Realize the function.
  • the CPU 301 is an arithmetic processing device mounted on a normal PC, executes various programs, and performs various controls.
  • the programs 213 and 313 for displaying images are a server connection function for connecting the server 100 to the computer, and a camera that requests the server 100 to play back the camera video related to one or a plurality of monitoring cameras 400 having the monitoring authority.
  • This is a program for realizing a video playback request function and a camera video display function for displaying a camera video transmitted from the server 100.
  • the server 100 identifies the connection partner by connecting using the IP address and port number when starting the connection with the monitoring camera 400 in the local environment, and authenticates with the user ID and password.
  • the server 100 uses the port forwarding (also expressed as port forwarding or the like) of the router 600 to identify the connection partner using the global IP address and the port number. Since the server 100 identifies the terminal device 200 based on the terminal-specific information using the UID at the start of connection with the terminal device 200, the camera image can be displayed by authenticating the user ID and the password and matching the terminal-specific information.
  • the server 100 and the terminal device 300 as a viewing PC are both configured as personal computers and have a clock function or the like possessed by a normal personal computer.
  • the terminal device 200 as a mobile terminal and the monitoring camera 400 also have a clock function and the like.
  • FIG. 11 is a schematic diagram illustrating an example of the operation of the monitoring camera system 10.
  • the terminal device 200 as a mobile terminal is illustrated, but the operation is the same even if the terminal device 300 is replaced with a browsing PC.
  • the server 100 requests the monitoring camera 400 to connect using the TCP / IP method.
  • the monitoring camera 400 includes one or more monitoring cameras, and the monitoring camera 400 and the server 100 authenticate with a user ID and a password.
  • the authentication database of the server 100 is referred to and the authentication is successful, the monitoring camera 400 and the server 100 are connected.
  • the server 100 requests the monitoring camera 400 to capture and transmit the camera video.
  • the monitoring camera 400 receives a video request from the server 100, it continuously captures the camera video, and converts the raw video data (RAW data) obtained by the imaging to H.264.
  • H.264 high profile video data is encoded.
  • the surveillance camera 400 creates a plurality of RAW data between an intra frame (I frame) that can be decoded independently and created at regular intervals, and an intra frame with reference to a past frame. Encoding is performed into moving image data including a prediction inter frame (P frame) and a bidirectional prediction inter frame (B frame) created by referring to both past and future frames.
  • the monitoring camera 400 transmits the encoded moving image data to the server 100 by a stream method.
  • the server 100 repeatedly transmits a live image request to the monitoring camera 400 at a predetermined snapshot time interval.
  • the snapshot time interval is arbitrarily set in advance, and is, for example, 1/10 second, 1 second, 3 seconds or 60 seconds. In the embodiment described below, the snapshot time interval is 1 second.
  • the monitoring camera 400 creates (cuts out) real-time still image data from raw video data (RAW data) and transmits it to the server 100 as a live image.
  • RAW data real-time still image data from raw video data
  • the server 100 acquires the moving image data from the monitoring camera 400 by the stream method, the server 100 stores it in the secondary video folder 112 as one file for each storage unit (for example, 10 minutes). Thereby, the reproduction of the past moving image data can be accepted after the moving image data is stored (for example, after 10 minutes).
  • the server 100 saves the moving image data received from the monitoring camera 400 in the secondary video folder 112 by adjusting the time according to the delay related to the encoding in the encoding unit 402. For example, when the moving image data transmitted from the monitoring camera 400 is delayed by one minute with respect to the real-time still image data transmitted simultaneously with the moving image data, the server 100, for example, 9: 51: 00-10 The moving image data received at 0:00:59 is saved in the secondary video folder 112 as one file with a time stamp of 9:55:00 to 9:59:59.
  • the primary video folder 111 is displayed as high-resolution data that may be frequently viewed during a predetermined browsing period (for example, 24 hours). When the viewing period is over, the primary video folder 111 is deleted.
  • the terminal device 200 accepts a connection request to the server 100 from the user.
  • a connection request to the server 100 is input to the terminal device 200
  • the terminal device 200 requests the server 100 to connect using the TCP / IP method.
  • connection may be made partially using a protocol unique to the mobile phone.
  • the terminal device 200 includes one or a plurality of terminal devices, and the terminal device 200 and the server 100 authenticate with a user ID and a password.
  • registration is performed by issuing an encryption key only for the first time, and a user ID and password are issued during registration. If the authentication is successful, the terminal device 200 and the server 100 are connected.
  • the terminal device 200 receives a camera list request from the user.
  • the terminal device 200 transmits the camera list request to the server 100.
  • the server 100 checks authority information about the monitoring camera 400 that can be monitored by the requested terminal device 200.
  • the server 100 transmits authorized camera list data to the terminal device 200.
  • the terminal device 200 displays the received camera list data on the display 203 and accepts a selection from the user regarding the monitoring camera to be played live.
  • the terminal device 200 When the terminal device 200 receives the selection of the monitoring camera from the user, the terminal device 200 requests the server 100 to perform live playback of the camera video for the selected monitoring camera. Upon receiving a live playback request, the server 100 transmits the latest still image data acquired from the selected monitoring camera 400 and stored in the primary video folder 111 to the terminal device 200 as a live image.
  • the terminal device 200 displays the still image data received from the server 100 on the display 203 as a live image.
  • the terminal device 200 repeats the request for live playback at the same time interval as the snapshot time interval in the monitoring camera 400 until the user deselects the monitoring camera or while other than live playback is selected. .
  • the terminal device 200 receives an input of a rewind request from the user while the live image is displayed.
  • the terminal device 200 requests a rewind image from the server 100.
  • the server 100 When the server 100 receives the rewind image request, the server 100 extracts still image data at a point in time that is a predetermined time earlier than the previously transmitted still image data from the primary video folder 111 and transmits it to the terminal device 200 as a rewind live image. To do.
  • the server 100 since the past time for a certain time is set to the past time for one second, and the live image is still image data acquired every second, when one rewind request is input for the first time, The previous still image data is transmitted to the terminal device 200.
  • the terminal device 200 While the rewind is selected, that is, while the rewind live image is being displayed, the terminal device 200 requests the rewind live image from the server 100 at regular intervals.
  • the fixed time at that time is a fixed time shorter than the request interval of the live image, for example, every 0.2 seconds. Therefore.
  • On the terminal device 200 a rewind live image created from still image data of the previous camera image by 1 second is displayed every 0.2 seconds.
  • the terminal device 200 accepts an input of a fast forward request from the user while displaying the rewind live image.
  • the terminal device 200 requests the server 100 for a fast forward live image.
  • the server 100 receives the fast-forward live image request, the server 100 extracts still image data at a time point that is a certain time later than the previously transmitted still image data from the primary video folder 111 and transmits it to the terminal device 200 as a fast-forward live image.
  • the future time is set to the future time of 1 second, and the live image is an image acquired every second. Therefore, when a fast-forward request is input for the first time, the rewind live image From the still image data that is currently displayed, still image data that is one second in the future is transmitted.
  • the terminal device 200 While the fast forward is selected, that is, while the fast forward live image is being displayed, the terminal device 200 requests the server 100 for a fast forward live image at regular intervals.
  • the fixed time at that time is a fixed time shorter than the request interval of the live image, for example, every 0.2 seconds. Accordingly, the fast-forward live image created from the still image data of the camera video after 1 second is displayed on the terminal device 200 every 0.2 seconds.
  • the fast-forward live image reaches the current image, the live image display is switched to, and the fast-forward is finished.
  • the server 100 extracts the received past moving image data from the secondary video folder 112 instead of the still image data, and transmits the extracted moving image data to the terminal device 200 for display.
  • the still image data is an image that is light enough to be continuously transmitted to the terminal device 200, and the moving image data is encoded with a high profile and the amount of data is extremely small while being a moving image. Light enough to send to.
  • both still image data and moving image data can be rewinded, played back after rewinding, and fast-forwarded, which is highly convenient.
  • the video data is stored in the baseline profile, so the video deteriorates due to the compression of the capacity, or in the case of the variable rate, the video is instantaneously moved. There was a limit to the mechanism itself, such as being disturbed.
  • the optimized data can be saved, but the conversion is performed using the CPU resource of the server serving as a recorder. There was a tendency for such costs to increase.
  • the encoding unit 402 of the monitoring camera 400 encodes raw video data into moving image data including a bidirectional prediction inter frame (B frame) and transmits the encoded video data to the server.
  • Optimized data storage is possible without imposing a load on the server.
  • the amount of data flowing on the network 500 is significantly reduced, and network traffic can be reduced.
  • the load on the network design and equipment is reduced, and the life of the network equipment can be extended.
  • a plurality of surveillance cameras 400 can be connected to the network 500.
  • encoding into moving image data including a bi-predictive inter frame (B frame) is not performed using the CPU resource of the server 100, but inside the surveillance camera 400.
  • B frame bi-predictive inter frame
  • the moving image data transmitted from the monitoring camera 400 includes a bidirectional prediction inter frame (B frame), and thus the live video generation unit 403 inevitably delays. Since the real-time video data is generated from the raw video data without delay and transmitted to the server 100, the server 100 can acquire the real-time video, and is essential as a surveillance camera system for browsing in real time. There is no loss of requirements.
  • the moving image data transmitted from the monitoring camera 400 is necessarily delayed due to encoding by the encoding unit 402, but the secondary video folder 112 is time-adjusted according to the delay. Therefore, when the server 100 specifies a time from the terminal device 200 and receives a playback request for past video data, the server 100 accurately stores the video data at the requested time from the secondary video folder 112. And transmitted to the terminal device 200.
  • a rewind request is transmitted from the terminal device 200 to the server 100.
  • the terminal device 200 can check back as it is, and by transmitting a fast-forward request from the terminal device 200 to the server 100, it is possible to confirm from the past to the present time by fast-forward.
  • FIG. 12 is a diagram illustrating a schematic configuration of a monitoring camera 400 according to the first modification.
  • the encoding unit 402B of the monitoring camera 400 converts raw video data (RAW data) generated by the imaging unit 401 into moving image data including B frames (for example, H.264 high profile video data), and then, in the GOP unit (for example, 1 second) in which the I frame is the first image from the video data, divided data is created in the order of shooting time and stored in the memory 4025. save.
  • RAW data raw video data
  • B frames for example, H.264 high profile video data
  • the GOP unit for example, 1 second
  • GOP Group Of Picture
  • I frames intra frames
  • P frames prediction interframes
  • B frames bidirectional prediction interframes
  • the encoded data is transferred from the memory 4025 to the server 100 one by one.
  • the server 100 repeatedly transmits the divided data request to the monitoring camera 400 and receives the divided data. Then, the server 100 combines the divided data received from the monitoring camera 400 in the order of shooting time, creates one file of moving image data for each storage unit (for example, 10 minutes), and stores it in the secondary video folder 112.
  • the server 100 when the server 100 is installed in a cloud environment and video data is transferred from the remote monitoring camera 400 to the server 100, the original video can be restored later. Since it can be divided into smaller units and transmitted while avoiding congestion of the network 500, it is possible to collect camera images stably in a cloud environment.
  • FIG. 13 is a diagram illustrating a schematic configuration of a monitoring camera 400 according to the second modification.
  • the live video generation unit 403B of the monitoring camera 400 converts the raw video data (RAW data) into real-time data that does not include a bidirectional prediction interframe (B frame). From moving image data, that is, an intra frame (I frame) that can be decoded independently and created at regular intervals, and a plurality of predicted inter frames (P frames) created between intra frames with reference to past frames
  • the encoded video data of the baseline profile is encoded and transmitted to the server 100 in a stream format.
  • the encoding unit 402 of the surveillance camera 400 encodes raw video data into moving image data including a bi-directional prediction interframe (B frame). Because the data is transmitted to 100, optimized data storage can be performed without imposing a load on the server 100, and the live video generation unit 403 creates real-time video data from unprocessed video data without delay, and the server Therefore, the server 100 can acquire a real-time video, and the essential requirement as a surveillance camera system for real-time browsing is not impaired.
  • B frame bi-directional prediction interframe
  • the monitoring camera 400 since the monitoring camera 400 transmits real-time video in a stream format, traffic on the network 500 may increase as compared with the above-described embodiment, but it may be in a local environment. Will not be a problem.
  • FIG. 14 is a diagram illustrating a schematic configuration of a monitoring camera 400 according to the third modification.
  • the live video generation unit 403C of the monitoring camera 400 performs unprocessed video at a snapshot time interval (for example, every second) set in advance by the timer 409 or the like. It repeats that real-time still image data is cut out from data (RAW data) and transmitted to the server 100.
  • a TCP / IP-based transfer method such as FTP or HTTP is used.
  • the encoding unit 402 of the surveillance camera 400 encodes raw video data into moving image data including a bi-directional prediction interframe (B frame). Because the data is transmitted to 100, optimized data storage can be performed without imposing a load on the server 100, and the live video generation unit 403 creates real-time video data from unprocessed video data without delay, and the server Therefore, the server 100 can acquire a real-time video, and the essential requirement as a surveillance camera system for real-time browsing is not impaired.
  • B frame bi-directional prediction interframe

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Closed-Circuit Television Systems (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

A monitoring camera system comprises a monitoring camera and a server connected to the monitoring camera via a network. The monitoring camera comprises: an image capture unit that captures light-receiving images to generate unprocessed image data; an encode unit that encodes the unprocessed image data into moving image data including bidirectional prediction interframes produced by referring to both past and future frames and then transmits the moving image data to the server; and a live image generation unit that produces real-time image data from the unprocessed image data and transmits the real-time image data to the server.

Description

監視カメラシステムSurveillance camera system
 本発明は、監視カメラシステムに関する。 The present invention relates to a surveillance camera system.
 防犯や調査あるいは管理といった目的で、集合住宅などの建物の出入口や店頭あるいは街頭、工場や配送センターなど、監視対象となるポイントを監視カメラにより適時撮像し、カメラ映像をインターネットなどの電気通信回線を介して、ユーザ端末のモニタに送出し、監視する監視カメラシステムが知られている(たとえば、特開2002-77882号公報参照)。監視カメラの業界では、リアルタイムの映像を表示することが必須の条件とされているため、IPカメラ(ネットワークカメラ)の技術が発達してきた今でもなお、アナログ構成が優位となっている。 For the purpose of crime prevention, investigation, or management, points to be monitored such as entrances and exits of buildings such as apartment buildings, storefronts, streets, factories, distribution centers, etc. are timely imaged by a surveillance camera, and the camera image is connected to an electric communication line such as the Internet. There is known a surveillance camera system that sends and monitors to a monitor of a user terminal (see, for example, JP-A-2002-77882). In the surveillance camera industry, since it is an essential condition to display real-time video, the analog configuration is still dominant even when the technology of the IP camera (network camera) has been developed.
 従来の監視カメラシステムにおいてIPカメラを使用する場合には、IPカメラからのストリーム出力は、H.264のベースラインプロファイルに限定されている。これは、監視カメラシステムが、リアルタイムでの閲覧を主としている性格上、遅延の発生するハイプロファイルでのストリームを許容しないからである。なお、本明細書において「ハイプロファイル」とは、双方向予測インターフレーム(未来予測フレーム、Bフレームなどとも表現される)を含む動画データを意味する。ハイプロファイルは、過去のフレームと未来のフレームの両方を参照して最適化を行うため、仕組み上必ず遅延が発生する。 When using an IP camera in a conventional surveillance camera system, the stream output from the IP camera is H.264. Limited to H.264 baseline profile. This is because the surveillance camera system does not allow a high-profile stream with a delay due to the nature of viewing in real time. In the present specification, “high profile” means moving image data including bi-directional prediction inter frames (also expressed as future prediction frames, B frames, etc.). The high profile is optimized by referring to both the past frame and the future frame.
 ところで、一般に、動画データをサーバに保存する場合、データ容量が膨大になるという問題や長期保存ができないという問題がある。また、大容量のサーバを多数用意する必要があり、コストがかさむという問題がある。 By the way, in general, when moving image data is stored in a server, there are problems that the data capacity becomes enormous and long-term storage is impossible. In addition, it is necessary to prepare a large number of large-capacity servers, and there is a problem that costs increase.
 動画データの容量を減らすために、従来の監視カメラシステムでは、ベースラインプロファイルの最適化や可変レートでの容量削減を行っている。また、本件出願人は、ベースラインプロファイルの動画データをハイプロファイルに変換して保存する監視カメラシステムを既に提案している(特願2016-079470)。 In order to reduce the volume of video data, conventional surveillance camera systems optimize the baseline profile and reduce the capacity at a variable rate. Further, the applicant of the present application has already proposed a surveillance camera system that converts and saves the video data of the baseline profile into a high profile (Japanese Patent Application No. 2016-077940).
 本開示の一側面に係る監視カメラシステムは、
 監視カメラと、
 前記監視カメラにネットワークを介して接続されたサーバと、
を備え、
 前記監視カメラは、
 受光した映像を撮像して未加工の映像データを生成する撮像部と、
 前記未加工の映像データを、過去のフレームと未来のフレームの両方を参照して作成される双方向予測インターフレームを含む動画データに符号化して、前記サーバに送信するエンコード部と、
 前記未加工の映像データからリアルタイムの映像データを作成して、前記サーバに送信するライブ映像生成部と、
を有する。
A surveillance camera system according to one aspect of the present disclosure includes:
A surveillance camera,
A server connected to the surveillance camera via a network;
With
The surveillance camera is
An imaging unit that captures the received video and generates raw video data;
An encoding unit that encodes the raw video data into video data including a bi-directional prediction interframe created by referring to both past and future frames, and transmits the encoded video data to the server;
Creating a real-time video data from the raw video data and transmitting it to the server;
Have
図1は、一実施の形態に係る監視カメラシステムの概略的な構成を示す図である。FIG. 1 is a diagram illustrating a schematic configuration of a surveillance camera system according to an embodiment. 図2は、図1に示す監視カメラシステムにおける監視カメラの概略的な構成を示す図である。FIG. 2 is a diagram showing a schematic configuration of the surveillance camera in the surveillance camera system shown in FIG. 図3は、図2に示す監視カメラのエンコード部の概略的な内部構成を示すブロック図である。FIG. 3 is a block diagram showing a schematic internal configuration of the encoding unit of the surveillance camera shown in FIG. 図4は、図3に示すエンコード部における出力フレームの遅延と並び替えを説明するための図である。FIG. 4 is a diagram for explaining output frame delay and rearrangement in the encoding unit shown in FIG. 3. 図5は、図3に示すエンコード部の原理的な動作を説明するための図である。FIG. 5 is a diagram for explaining the principle operation of the encoding unit shown in FIG. 図6は、図3に示すエンコード部のハードウェア構成の一例を示すブロック図である。FIG. 6 is a block diagram illustrating an example of a hardware configuration of the encoding unit illustrated in FIG. 3. 図7は、図3に示すエンコード部のハードウェア構成の変形例を示すブロック図である。FIG. 7 is a block diagram showing a modification of the hardware configuration of the encoding unit shown in FIG. 図8は、図1に示す監視カメラシステムにおけるサーバの概略的な構成を示す図である。FIG. 8 is a diagram showing a schematic configuration of a server in the monitoring camera system shown in FIG. 図9は、図1に示す監視カメラシステムにおける端末装置(モバイル端末)の概略的な構成を示す図である。FIG. 9 is a diagram showing a schematic configuration of a terminal device (mobile terminal) in the surveillance camera system shown in FIG. 図10は、図1に示す監視カメラシステムにおける端末装置(閲覧PC)の概略的な構成を示す図である。FIG. 10 is a diagram showing a schematic configuration of a terminal device (browsing PC) in the surveillance camera system shown in FIG. 図11は、図1に示す監視カメラシステムの動作の一例を示す概略図である。FIG. 11 is a schematic diagram showing an example of the operation of the surveillance camera system shown in FIG. 図12は、一実施の形態の第1の変形例に係る監視カメラの概略的な構成を示す図である。FIG. 12 is a diagram illustrating a schematic configuration of a monitoring camera according to a first modification of the embodiment. 図13は、一実施の形態の第2の変形例に係る監視カメラの概略的な構成を示す図である。FIG. 13 is a diagram illustrating a schematic configuration of a monitoring camera according to a second modification of the embodiment. 図14は、一実施の形態の第3の変形例に係る監視カメラの概略的な構成を示す図である。FIG. 14 is a diagram illustrating a schematic configuration of a surveillance camera according to a third modification of the embodiment.
 従来の監視カメラシステムでは、動画データの保存をベースラインプロファイルで行っているため、容量の圧縮により映像が劣化したり、可変レートの場合は瞬時の動きに映像が乱れたりするなど、仕組み自体に限界が来ている。 In conventional surveillance camera systems, video data is stored in a baseline profile, so the video itself deteriorates due to capacity compression, or the video is distorted by instantaneous movement in the case of variable rate. The limit is coming.
 一方、ベースラインプロファイルの動画データをハイプロファイルに変換して保存する構成では、最適化したデータ保存が可能とされるものの、レコーダとなるサーバのCPUリソースを使用して変換を行うため、システムにかかるコストが大きくなる傾向がある。 On the other hand, in the configuration in which the video data of the baseline profile is converted to a high profile and saved, the optimized data can be saved, but the conversion is performed using the CPU resource of the server serving as a recorder. Such costs tend to increase.
 また、カメラからのストリーム出力はベースラインプロファイルに限定されていることから、ネットワークに複数台のカメラを接続すると、トラフィックが増大し、ネットワーク設計や機器への負荷が発生する。 Also, since the stream output from the camera is limited to the baseline profile, connecting multiple cameras to the network increases the traffic, causing a load on the network design and equipment.
[実施形態]
 以下で説明する実施形態は、以上のような点を考慮してなされたものであり、その目的は、リアルタイムの映像を取得可能な監視カメラシステムであって、サーバに負荷をかけることなく、最適化したデータ保存を可能とする監視カメラシステムを提供することにある。
[Embodiment]
The embodiment described below has been made in consideration of the above points, and the purpose thereof is a surveillance camera system capable of acquiring real-time video, and is optimal without imposing a load on the server. Another object of the present invention is to provide a surveillance camera system that can store data in an organized manner.
 実施形態の第1の態様に係る監視カメラシステムは、
 監視カメラと、
 前記監視カメラにネットワークを介して接続されたサーバと、
を備え、
 前記監視カメラは、
 受光した映像を撮像して未加工の映像データを生成する撮像部と、
 前記未加工の映像データを、過去のフレームと未来のフレームの両方を参照して作成される双方向予測インターフレームを含む動画データに符号化して、前記サーバに送信するエンコード部と、
 前記未加工の映像データからリアルタイムの映像データを作成して、前記サーバに送信するライブ映像生成部と、
を有する。
The surveillance camera system according to the first aspect of the embodiment includes:
A surveillance camera,
A server connected to the surveillance camera via a network;
With
The surveillance camera is
An imaging unit that captures the received video and generates raw video data;
An encoding unit that encodes the raw video data into video data including a bi-directional prediction interframe created by referring to both past and future frames, and transmits the encoded video data to the server;
Creating a real-time video data from the raw video data and transmitting it to the server;
Have
 このような態様によれば、監視カメラのエンコード部が、未加工の映像データを、双方向予測インターフレーム(Bフレーム)を含む動画データに符号化してサーバに送信するため、サーバに負荷をかけることなく、最適化したデータ保存が可能となる。また、ネットワーク上を流れるデータ量が大幅に小さくなり、ネットワークトラフィックを低減することができる。これにより、ネットワーク設計や機器への負荷が低減し、ネットワーク機器を長寿命化できる。また、ネットワークに複数台のカメラを接続することが可能となる。 According to such an aspect, the encoding unit of the surveillance camera encodes the raw video data into moving image data including a bi-directional prediction inter frame (B frame) and transmits the encoded video data to the server, which places a load on the server. And optimized data storage is possible. In addition, the amount of data flowing on the network is significantly reduced, and network traffic can be reduced. As a result, the load on the network design and equipment is reduced, and the life of the network equipment can be extended. In addition, a plurality of cameras can be connected to the network.
 また、このような態様によれば、双方向予測インターフレーム(Bフレーム)を含む動画データへの符号化が、サーバのCPUリソースを使用して行われるのではなく、監視カメラの内部で行われるため、サーバの負荷が低減し、システムにかかるコストを低減できる。また、サーバの負荷が低減するので、サーバ1台当たりの監視カメラの収容台数を増やすことができる。 Further, according to such an aspect, encoding into moving image data including a bi-predictive inter frame (B frame) is not performed using the CPU resource of the server, but is performed inside the surveillance camera. Therefore, the load on the server is reduced and the cost for the system can be reduced. Moreover, since the load on the server is reduced, the number of surveillance cameras that can be accommodated per server can be increased.
 また、このような態様によれば、監視カメラから送信される動画データは、双方向予測インターフレーム(Bフレーム)を含むことから、必然的に遅延しているものの、ライブ映像生成部が、遅延のない未加工の映像データからリアルタイムの映像データを作成してサーバに送信するため、サーバは、リアルタイムの映像を取得可能であり、リアルタイムでの閲覧という監視カメラシステムとしての必須の要件は何ら損なわれることがない。 In addition, according to such an aspect, the moving image data transmitted from the surveillance camera includes a bidirectional prediction inter frame (B frame), and thus is inevitably delayed, but the live video generation unit performs the delay. Since real-time video data is created from raw video data without data and transmitted to the server, the server can acquire real-time video, and the essential requirement of the surveillance camera system for real-time browsing is lost. It will not be.
 実施形態の第2の態様に係る監視カメラシステムは、上記第1の態様に係る監視カメラシステムであって、
 前記エンコード部は、前記動画データをストリーム方式で前記サーバに送信する。
A surveillance camera system according to a second aspect of the embodiment is the surveillance camera system according to the first aspect,
The encoding unit transmits the moving image data to the server by a stream method.
 実施形態の第3の態様に係る監視カメラシステムは、上記第1の態様に係る監視カメラシステムであって、
 前記エンコード部は、前記動画データからGOP単位で撮影時刻順に分割データを作成してメモリに保存し、前記サーバから分割データ要求を受け付ける毎に、前記メモリから前記サーバに1つずつ分割データを転送し、
 前記サーバは、前記監視カメラから受け取る分割データを撮影時刻順に結合して動画データを作成する。
A surveillance camera system according to a third aspect of the embodiment is the surveillance camera system according to the first aspect,
The encoding unit creates divided data in order of shooting time from the moving image data in GOP units, stores the divided data in a memory, and transfers the divided data from the memory to the server one by one each time a divided data request is received from the server. And
The server creates moving image data by combining divided data received from the monitoring camera in order of photographing time.
 実施形態の第4の態様に係る監視カメラシステムは、上記第1~第3のいずれかの態様に係る監視カメラシステムであって、
 前記ライブ映像生成部は、前記サーバからライブ画像要求を受け付ける毎に、前記未加工の映像データからリアルタイムの静止画像データを切り出して前記サーバに送信する。
A surveillance camera system according to a fourth aspect of the embodiment is the surveillance camera system according to any one of the first to third aspects,
Each time the live video generation unit receives a live image request from the server, the live video generation unit cuts out real-time still image data from the raw video data and transmits it to the server.
 実施形態の第5の態様に係る監視カメラシステムは、上記第1~第3のいずれかの態様に係る監視カメラシステムであって、
 前記ライブ映像生成部は、前記未加工の映像データを、前記双方向予測インターフレームを含まないリアルタイムの動画データに符号化して、ストリーム方式で前記サーバに送信する。
A surveillance camera system according to a fifth aspect of the embodiment is the surveillance camera system according to any one of the first to third aspects,
The live video generation unit encodes the raw video data into real-time video data not including the bidirectional prediction interframe, and transmits the encoded video data to the server in a stream format.
 実施形態の第6の態様に係る監視カメラシステムは、上記第1~第3のいずれかの態様に係る監視カメラシステムであって、
 前記ライブ映像生成部は、予め定められたスナップショット時間間隔で、前記未加工の映像データからリアルタイムの静止画像データを切り出して前記サーバに送信することを繰り返す。
A surveillance camera system according to a sixth aspect of the embodiment is the surveillance camera system according to any one of the first to third aspects,
The live video generation unit repeats cutting out real-time still image data from the raw video data and transmitting it to the server at a predetermined snapshot time interval.
 実施形態の第7の態様に係る監視カメラシステムは、上記第4または第6の態様に係る監視カメラシステムであって、
 前記サーバは、
 前記監視カメラから受け取る静止画像データを記憶装置に保存する一次映像保存部と、
 ネットワークを介して接続された端末装置に前記静止画像データをライブ画像として送信するライブ画像送信部と、
 前記端末装置から、ライブ画像からの巻き戻し要求を受け付けると、前回送信した静止画像データより一定時間分過去の時点の静止画像データを前記記憶装置から取得して、巻き戻しライブ画像として前記端末装置に送信する巻き戻しライブ画像送信部と、
 前記端末装置から、巻き戻しライブ画像からの早送り要求を受け付けると、現時点の静止画像データに達するまで、前回送信した静止画像データより一定時間分未来の時点の静止画像データを前記記憶装置から取得して、早送りライブ画像として前記端末装置に送信する早送りライブ画像送信部と、
を有する。
A surveillance camera system according to a seventh aspect of the embodiment is the surveillance camera system according to the fourth or sixth aspect,
The server
A primary video storage unit for storing still image data received from the surveillance camera in a storage device;
A live image transmission unit that transmits the still image data as a live image to a terminal device connected via a network;
When receiving a rewind request from a live image from the terminal device, the terminal device acquires still image data at a time point that is a certain time past from the previously transmitted still image data from the storage device, and the terminal device as a rewind live image. A rewind live image transmitter to transmit to
When a fast-forward request from a rewind live image is received from the terminal device, still image data at a time point that is a certain time later than the previously transmitted still image data is acquired from the storage device until the current still image data is reached. A fast-forward live image transmission unit for transmitting to the terminal device as a fast-forward live image;
Have
 このような態様によれば、端末装置にてライブ画像を表示しているときに、ユーザが見落とし等に気付いた場合には、端末装置からサーバに巻き戻し要求を送信することで、端末装置ではそのまま過去にさかのぼって確認することができ、また、端末装置からサーバに早送り要求を送信することで、過去から現時点までを早送りで確認することができる。 According to such an aspect, when the user notices an oversight or the like while displaying a live image on the terminal device, the terminal device transmits a rewind request from the terminal device to the server. It is possible to check back in the past as it is, and by sending a fast-forward request from the terminal device to the server, it is possible to check from the past to the current time by fast-forward.
 実施形態の第8の態様に係る監視カメラシステムは、上記第1~第7のいずれかの態様に係る監視カメラシステムであって、
 前記サーバは、前記監視カメラから受け取る動画データを、前記エンコード部での符号化に係る遅延に応じて時間調整して記憶装置に保存する二次映像保存部を有する。
A surveillance camera system according to an eighth aspect of the embodiment is the surveillance camera system according to any one of the first to seventh aspects,
The server includes a secondary video storage unit that stores the moving image data received from the monitoring camera in a storage device by adjusting the time according to the delay associated with the encoding by the encoding unit.
 このような態様によれば、監視カメラから送信される動画データはエンコード部での符号化により必然的に遅延しているものの、記憶装置には遅延時間に応じて時間調整された状態で保存されるため、サーバは、端末装置から時刻を特定して過去の動画データの再生要求を受け付けた場合に、要求された時刻の動画データを記憶装置から正確に抽出して端末装置に送信することができる。 According to such an aspect, the moving image data transmitted from the surveillance camera is inevitably delayed by encoding in the encoding unit, but is stored in the storage device in a state adjusted in time according to the delay time. Therefore, when the server specifies the time from the terminal device and receives a reproduction request for past moving image data, the server can accurately extract the moving image data at the requested time from the storage device and transmit it to the terminal device. it can.
 実施形態の第9の態様に係る監視カメラは、
 サーバにネットワークを介して接続される監視カメラであって、
 受光した映像を撮像して未加工の映像データを生成する撮像部と、
 前記未加工の映像データを、過去のフレームと未来のフレームの両方を参照して作成される双方向予測インターフレームを含む動画データに符号化して、前記サーバに送信するエンコード部と、
 前記未加工の映像データからリアルタイムの映像データを作成して、前記サーバに送信するライブ映像生成部と、
を有する。
The surveillance camera according to the ninth aspect of the embodiment is
A surveillance camera connected to a server via a network,
An imaging unit that captures the received video and generates raw video data;
An encoding unit that encodes the raw video data into video data including a bi-directional prediction interframe created by referring to both past and future frames, and transmits the encoded video data to the server;
Creating a real-time video data from the raw video data and transmitting it to the server;
Have
 実施形態の第10の態様に係るサーバは、
 受光した映像を撮像して未加工の映像データを生成する撮像部と、
 前記未加工の映像データを、過去のフレームと未来のフレームの両方を参照して作成される双方向予測インターフレームを含む動画データに符号化して、前記サーバに送信するエンコード部と、
 前記未加工の映像データからリアルタイムの映像データを作成して、前記サーバに送信するライブ映像生成部と、
を有する監視カメラにネットワークを介して接続されるサーバであって、
 前記監視カメラから受け取るリアルタイムの映像データを、記憶装置に保存する一次映像保存部と、
 前記監視カメラから受け取る動画データを、前記エンコード部での符号化に係る遅延に応じて時間調整して前記記憶装置に保存する二次映像保存部と、
を備える。
The server according to the tenth aspect of the embodiment is
An imaging unit that captures the received video and generates raw video data;
An encoding unit that encodes the raw video data into video data including a bi-directional prediction interframe created by referring to both past and future frames, and transmits the encoded video data to the server;
Creating a real-time video data from the raw video data and transmitting it to the server;
A server connected to a surveillance camera via a network,
A primary video storage unit that stores real-time video data received from the surveillance camera in a storage device;
A secondary video storage unit that saves the video data received from the monitoring camera in the storage device by adjusting the time according to the delay related to the encoding in the encoding unit;
Is provided.
 以下に、添付の図面を参照して、実施の形態の具体例を詳細に説明する。なお、各図において同等の機能を有する構成要素には同一の符号を付し、同一符号の構成要素の詳しい説明は繰り返さない。 Hereinafter, specific examples of the embodiment will be described in detail with reference to the accompanying drawings. In addition, in each figure, the component which has an equivalent function is attached | subjected the same code | symbol, and detailed description of the component of the same code | symbol is not repeated.
 図1は、一実施の形態に係る監視カメラシステム10の概略的な構成を示す図である。 FIG. 1 is a diagram illustrating a schematic configuration of a surveillance camera system 10 according to an embodiment.
 図1に示すように、監視カメラシステム10は、監視カメラ400A、400Bと、サーバ100と、端末装置200A、200B、300A~300Cとを備えている。監視カメラ400A、400Bとサーバ100と端末装置200A、200B、300A~300Cとは、ネットワーク500を介して互いに接続されている。 As shown in FIG. 1, the surveillance camera system 10 includes surveillance cameras 400A and 400B, a server 100, and terminal devices 200A, 200B, and 300A to 300C. The monitoring cameras 400A and 400B, the server 100, and the terminal devices 200A, 200B, and 300A to 300C are connected to each other via the network 500.
 ネットワーク500は、有線回線と無線回線のいずれでもよく、回線の種類や形態は問わない。本実施の形態では、ネットワーク500は、LANやインターネット、Wi-Fi回線、3G/LTE回線などの通信網から構成される。 The network 500 may be either a wired line or a wireless line, and the type and form of the line are not limited. In the present embodiment, the network 500 includes a communication network such as a LAN, the Internet, a Wi-Fi line, or a 3G / LTE line.
 本実施の形態では、監視カメラ400A、400Bは、サーバ100からライブ画像要求を受け付ける毎に、カメラ映像からリアルタイムの静止画像を作成してサーバ100に送信する。サーバ100は、監視カメラ400から静止画像を取得して蓄積し、端末装置200A、200B、300A~300Cから閲覧要求を受け付ける毎に、閲覧用の静止画像を端末装置200A、200B、300A~300Cに送信する。図示された例では、監視カメラ400A、400Bは、サーバ100と同じLAN内に存在している。監視カメラ400A、400Bとサーバ100との通信は、TCP/IP方式であり、確実にカメラ映像を送受信することができ、信頼性が高い。 In this embodiment, every time the monitoring cameras 400A and 400B accept a live image request from the server 100, the monitoring cameras 400A and 400B create a real-time still image from the camera video and transmit it to the server 100. The server 100 acquires and accumulates still images from the monitoring camera 400, and each time a browsing request is received from the terminal devices 200A, 200B, 300A to 300C, the still image for browsing is stored in the terminal devices 200A, 200B, 300A to 300C. Send. In the illustrated example, the monitoring cameras 400A and 400B exist in the same LAN as the server 100. The communication between the monitoring cameras 400A and 400B and the server 100 is a TCP / IP system, can reliably transmit and receive camera images, and has high reliability.
 端末装置200A、200B、300A~300Cは、ユーザが使用するものであり、スマートフォン、タブレットなどのモバイル端末、ノートブックコンピュータまたはデスクトップコンピュータなどである。端末装置200A、200B、300A~300Cは、サーバ100と同じLAN内に存在していてもよいし、ネットワーク500にルータ600を介して接続されていれば、インターネットを介してサーバ100に接続されていてもよい。 The terminal devices 200A, 200B, 300A to 300C are used by the user, and are mobile terminals such as smartphones and tablets, notebook computers or desktop computers. The terminal devices 200A, 200B, and 300A to 300C may exist in the same LAN as the server 100. If the terminal devices 200A, 200B, and 300A to 300C are connected to the network 500 via the router 600, they are connected to the server 100 via the Internet. May be.
 以下の説明では、監視カメラ400A、400Bに符号400を付してまとめて表現することがある。また、モバイル端末としての端末装置200A、200Bに符号200、閲覧PCとしての端末装置300A~300Cに符号300を付してまとめて表現することがある。 In the following description, the monitoring cameras 400A and 400B may be collectively expressed by adding a reference numeral 400. In some cases, the terminal devices 200A and 200B as mobile terminals are denoted by reference numeral 200, and the terminal devices 300A to 300C as viewing PCs are denoted by reference numeral 300 and collectively represented.
 次に、監視カメラ400の構成を説明する。図2は、監視カメラ400の概略構成を示す図である。 Next, the configuration of the monitoring camera 400 will be described. FIG. 2 is a diagram illustrating a schematic configuration of the monitoring camera 400.
 図2に示すように、監視カメラ400は、撮像部401と、エンコード部402と、ライブ映像生成部403と、ネットワークインターフェース404などの通信制御部と、を有している。これらの各部401~404は、同一の筐体405に一体に収容されている。 As shown in FIG. 2, the monitoring camera 400 includes an imaging unit 401, an encoding unit 402, a live video generation unit 403, and a communication control unit such as a network interface 404. Each of these parts 401 to 404 is housed in the same housing 405.
 図示された例では、撮像部401は、レンズ4011と、レンズ4011を通して受光した映像を撮像する映像データ受信部4012と、撮像結果に基づいて未加工の映像データ(RAWデータなどとも表現される)を生成するRAWデータ生成部4013と、を有している。撮像部401の構成は、従来の監視カメラのものと同様であり、詳細な説明を省略する。 In the illustrated example, the imaging unit 401 includes a lens 4011, a video data receiving unit 4012 that captures an image received through the lens 4011, and raw video data (also expressed as RAW data) based on the imaging result. And a RAW data generation unit 4013 for generating. The configuration of the imaging unit 401 is the same as that of a conventional surveillance camera, and detailed description thereof is omitted.
 エンコード部402は、撮像部401に接続されており、撮像部401にて生成された未加工の映像データ(RAWデータ)を、過去のフレームと未来のフレームの両方を参照して作成される双方向予測インターフレーム(Bフレーム)を含む動画データに符号化して、ネットワークインターフェース404を介してサーバ100に送信する。本実施の形態では、エンコード部402は、動画データをストリーム方式(ストリーミング方式ともいう)でサーバ100に送信する。 The encoding unit 402 is connected to the imaging unit 401, and both of the raw video data (RAW data) generated by the imaging unit 401 are created with reference to both past frames and future frames. It encodes into the moving image data containing a direction prediction inter frame (B frame), and transmits to the server 100 via the network interface 404. In the present embodiment, the encoding unit 402 transmits moving image data to the server 100 by a stream method (also referred to as a streaming method).
 図3は、エンコード部402の概略的な内部構成を示すブロック図である。 FIG. 3 is a block diagram showing a schematic internal configuration of the encoding unit 402.
 図3に示すように、エンコード部402は、制御回路4021と、フレーム内圧縮ブロック4022と、フレーム間圧縮ブロック4023と、可変長符号圧縮/フレーム並べ替えブロック4024と、メモリ4025と、を有している。 As shown in FIG. 3, the encoding unit 402 includes a control circuit 4021, an intra-frame compression block 4022, an inter-frame compression block 4023, a variable length code compression / frame rearrangement block 4024, and a memory 4025. ing.
 制御回路4021は、圧縮パラメータの設定など、各ブロックの制御を行う。フレーム内圧縮ブロック4022は、フレーム内の空間的相関関係による情報圧縮(1画面内のデータ圧縮)を行う。フレーム間圧縮ブロック4023は、前後のフレームとの時間的相関関係による情報圧縮を行う。可変長符号圧縮/フレーム並べ替えブロック4024は、出現頻度の高いデータに短い符号を割り当てることで情報を圧縮する(エントロピー符号化)とともに、符号化したフレームデータの出力の順番を並べ替える。メモリ4025は、双方向予測インターフレーム(Bフレーム)を生成するために、複数(図示された例では5フレーム)のフレームデータを蓄積する。 The control circuit 4021 controls each block such as setting compression parameters. The intra-frame compression block 4022 performs information compression (data compression within one screen) based on the spatial correlation within the frame. The inter-frame compression block 4023 performs information compression based on temporal correlation with previous and subsequent frames. The variable-length code compression / frame rearrangement block 4024 compresses information by assigning a short code to frequently occurring data (entropy coding) and rearranges the output order of the encoded frame data. The memory 4025 accumulates a plurality of (5 frames in the illustrated example) frame data in order to generate a bidirectional prediction inter frame (B frame).
 図4は、エンコード部402における出力フレームの遅延と並び替えを説明するための図である。 FIG. 4 is a diagram for explaining output frame delay and rearrangement in the encoding unit 402.
 図4に示すように、エンコード部402は、RAWデータを、単独でデコーディング可能で一定時間ごとに作成されるイントラフレーム(Iフレーム)と、過去のフレームを参照してイントラフレームの間に複数作成される予測インターフレーム(Pフレーム)と、過去のフレームと未来のフレームの両方を参照して作成される双方向予測インターフレーム(Bフレーム)とを含む動画データに符号化する。ここで、双方向予測インターフレーム(Bフレーム)は、予測インターフレーム(Pフレーム)が確定しなければ生成できない。そのため、双方向予測インターフレーム(Bフレーム)を含むハイプロファイル出力は、双方向予測インターフレーム(Bフレーム)を含まないベースライン出力に比べて、出力までの遅延時間が長く、かつフレームの並べ替えが必要である。遅延時間が長いため、エンコード部402には、その期間のフレームを保持するメモリ4025が設けられている。 As shown in FIG. 4, the encoding unit 402 includes a plurality of pieces of RAW data between an intra frame (I frame) that can be decoded independently and created at regular intervals, and an intra frame with reference to a past frame. It is encoded into moving image data including a prediction inter frame (P frame) to be created and a bidirectional prediction inter frame (B frame) created by referring to both the past frame and the future frame. Here, a bidirectional prediction inter frame (B frame) cannot be generated unless a prediction inter frame (P frame) is determined. Therefore, a high profile output including a bi-predictive interframe (B frame) has a longer delay time to output and a rearrangement of frames than a baseline output not including a bi-predictive interframe (B frame). is required. Since the delay time is long, the encoding unit 402 is provided with a memory 4025 that holds a frame for that period.
 図5は、エンコード部402の原理的な動作を説明するための図である。 FIG. 5 is a diagram for explaining the principle operation of the encoding unit 402.
 図5に示すように、エンコード部402は、順次入力されるRAWデータのうち最初のフレームデータについては、そのままフレーム内圧縮(空間的情報圧縮)を行う(ステップS41)。一方、エンコード部402は、2枚目以降のフレームデータについては、まず、前後のフレームのフレーム内圧縮を展開し(ステップS42)、それらを参照してフレーム間圧縮(時間的情報圧縮)を行ったのち(ステップS43)、フレーム内圧縮(空間的情報圧縮)を行う(ステップS41)。そして、エンコード部402は、フレーム内圧縮されたフレームデータに対して、符号の出現頻度による圧縮(エントロピー符号化)および並び替えを行って出力する(ステップS44)。 As shown in FIG. 5, the encoding unit 402 performs intra-frame compression (spatial information compression) for the first frame data among the sequentially input RAW data (step S41). On the other hand, for the second and subsequent frame data, the encoding unit 402 first decompresses the frames before and after the frame (step S42), and performs inter-frame compression (temporal information compression) with reference to them. After that (step S43), intra-frame compression (spatial information compression) is performed (step S41). Then, the encoding unit 402 performs compression (entropy encoding) and rearrangement on the frame data compressed in the frame based on the appearance frequency of the code and outputs the result (step S44).
 図6は、エンコード部402のハードウェア構成の一例を示すブロック図である。 FIG. 6 is a block diagram illustrating an example of a hardware configuration of the encoding unit 402.
 図6に示す例では、エンコード部402は、エンコードLSI410と、インターフェースブロック411A、411Bと、を有している。エンコードLSI410は、ハイプロファイルのエンコードを行うハードウェアであり、制御回路4021と、フレーム内圧縮ブロック4022と、フレーム間圧縮ブロック4023と、可変長符号圧縮/フレーム並べ替えブロック4024(図3参照)とを含んでいる。また、図示された例では、エンコードLSI410は、内部にメモリ4025を含んでいる。インターフェースブロック411A、411Bは、データ入出力とLSIの入出力とが異なる場合に変換を行う。 In the example shown in FIG. 6, the encoding unit 402 includes an encoding LSI 410 and interface blocks 411A and 411B. The encoding LSI 410 is hardware that performs high-profile encoding, and includes a control circuit 4021, an intra-frame compression block 4022, an inter-frame compression block 4023, and a variable-length code compression / frame rearrangement block 4024 (see FIG. 3). Is included. In the illustrated example, the encoding LSI 410 includes a memory 4025 inside. The interface blocks 411A and 411B perform conversion when data input / output and LSI input / output are different.
 エンコードLSI410としては、具体的には、たとえば、ソシオネクスト社のH.264トランスレート対応メモリ内蔵トランスコーダLSI MB86M0xが用いられる。従来、このようなエンコードLSI410は、Bフレームを生成することから、必然的に遅延を発生させるため、その性格上、テレビ、ハードディスクレコーダ、パソコンなどデジタル放送の録画機能を搭載する据置き機器、またはスマートフォンやタブレット製品向けのモバイル関連機器に利用されるものであり、本件発明者らの知るかぎり、監視カメラに利用された例は存在しない。 As the encode LSI 410, specifically, for example, H.S. H.264 translation compatible memory built-in transcoder LSI MB86M0x is used. Conventionally, such an encoding LSI 410 generates a B frame, and therefore inevitably generates a delay. Therefore, in terms of its nature, a stationary device equipped with a digital broadcast recording function such as a television, a hard disk recorder, or a personal computer, or It is used for mobile-related devices for smartphones and tablet products, and as far as the inventors know, there is no example used for surveillance cameras.
 なお、図6に示す例では、メモリ4025はエンコードLSI410に内蔵されていたが、これに限定されず、図7に示すように、メモリ4025はエンコードLSI410に外付けされていてもよい。 In the example shown in FIG. 6, the memory 4025 is built in the encode LSI 410, but the present invention is not limited to this, and the memory 4025 may be externally attached to the encode LSI 410 as shown in FIG.
 また、エンコード部402のハードウェア構成は、図6および図7に示すようなエンコードLSI410を用いて実装される態様に限定されず、マイコンとソフトウェアを用いて実装される態様であってもよい。 Further, the hardware configuration of the encoding unit 402 is not limited to an aspect implemented using the encoding LSI 410 as shown in FIGS. 6 and 7, and may be an aspect implemented using a microcomputer and software.
 図2に戻って、ライブ映像生成部403は、撮像部401に接続されており、遅延のない未加工の映像データ(RAWデータ)からリアルタイムの映像データを作成して、サーバ100に送信する。本実施の形態では、ライブ映像生成部403は、サーバ100からライブ画像要求を受け付ける毎に、遅延のない未加工の映像データ(RAWデータ)からリアルタイムの静止画像データを作成して(たとえば、そのまま切り出して)、ライブ画像として、ネットワークインターフェース404を介してサーバ100に送信する。なお、本明細書において「ライブ画像」とは、リアルタイムのカメラ映像を表す1枚の静止画像を意味する。 Returning to FIG. 2, the live video generation unit 403 is connected to the imaging unit 401, creates real-time video data from raw video data (RAW data) without delay, and transmits the real-time video data to the server 100. In the present embodiment, every time a live image request is received from the server 100, the live video generation unit 403 creates real-time still image data from raw video data (RAW data) without delay (for example, as it is). The live image is cut out and transmitted to the server 100 via the network interface 404. In this specification, “live image” means one still image representing a real-time camera image.
 次に、サーバ100の構成を説明する。図8は、サーバ100の概略的な構成を示す図である。 Next, the configuration of the server 100 will be described. FIG. 8 is a diagram illustrating a schematic configuration of the server 100.
 図8に示すように、サーバ100は、キャッシュメモリであるメモリ102を伴うCPU101やデバイスドライバなどを有する制御・演算装置と、DRAMなどの主記憶装置やハードディスクなどの補助記憶装置を有する記憶装置110と、ネットワークインターフェース104などの通信制御装置と、表示装置としてのディスプレイ103と、キーボード105、マウス106などを含む入力装置と、を備えている。 As shown in FIG. 8, the server 100 includes a control / arithmetic unit having a CPU 101 and a device driver with a memory 102 as a cache memory, and a storage device 110 having a main storage device such as a DRAM and an auxiliary storage device such as a hard disk. And a communication control device such as a network interface 104, a display 103 as a display device, and an input device including a keyboard 105, a mouse 106, and the like.
 記憶装置110には、一次映像フォルダ111と、二次映像フォルダ112と、プログラム113と、オペレーティングシステム114と、認証用データベースや環境設定フォルダ(図示しない)などと、が格納されている。 The storage device 110 stores a primary video folder 111, a secondary video folder 112, a program 113, an operating system 114, an authentication database, an environment setting folder (not shown), and the like.
 一次映像フォルダ111には、監視カメラ400から受け取るリアルタイムの映像データが保存される。本実施の形態では、一次映像フォルダ111には、監視カメラ400から受け取るリアルタイムの静止画像データが順次保存される。二次映像フォルダ112には、監視カメラ400から受け取る動画データが、予め定められた保存単位(たとえば10分)ごとに1つのファイルとして保存される。 In the primary video folder 111, real-time video data received from the monitoring camera 400 is stored. In the present embodiment, real-time still image data received from the monitoring camera 400 is sequentially stored in the primary video folder 111. In the secondary video folder 112, moving image data received from the monitoring camera 400 is stored as one file for each predetermined storage unit (for example, 10 minutes).
 プログラム113は、通常、記憶装置110の補助記憶装置に格納されており、実行時には主記憶装置にロードされる。認証用データベースには、ID、パスワード、各監視カメラ400や端末装置200、300のポート番号とIPアドレス、IPアドレスの無いモバイル端末200では個体識別番号(UID)などが記憶されている。環境設定フォルダには、各監視カメラ400の映像取得タイミング、各監視カメラ400のエンコード部402での符号化に係る遅延時間などが予め記憶されている。 The program 113 is normally stored in the auxiliary storage device of the storage device 110, and is loaded into the main storage device at the time of execution. The authentication database stores IDs, passwords, port numbers and IP addresses of the respective monitoring cameras 400 and terminal devices 200 and 300, and individual identification numbers (UIDs) in the mobile terminal 200 having no IP address. In the environment setting folder, video acquisition timing of each monitoring camera 400, delay time related to encoding by the encoding unit 402 of each monitoring camera 400, and the like are stored in advance.
 本実施の形態では、サーバ100は、サーバと端末装置とが一体となっており、自らカメラ映像を表示する端末装置の役割も果たすため、また、メンテナンスや管理のため、表示装置としてのディスプレイ103と、入力装置としてのキーボード105やマウス106とを有している。サーバ109にてカメラ映像の表示を要しない場合は、表示装置としての端末機能は無くてもよい。 In this embodiment, since the server and the terminal device are integrated with each other in the present embodiment, the server 100 also serves as a terminal device that displays the camera video by itself, and for maintenance and management, the display 103 as a display device. And a keyboard 105 and a mouse 106 as input devices. If the server 109 does not require display of camera video, the terminal function as a display device may not be provided.
 サーバ100は、CPU101が、プログラム113をメモリ102にロードして実行することにより、監視カメラ400からのカメラ映像の取得および端末装置200、300へのカメラ映像の送出が可能なコンピュータの機能を実現する。CPU101は、通常のコンピュータに搭載される演算処理装置であり、各種プログラムを実行し、各種制御等を行う。 The server 100 implements a computer function that allows the CPU 101 to load the program 113 into the memory 102 and execute it, thereby acquiring the camera video from the monitoring camera 400 and sending the camera video to the terminal devices 200 and 300. To do. The CPU 101 is an arithmetic processing device mounted on a normal computer, executes various programs, and performs various controls.
 サーバ100は、1台のサーバであってもよいし、複数のサーバからなるサーバ群であってもよい。たとえば、二次映像フォルダ112について、一定期間(たとえば24時間)経過後の動画データの保存先を、監視カメラ400から動画データを取得するサーバとは別のサーバに設けられた二次映像フォルダ112としてもよい。頻繁には再生しない過去の動画データの保存先を別にすることで、サーバ100の負荷を低減でき、さらに多くの台数のカメラを同一ネットワーク上で監視可能となる。 The server 100 may be a single server or a server group including a plurality of servers. For example, for the secondary video folder 112, the secondary video folder 112 provided in a server different from the server that acquires the video data from the monitoring camera 400 as the storage destination of the video data after a certain period (for example, 24 hours) has elapsed. It is good. By separating the storage destination of past video data that is not frequently played back, the load on the server 100 can be reduced, and a larger number of cameras can be monitored on the same network.
 プログラム113は、コンピュータに、監視カメラ400から受け取る動画データを、エンコード部402での符号化に係る遅延に応じて時間調整して二次映像フォルダ112に保存する二次映像保存機能を実現させるためのプログラムである。サーバ100は、CPU101が、プログラム113をメモリ102にロードして実行することにより、監視カメラ400から受け取る動画データを、エンコード部402での符号化に係る遅延に応じて時間調整して二次映像フォルダ112に保存する二次映像保存部として機能する。 The program 113 causes the computer to realize a secondary video storage function that saves the video data received from the monitoring camera 400 in the secondary video folder 112 by adjusting the time according to the delay related to the encoding in the encoding unit 402. It is a program. In the server 100, the CPU 101 loads the program 113 to the memory 102 and executes it so that the moving image data received from the monitoring camera 400 is time-adjusted according to the delay related to the encoding in the encoding unit 402 and the secondary video. It functions as a secondary video storage unit stored in the folder 112.
 また、プログラム113は、コンピュータに、監視カメラ400から受け取る静止画像データを、たとえば予め定められた閲覧期間(たとえば24時間)、一次映像フォルダ111に保存する一次映像保存機能を実現させるためのプログラムである。サーバ100は、CPU101が、プログラム113をメモリ102にロードして実行することにより、監視カメラ400から受け取る静止画像データを、たとえば予め定められた閲覧期間、一次映像フォルダ111に保存する一次映像保存部として機能する。なお、静止画像データの閲覧期間(一次映像フォルダ111での保存期間)は、予め定められている態様に限定されず、その時の記憶装置110の空き容量に応じて適宜設定されるようになっていてもよい。 The program 113 is a program for causing a computer to realize a primary video storage function for storing still image data received from the surveillance camera 400 in the primary video folder 111, for example, for a predetermined browsing period (for example, 24 hours). is there. In the server 100, a primary video storage unit that stores still image data received from the monitoring camera 400 in the primary video folder 111, for example, for a predetermined viewing period, by the CPU 101 loading and executing the program 113 in the memory 102. Function as. Note that the still image data browsing period (storage period in the primary video folder 111) is not limited to a predetermined mode, and is set as appropriate according to the free capacity of the storage device 110 at that time. May be.
 また、プログラム113は、コンピュータに、端末装置200、300からのカメラ映像の再生要求を受付するカメラ映像再生要求受付機能、一次映像フォルダ111に保存された最新の静止画像データを、ライブ画像として端末装置200、300に送信するライブ画像送信機能を実現させるためのプログラムである。サーバ100は、CPU101が、プログラム113をメモリ102にロードして実行することにより、端末装置200、300からカメラ映像再生要求を受け付けると、一次映像フォルダ111に保存された最新の静止画像データを、ライブ画像として端末装置200、300に送信するライブ画像送信部として機能する。 Further, the program 113 has a camera video playback request receiving function for receiving camera video playback requests from the terminal devices 200 and 300 in the computer, and the latest still image data stored in the primary video folder 111 as a live image as a terminal. This is a program for realizing a live image transmission function to be transmitted to the devices 200 and 300. When the CPU 101 receives a camera video playback request from the terminal devices 200 and 300 by the CPU 101 loading and executing the program 113 in the memory 102, the latest still image data stored in the primary video folder 111 is It functions as a live image transmission unit that transmits to the terminal devices 200 and 300 as a live image.
 また、プログラム113は、コンピュータに、端末装置200、300からの、ライブ画像からの巻き戻し要求を受付する巻き戻し要求受付機能、ライブ画像からの巻き戻し要求を受け付けると、前回送信した静止画像データより一定時間分過去の時点の静止画像データを一次映像フォルダ111から取得して、巻き戻しライブ画像として端末装置200、300に送信する巻き戻しライブ画像送信機能、端末装置200、300からの、巻き戻しライブ画像からの早送り要求を受付する早送り要求受付機能、巻き戻しライブ画像からの早送り要求を受け付けると、現時点の静止画像データに達するまで、送信した静止画像データより一定時間分未来の時点の静止画像データを一次映像フォルダ111から取得して、早送りライブ画像として端末装置200、300に送信する早送りライブ画像送信機能を実現させるためのプログラムである。サーバ100は、CPU101が、プログラム113をメモリ102にロードして実行することにより、端末装置200、300から、ライブ画像からの巻き戻し要求を受け付けると、前回送信した静止画像データより一定時間分過去の時点の静止画像データを一次映像フォルダ111から取得して、巻き戻しライブ画像として端末装置200、300に送信する巻き戻しライブ画像送信部と、端末装置200、300から、巻き戻しライブ画像からの早送り要求を受け付けると、現時点の静止画像データに達するまで、前回送信した静止画像データより一定時間分未来の時点の静止画像データを一次映像フォルダ111から取得して、早送りライブ画像として端末装置200、300に送信する早送りライブ画像送信部と、して機能する。 In addition, when the program 113 receives a rewind request receiving function from the terminal devices 200 and 300 for receiving a rewind request from a live image, or when receiving a rewind request from a live image, the program 113 transmits the still image data transmitted last time. A rewind live image transmission function for acquiring still image data of a certain time in the past from the primary video folder 111 and transmitting the still image data to the terminal devices 200 and 300 as a rewind live image. A fast-forwarding request acceptance function that accepts a fast-forward request from a return live image, and when a fast-forward request from a rewind live image is accepted, the still image at a future point in time from the transmitted still image data until the current still image data is reached Acquire image data from the primary video folder 111 as a fast-forward live image Is a program for realizing the fast-forward live image transmission function of transmitting to the terminal device 200 and 300. When the CPU 101 loads the program 113 to the memory 102 and executes it, the server 100 accepts a rewind request from the live image from the terminal devices 200 and 300, and then the past time from the previously transmitted still image data. Still image data from the primary video folder 111 is transmitted from the primary video folder 111 and transmitted to the terminal devices 200 and 300 as a rewind live image. From the terminal devices 200 and 300, the rewind live image is transmitted from the rewind live image. When the fast-forwarding request is received, still image data at a time point that is a certain time later than the previously transmitted still image data is acquired from the primary video folder 111 until the current still image data is reached, and the terminal device 200 as a fast-forwarding live image is acquired. It functions as a fast-forward live image transmission unit that transmits to 300.
 また、プログラム113は、コンピュータに、端末装置200、300から過去の動画データの再生を受付する過去動画再生受付機能、受付した過去の動画データを二次映像フォルダ112から抽出し、抽出した動画データを端末装置200、300に送信する過去動画送信機能を実現させるためのプログラムである。 In addition, the program 113 extracts a past video reproduction reception function for accepting reproduction of past video data from the terminal devices 200 and 300 to the computer, extracts the received past video data from the secondary video folder 112, and extracts the extracted video data. Is a program for realizing a past moving image transmission function for transmitting to the terminal devices 200 and 300.
 また、プログラム113は、コンピュータに、1または複数台の監視カメラ400との接続を行う監視カメラ接続機能、1または複数台の端末装置200、300との接続を行う端末装置接続機能を実現させるためのプログラムである。 Further, the program 113 causes the computer to realize a monitoring camera connection function for connecting to one or a plurality of monitoring cameras 400, and a terminal device connection function for connecting to one or a plurality of terminal devices 200 and 300. It is a program.
 本実施の形態では、監視カメラ400とサーバ100との接続は、TCP/IP方式で行い、接続すべき監視カメラ400を特定するため、サーバ100側で設定したユーザIDとパスワードで認証を行ってから、カメラ映像の取得などの要求を行う。端末装置200、300とサーバ100との接続もTCP/IP方式で行い、ユーザIDとパスワードで認証を行って、端末装置200、300がサーバ100に登録してある端末装置であることを確認してから、カメラ映像の送信を行う。認証は、サーバ100上の認証用データベースによる認証が好ましい。 In the present embodiment, the connection between the monitoring camera 400 and the server 100 is performed using the TCP / IP method, and authentication is performed using the user ID and password set on the server 100 side in order to identify the monitoring camera 400 to be connected. To request camera image acquisition. The connection between the terminal devices 200 and 300 and the server 100 is also performed by the TCP / IP method, and authentication is performed with a user ID and a password, and it is confirmed that the terminal devices 200 and 300 are terminal devices registered in the server 100. Then send the camera video. The authentication is preferably performed by an authentication database on the server 100.
 端末装置200、300や監視カメラ400が同じLAN内に存在しなくても、ルータ600を介してネットワーク500で繋がっていれば、接続するルータ600のIPアドレスとそれぞれに割り振られたポート番号にて特定が行われる。インターネットで接続されているときは、両者の間のルータ600のポート番号が「有効」にされる。IPアドレスのないモバイル端末200ではUIDが利用される。 Even if the terminal devices 200 and 300 and the monitoring camera 400 do not exist in the same LAN, as long as they are connected to the network 500 via the router 600, the IP address of the router 600 to be connected and the port number assigned to each are used. Identification is done. When connected via the Internet, the port number of the router 600 between the two is made “valid”. The mobile terminal 200 having no IP address uses a UID.
 本実施の形態では、監視カメラ400からサーバ100への動画データの送信は、TCP/IPベースのRTSP(リアルストリーミングプロトコル)で行われる。本実施の形態では、動画データはH.264のハイプロファイルで圧縮して送信されるが、H.265等、次世代コーディックのハイプロファイルが用いられてもよい。 In the present embodiment, transmission of moving image data from the monitoring camera 400 to the server 100 is performed by TCP / IP-based RTSP (Real Streaming Protocol). In this embodiment, the moving image data is H.264. H.264 is compressed with high profile and transmitted. A high profile of a next generation codec such as H.265 may be used.
 次に、モバイル端末としての端末装置200の構成を説明する。図9は、端末装置200の概略構成を示す図である。 Next, the configuration of the terminal device 200 as a mobile terminal will be described. FIG. 9 is a diagram illustrating a schematic configuration of the terminal device 200.
 図9に示すように、端末装置200は、メモリ202を伴うCPU201やデバイスドライバなどを有する制御・演算装置と、記憶装置210と、データの送受信などを行う通信制御装置(図示しない)と、表示装置としてのディスプレイ204と、操作ボタンあるいはタッチパネルなどの出力装置(図示しない)とを備えている。記憶装置210には、画像表示のためのプログラム213やオペレーティングシステム214が格納されている。 As illustrated in FIG. 9, the terminal device 200 includes a control / arithmetic apparatus including a CPU 201 with a memory 202 and a device driver, a storage device 210, a communication control device (not shown) that transmits and receives data, a display, and the like. A display 204 as an apparatus and an output device (not shown) such as operation buttons or a touch panel are provided. The storage device 210 stores an image display program 213 and an operating system 214.
 端末装置200は、たとえばスマートフォンなどの携帯電話などであり、CPU201がプログラム213をメモリ202にロードして実行することにより、サーバ100から送信されたカメラ映像の表示が可能なコンピュータの機能を実現する。CPU201は、通常のモバイル端末に搭載される演算処理装置であり、各種プログラムを実行し、各種制御などを行う。 The terminal device 200 is, for example, a mobile phone such as a smartphone, and the CPU 201 loads the program 213 into the memory 202 and executes it, thereby realizing a computer function capable of displaying the camera video transmitted from the server 100. . The CPU 201 is an arithmetic processing device mounted on a normal mobile terminal, executes various programs, and performs various controls.
 次に、閲覧PCとしての端末装置300の構成を説明する。図10は、端末装置300の概略構成を示す図である。 Next, the configuration of the terminal device 300 as a browsing PC will be described. FIG. 10 is a diagram illustrating a schematic configuration of the terminal device 300.
 図10に示すように、端末装置300は、メモリ302を伴うCPU301やデバイスドライバなどを有する制御・演算装置と、DRAMなどの主記憶装置やハードディスクなどの補助記憶装置を有する記憶装置310と、ネットワークインターフェース304などの通信制御装置と、表示装置としてのディスプレイ303と、キーボード305、マウス306などの出力装置と、を有している。記憶装置310には、画像表示のためのプログラム313やオペレーティングシステム314が格納されている。 As shown in FIG. 10, a terminal device 300 includes a control / arithmetic unit having a CPU 301 and a device driver with a memory 302, a main storage device such as a DRAM, a storage device 310 having an auxiliary storage device such as a hard disk, and a network. A communication control device such as an interface 304, a display 303 as a display device, and output devices such as a keyboard 305 and a mouse 306 are included. The storage device 310 stores an image display program 313 and an operating system 314.
 端末装置300は、たとえばデスクトップコンピュータやノートブックコンピュータ、タブレット端末などであり、CPU301がプログラム313をメモリ302にロードして実行することにより、サーバ100から送信されたカメラ映像の表示が可能なコンピュータの機能を実現する。CPU301は、通常のPCに搭載される演算処理装置であり、各種プログラムを実行し、各種制御を行う。 The terminal device 300 is, for example, a desktop computer, a notebook computer, a tablet terminal, or the like. The CPU 301 loads a program 313 into the memory 302 and executes it, so that the camera image transmitted from the server 100 can be displayed. Realize the function. The CPU 301 is an arithmetic processing device mounted on a normal PC, executes various programs, and performs various controls.
 画像表示のためのプログラム213、313は、コンピュータに、サーバ100との接続を行うサーバ接続機能、監視権限のある1または複数台の監視カメラ400に係るカメラ映像の再生をサーバ100に要求するカメラ映像再生要求機能、サーバ100から送信されたカメラ映像を表示するカメラ映像表示機能を実現させるためのプログラムである。 The programs 213 and 313 for displaying images are a server connection function for connecting the server 100 to the computer, and a camera that requests the server 100 to play back the camera video related to one or a plurality of monitoring cameras 400 having the monitoring authority. This is a program for realizing a video playback request function and a camera video display function for displaying a camera video transmitted from the server 100.
 サーバ100は、ローカル環境にある監視カメラ400との接続開始時にIPアドレスおよびポート番号を用いた接続によって接続相手を特定し、ユーザIDとパスワードで認証する。サーバ100は、遠隔環境にある監視カメラ400との接続では、ルータ600のポート転送(ポートフォワーディングなどとも表現される)を用いることで、グローバルIPアドレスおよびポート番号を用いて接続相手を特定する。サーバ100は、端末装置200との接続開始時にUIDを用いた端末固有情報をもとに端末装置200を特定するため、ユーザIDとパスワードの認証および端末固有情報の一致によりカメラ映像の表示を可能とする。 The server 100 identifies the connection partner by connecting using the IP address and port number when starting the connection with the monitoring camera 400 in the local environment, and authenticates with the user ID and password. When connecting to the monitoring camera 400 in the remote environment, the server 100 uses the port forwarding (also expressed as port forwarding or the like) of the router 600 to identify the connection partner using the global IP address and the port number. Since the server 100 identifies the terminal device 200 based on the terminal-specific information using the UID at the start of connection with the terminal device 200, the camera image can be displayed by authenticating the user ID and the password and matching the terminal-specific information. And
 本実施の形態では、サーバ100と閲覧PCとしての端末装置300とは、ともにパーソナルコンピュータとして構成され、通常のパーソナルコンピュータが有するクロック機能などを備えている。モバイル端末としての端末装置200および監視カメラ400もクロック機能などを備えている。 In the present embodiment, the server 100 and the terminal device 300 as a viewing PC are both configured as personal computers and have a clock function or the like possessed by a normal personal computer. The terminal device 200 as a mobile terminal and the monitoring camera 400 also have a clock function and the like.
 次に、このような構成からなる監視カメラシステム10の動作の一例について、図11を参照して説明する。図11は、監視カメラシステム10の動作の一例を示す概略図である。なお、図11に示す例では、モバイル端末としての端末装置200が図示されているが、閲覧PCとしての端末装置300に置き換えても動作は同様である。 Next, an example of the operation of the monitoring camera system 10 having such a configuration will be described with reference to FIG. FIG. 11 is a schematic diagram illustrating an example of the operation of the monitoring camera system 10. In the example illustrated in FIG. 11, the terminal device 200 as a mobile terminal is illustrated, but the operation is the same even if the terminal device 300 is replaced with a browsing PC.
 図11を参照し、まず、サーバ100は、監視カメラ400にTCP/IP方式で接続を要求する。監視カメラ400は1または複数台の監視カメラで構成され、監視カメラ400とサーバ100とはユーザIDおよびパスワードで認証を行う。サーバ100の認証用データベースが参照され、認証が成功すると、監視カメラ400とサーバ100とは接続状態となる。 Referring to FIG. 11, first, the server 100 requests the monitoring camera 400 to connect using the TCP / IP method. The monitoring camera 400 includes one or more monitoring cameras, and the monitoring camera 400 and the server 100 authenticate with a user ID and a password. When the authentication database of the server 100 is referred to and the authentication is successful, the monitoring camera 400 and the server 100 are connected.
 次に、サーバ100は、監視カメラ400にカメラ映像を撮像して送信することを要求する。監視カメラ400は、サーバ100からの映像要求を受け付けると、カメラ映像を連続して撮像し、撮像して得られた未加工の映像データ(RAWデータ)をH.264のハイプロファイルの動画データに符号化する。より詳しくは、監視カメラ400は、RAWデータを、単独でデコーディング可能で一定時間ごとに作成されるイントラフレーム(Iフレーム)と、過去のフレームを参照してイントラフレームの間に複数作成される予測インターフレーム(Pフレーム)と、過去のフレームと未来のフレームの両方を参照して作成される双方向予測インターフレーム(Bフレーム)とを含む動画データに符号化する。そして、監視カメラ400は、符号化した動画データをストリーム方式でサーバ100に送信する。 Next, the server 100 requests the monitoring camera 400 to capture and transmit the camera video. When the monitoring camera 400 receives a video request from the server 100, it continuously captures the camera video, and converts the raw video data (RAW data) obtained by the imaging to H.264. H.264 high profile video data is encoded. More specifically, the surveillance camera 400 creates a plurality of RAW data between an intra frame (I frame) that can be decoded independently and created at regular intervals, and an intra frame with reference to a past frame. Encoding is performed into moving image data including a prediction inter frame (P frame) and a bidirectional prediction inter frame (B frame) created by referring to both past and future frames. Then, the monitoring camera 400 transmits the encoded moving image data to the server 100 by a stream method.
 また、サーバ100は、予め定められたスナップショット時間間隔で、監視カメラ400にライブ画像要求を送信することを繰り返す。スナップショット時間間隔は、予め任意で設定されており、たとえば10分の1秒、1秒、3秒あるいは60秒などである。以下に説明する実施例では、スナップショット時間間隔を1秒とする。監視カメラ400は、サーバ100からライブ画像要求を受け付ける毎に、未加工の映像データ(RAWデータ)からリアルタイムの静止画像データを作成し(切り出し)、ライブ画像としてサーバ100に送信する。 In addition, the server 100 repeatedly transmits a live image request to the monitoring camera 400 at a predetermined snapshot time interval. The snapshot time interval is arbitrarily set in advance, and is, for example, 1/10 second, 1 second, 3 seconds or 60 seconds. In the embodiment described below, the snapshot time interval is 1 second. Each time the monitoring camera 400 receives a live image request from the server 100, the monitoring camera 400 creates (cuts out) real-time still image data from raw video data (RAW data) and transmits it to the server 100 as a live image.
 サーバ100は、監視カメラ400からストリーム方式で動画データを取得すると、保存単位(たとえば10分)ごとに1つのファイルとして二次映像フォルダ112に格納する。これにより、過去の動画データの再生は動画データ保存後(たとえば10分後)から受付可能となる。 When the server 100 acquires the moving image data from the monitoring camera 400 by the stream method, the server 100 stores it in the secondary video folder 112 as one file for each storage unit (for example, 10 minutes). Thereby, the reproduction of the past moving image data can be accepted after the moving image data is stored (for example, after 10 minutes).
 本実施の形態では、サーバ100は、監視カメラ400から受け取る動画データを、エンコード部402での符号化に係る遅延に応じて時間調整して二次映像フォルダ112に保存する。たとえば、監視カメラ400から送信される動画データが、動画データと同時に送信されるリアルタイムの静止画像データに対して1分遅延している場合には、サーバ100は、たとえば9:51:00~10:00:59に受け取る動画データに、9:50:00~9:59:59のタイムスタンプを付して1つのファイルとして二次映像フォルダ112に保存する。 In the present embodiment, the server 100 saves the moving image data received from the monitoring camera 400 in the secondary video folder 112 by adjusting the time according to the delay related to the encoding in the encoding unit 402. For example, when the moving image data transmitted from the monitoring camera 400 is delayed by one minute with respect to the real-time still image data transmitted simultaneously with the moving image data, the server 100, for example, 9: 51: 00-10 The moving image data received at 0:00:59 is saved in the secondary video folder 112 as one file with a time stamp of 9:55:00 to 9:59:59.
 また、サーバ100は、監視カメラ400からライブ画像としての静止画像データを取得すると、予め定められた閲覧期間(たとえば24時間)は、頻繁に見る可能性がある高解像度のデータとして一次映像フォルダ111に保存し、閲覧期間を過ぎると、一次映像フォルダ111から削除する。 In addition, when the server 100 acquires still image data as a live image from the monitoring camera 400, the primary video folder 111 is displayed as high-resolution data that may be frequently viewed during a predetermined browsing period (for example, 24 hours). When the viewing period is over, the primary video folder 111 is deleted.
 端末装置200は、ユーザによるサーバ100への接続要求を受付ける。サーバ100への接続要求が端末装置200に入力されると、端末装置200は、サーバ100にTCP/IP方式で接続を要求する。モバイル端末200では、部分的に携帯電話独自のプロトコルを使用して接続することもある。端末装置200は、1または複数台の端末装置で構成され、端末装置200とサーバ100とはユーザIDおよびパスワードで認証を行う。サーバ100の認証用データベースが参照され、端末装置200の認証登録がされていない場合は、初回のみ暗号化キーを発行して登録が行われ、登録時にユーザIDとパスワードが発行される。認証が成功すると、端末装置200とサーバ100とは接続状態となる。 The terminal device 200 accepts a connection request to the server 100 from the user. When a connection request to the server 100 is input to the terminal device 200, the terminal device 200 requests the server 100 to connect using the TCP / IP method. In the mobile terminal 200, connection may be made partially using a protocol unique to the mobile phone. The terminal device 200 includes one or a plurality of terminal devices, and the terminal device 200 and the server 100 authenticate with a user ID and a password. When the authentication database of the server 100 is referred to and the terminal device 200 is not registered for authentication, registration is performed by issuing an encryption key only for the first time, and a user ID and password are issued during registration. If the authentication is successful, the terminal device 200 and the server 100 are connected.
 端末装置200は、ユーザからのカメラ一覧要求を受け付ける。端末装置200は、カメラ一覧要求が入力されると、サーバ100にカメラ一覧要求を送信する。サーバ100は、カメラ一覧要求を受信すると、要求した端末装置200が監視できる監視カメラ400についての権限情報を確認する。サーバ100は、権限のあるカメラ一覧データを端末装置200に送信する。端末装置200は、受信したカメラ一覧データをディスプレイ203に表示し、ライブ再生させるべき監視カメラについてのユーザからの選択を受け付ける。 The terminal device 200 receives a camera list request from the user. When the camera list request is input, the terminal device 200 transmits the camera list request to the server 100. When the server 100 receives the camera list request, the server 100 checks authority information about the monitoring camera 400 that can be monitored by the requested terminal device 200. The server 100 transmits authorized camera list data to the terminal device 200. The terminal device 200 displays the received camera list data on the display 203 and accepts a selection from the user regarding the monitoring camera to be played live.
 端末装置200は、ユーザからの監視カメラの選択が入力されると、選択された監視カメラについてカメラ映像のライブ再生をサーバ100に要求する。サーバ100は、ライブ再生の要求を受け付けると、選択された監視カメラ400から取得して一次映像フォルダ111に保存された最新の静止画像データを、ライブ画像として端末装置200へ送信する。 When the terminal device 200 receives the selection of the monitoring camera from the user, the terminal device 200 requests the server 100 to perform live playback of the camera video for the selected monitoring camera. Upon receiving a live playback request, the server 100 transmits the latest still image data acquired from the selected monitoring camera 400 and stored in the primary video folder 111 to the terminal device 200 as a live image.
 端末装置200は、サーバ100から受け取る静止画像データを、ライブ画像としてディスプレイ203に表示する。ユーザによって監視カメラの選択が解除されるまで、あるいはライブ再生以外が選択されている間を除き、端末装置200は、監視カメラ400におけるスナップショット時間間隔と同じ時間間隔で、ライブ再生の要求を繰り返す。 The terminal device 200 displays the still image data received from the server 100 on the display 203 as a live image. The terminal device 200 repeats the request for live playback at the same time interval as the snapshot time interval in the monitoring camera 400 until the user deselects the monitoring camera or while other than live playback is selected. .
 端末装置200は、ライブ画像の表示中に、ユーザからの巻き戻し要求の入力を受け付ける。端末装置200は、ユーザから巻き戻し要求が入力されると、サーバ100に巻き戻し画像を要求する。 The terminal device 200 receives an input of a rewind request from the user while the live image is displayed. When a rewind request is input from the user, the terminal device 200 requests a rewind image from the server 100.
 サーバ100は、巻き戻し画像の要求を受信すると、前回送信した静止画像データより一定時間分過去の時点の静止画像データを、一次映像フォルダ111から抽出し、巻き戻しライブ画像として端末装置200へ送信する。本実施例では、一定時間分過去の時点を1秒分過去の時点とし、ライブ画像が1秒ごとに取得した静止画像データであるので、最初に巻き戻し要求が入力されたときは、1枚前の静止画像データが端末装置200に送信される。 When the server 100 receives the rewind image request, the server 100 extracts still image data at a point in time that is a predetermined time earlier than the previously transmitted still image data from the primary video folder 111 and transmits it to the terminal device 200 as a rewind live image. To do. In this embodiment, since the past time for a certain time is set to the past time for one second, and the live image is still image data acquired every second, when one rewind request is input for the first time, The previous still image data is transmitted to the terminal device 200.
 巻き戻しが選択されている間中、すなわち、巻き戻しライブ画像の表示中は、端末装置200は、サーバ100に、一定時間ごとに巻き戻しライブ画像を要求する。その際の一定時間は、ライブ画像の要求間隔より短い一定時間であって、たとえば0.2秒ごとである。したがって。端末装置200には、0.2秒ごとに、1秒ずつ前のカメラ映像の静止画像データから作成された巻き戻しライブ画像が表示されることになる。 While the rewind is selected, that is, while the rewind live image is being displayed, the terminal device 200 requests the rewind live image from the server 100 at regular intervals. The fixed time at that time is a fixed time shorter than the request interval of the live image, for example, every 0.2 seconds. Therefore. On the terminal device 200, a rewind live image created from still image data of the previous camera image by 1 second is displayed every 0.2 seconds.
 さらに、端末装置200は、巻き戻しライブ画像の表示中に、ユーザからの早送り要求の入力を受け付ける。端末装置200は、早送り要求が入力されると、サーバ100に早送りライブ画像を要求する。サーバ100は、早送りライブ画像の要求を受信すると、前回送信した静止画像データより一定時間分未来の時点の静止画像データを、一次映像フォルダ111から抽出し、早送りライブ画像として端末装置200に送信する。本実施の形態では、一定時間分未来の時点を1秒分未来の時点とし、ライブ画像が1秒ごとに取得した画像であるので、最初に早送り要求が入力されたときは、巻き戻しライブ画像として現在表示されている静止画像データより、1秒未来の静止画像データが送信される。 Furthermore, the terminal device 200 accepts an input of a fast forward request from the user while displaying the rewind live image. When the fast forward request is input, the terminal device 200 requests the server 100 for a fast forward live image. When the server 100 receives the fast-forward live image request, the server 100 extracts still image data at a time point that is a certain time later than the previously transmitted still image data from the primary video folder 111 and transmits it to the terminal device 200 as a fast-forward live image. . In the present embodiment, the future time is set to the future time of 1 second, and the live image is an image acquired every second. Therefore, when a fast-forward request is input for the first time, the rewind live image From the still image data that is currently displayed, still image data that is one second in the future is transmitted.
 早送りが選択されている間中、すなわち、早送りライブ画像の表示中は、端末装置200は、サーバ100に、一定時間ごとに早送りライブ画像を要求する。その際の一定時間は、ライブ画像の要求間隔より短い一定時間であって、たとえば0.2秒ごとである。したがって、端末装置200には、0.2秒ごとに、1秒ずつ後のカメラ映像の静止画像データから作成された早送りライブ画像が表示されることになる。早送りライブ画像が現時点の画像までたどり着いたら、ライブ画像表示に切り替わり、早送りは終了する。 While the fast forward is selected, that is, while the fast forward live image is being displayed, the terminal device 200 requests the server 100 for a fast forward live image at regular intervals. The fixed time at that time is a fixed time shorter than the request interval of the live image, for example, every 0.2 seconds. Accordingly, the fast-forward live image created from the still image data of the camera video after 1 second is displayed on the terminal device 200 every 0.2 seconds. When the fast-forward live image reaches the current image, the live image display is switched to, and the fast-forward is finished.
 ライブ再生に代えて、過去の動画データの再生も同様に行うことができる。サーバ100は、静止画像データの代わりに、受付した過去の動画データを、二次映像フォルダ112から抽出し、抽出した動画データを端末装置200に送信して表示させる。 Instead of live playback, past video data can be played in the same way. The server 100 extracts the received past moving image data from the secondary video folder 112 instead of the still image data, and transmits the extracted moving image data to the terminal device 200 for display.
 静止画像データは、画像であって端末装置200に連続して送信できるほど軽く、また、動画データも、ハイプロファイルで符号化されていて、動画でありながらデータ量が極めて小さいため、端末装置200に送信できるほど軽い。また、静止画像データも、動画データも、ともに、巻き戻し、巻き戻してからの再生や早送りが可能であり、利便性が高い。 The still image data is an image that is light enough to be continuously transmitted to the terminal device 200, and the moving image data is encoded with a high profile and the amount of data is extremely small while being a moving image. Light enough to send to. In addition, both still image data and moving image data can be rewinded, played back after rewinding, and fast-forwarded, which is highly convenient.
 ところで、上記でも言及したように、従来の監視カメラシステムでは、動画データの保存をベースラインプロファイルで行っているため、容量の圧縮により映像が劣化したり、可変レートの場合は瞬時の動きに映像が乱れたりするなど、仕組み自体に限界が来ていた。 By the way, as mentioned above, in the conventional surveillance camera system, the video data is stored in the baseline profile, so the video deteriorates due to the compression of the capacity, or in the case of the variable rate, the video is instantaneously moved. There was a limit to the mechanism itself, such as being disturbed.
 また、ベースラインプロファイルの動画データをハイプロファイルに変換して保存する構成では、最適化したデータ保存が可能とされるものの、レコーダとなるサーバのCPUリソースを使用して変換を行うため、システムにかかるコストが大きくなる傾向があった。 In addition, in the configuration in which the video data of the baseline profile is converted to a high profile and saved, the optimized data can be saved, but the conversion is performed using the CPU resource of the server serving as a recorder. There was a tendency for such costs to increase.
 また、カメラからのストリーム出力はベースラインプロファイルに限定されていることから、ネットワークに複数台のカメラを接続すると、トラフィックが増大し、ネットワーク設計や機器への負荷が発生していた。 Also, since the stream output from the camera is limited to the baseline profile, connecting multiple cameras to the network increased the traffic, causing a load on the network design and equipment.
 これに対し、本実施の形態によれば、監視カメラ400のエンコード部402が、未加工の映像データを、双方向予測インターフレーム(Bフレーム)を含む動画データに符号化してサーバに送信するため、サーバに負荷をかけることなく、最適化したデータ保存が可能となる。また、ネットワーク500上を流れるデータ量が大幅に小さくなり、ネットワークトラフィックを低減することができる。これにより、ネットワーク設計や機器への負荷が低減し、ネットワーク機器を長寿命化できる。また、ネットワーク500に複数台の監視カメラ400を接続することが可能となる。 On the other hand, according to the present embodiment, the encoding unit 402 of the monitoring camera 400 encodes raw video data into moving image data including a bidirectional prediction inter frame (B frame) and transmits the encoded video data to the server. Optimized data storage is possible without imposing a load on the server. In addition, the amount of data flowing on the network 500 is significantly reduced, and network traffic can be reduced. As a result, the load on the network design and equipment is reduced, and the life of the network equipment can be extended. In addition, a plurality of surveillance cameras 400 can be connected to the network 500.
 また、本実施の形態によれば、双方向予測インターフレーム(Bフレーム)を含む動画データへの符号化が、サーバ100のCPUリソースを使用して行われるのではなく、監視カメラ400の内部で行われるため、サーバ100の負荷が低減し、システムにかかるコストを低減できる。また、サーバ100の負荷が低減するので、サーバ1台当たりの監視カメラ400の収容台数を増やすことができる。 Further, according to the present embodiment, encoding into moving image data including a bi-predictive inter frame (B frame) is not performed using the CPU resource of the server 100, but inside the surveillance camera 400. As a result, the load on the server 100 is reduced and the cost of the system can be reduced. Further, since the load on the server 100 is reduced, the number of monitoring cameras 400 per server can be increased.
 また、本実施の形態によれば、監視カメラ400から送信される動画データは、双方向予測インターフレーム(Bフレーム)を含むことから、必然的に遅延しているものの、ライブ映像生成部403が、遅延のない未加工の映像データからリアルタイムの映像データを作成してサーバ100に送信するため、サーバ100は、リアルタイムの映像を取得可能であり、リアルタイムでの閲覧という監視カメラシステムとしての必須の要件は何ら損なわれることがない。 In addition, according to the present embodiment, the moving image data transmitted from the monitoring camera 400 includes a bidirectional prediction inter frame (B frame), and thus the live video generation unit 403 inevitably delays. Since the real-time video data is generated from the raw video data without delay and transmitted to the server 100, the server 100 can acquire the real-time video, and is essential as a surveillance camera system for browsing in real time. There is no loss of requirements.
 また、本実施の形態によれば、監視カメラ400から送信される動画データはエンコード部402での符号化により必然的に遅延しているものの、二次映像フォルダ112には遅延に応じて時間調整された状態で保存されるため、サーバ100は、端末装置200から時刻を特定して過去の動画データの再生要求を受け付けた場合に、要求された時刻の動画データを二次映像フォルダ112から正確に抽出して端末装置200に送信することができる。 In addition, according to the present embodiment, the moving image data transmitted from the monitoring camera 400 is necessarily delayed due to encoding by the encoding unit 402, but the secondary video folder 112 is time-adjusted according to the delay. Therefore, when the server 100 specifies a time from the terminal device 200 and receives a playback request for past video data, the server 100 accurately stores the video data at the requested time from the secondary video folder 112. And transmitted to the terminal device 200.
 また、本実施の形態によれば、端末装置200にてライブ画像を表示しているときに、ユーザが見落とし等に気付いた場合には、端末装置200からサーバ100に巻き戻し要求を送信することで、端末装置200ではそのままさかのぼって確認することができ、また、端末装置200からサーバ100に早送り要求を送信することで、過去から現時点までを早送りで確認することができる。 In addition, according to the present embodiment, when a user notices an oversight or the like while displaying a live image on the terminal device 200, a rewind request is transmitted from the terminal device 200 to the server 100. Thus, the terminal device 200 can check back as it is, and by transmitting a fast-forward request from the terminal device 200 to the server 100, it is possible to confirm from the past to the present time by fast-forward.
 なお、上述した実施の形態に対して様々な変更を加えることが可能である。以下、図面を参照しながら、変形の一例について説明する。以下の説明および以下の説明で用いる図面では、上述した実施の形態と同様に構成され得る部分について、上述した実施の形態における対応する部分に対して用いた符号と同一の符号を用いるとともに、重複する説明を省略する。 Note that various modifications can be made to the above-described embodiment. Hereinafter, an example of modification will be described with reference to the drawings. In the following description and the drawings used in the following description, for parts that can be configured in the same manner as the above-described embodiment, the same reference numerals as those used for the corresponding parts in the above-described embodiment are used, and overlapping Description to be omitted is omitted.
(第1の変形例)
 第1の変形例について、図12を参照して説明する。図12は、第1の変形例に係る監視カメラ400の概略的な構成を示す図である。
(First modification)
A first modification will be described with reference to FIG. FIG. 12 is a diagram illustrating a schematic configuration of a monitoring camera 400 according to the first modification.
 図12に示すように、第1の変形例では、監視カメラ400のエンコード部402Bは、撮像部401にて生成された未加工の映像データ(RAWデータ)を、Bフレームを含む動画データ(たとえば、H.264のハイプロファイルの動画データ)に符号化したのち、動画データからIフレームが先頭の画像となるGOP単位(たとえば1秒分)で、撮影時刻順に分割データを作成してメモリ4025に保存する。 As shown in FIG. 12, in the first modified example, the encoding unit 402B of the monitoring camera 400 converts raw video data (RAW data) generated by the imaging unit 401 into moving image data including B frames (for example, H.264 high profile video data), and then, in the GOP unit (for example, 1 second) in which the I frame is the first image from the video data, divided data is created in the order of shooting time and stored in the memory 4025. save.
 なお、GOP(Group Of Picture)は、1つ以上のIフレーム(イントラフレーム)と複数のPフレーム(予測インターフレーム)およびBフレーム(双方向予測インターフレーム)からなるフレーム群である。エンコード部402は、たとえば、動画データからIフレームを検出すると、次のIフレームを検出した直前のフレームまでを1つの分割データとする。この場合、1つの分割データが1つのIフレームを含むことになる。なお、1つの分割データが2つ以上のIフレームを含んでもよい。 Note that GOP (Group Of Picture) is a frame group including one or more I frames (intra frames), a plurality of P frames (prediction interframes), and B frames (bidirectional prediction interframes). For example, when detecting an I frame from the moving image data, the encoding unit 402 sets one frame of data up to the frame immediately before the next I frame is detected. In this case, one divided data includes one I frame. One divided data may include two or more I frames.
 また、監視カメラ400のエンコード部402Bは、サーバ100から分割データ要求を受け付ける毎に、メモリ4025からサーバ100に1つずつ分割データを転送する。 Moreover, every time the encoding unit 402B of the monitoring camera 400 receives a divided data request from the server 100, the encoded data is transferred from the memory 4025 to the server 100 one by one.
 一方、サーバ100は、監視カメラ400に分割データ要求を送信して分割データを受け取ることを繰り返す。そして、サーバ100は、監視カメラ400から受け取る分割データを撮影時刻順に結合して、保存単位(たとえば10分)ごとに1つのファイルの動画データを作成し、二次映像フォルダ112に保存する。 On the other hand, the server 100 repeatedly transmits the divided data request to the monitoring camera 400 and receives the divided data. Then, the server 100 combines the divided data received from the monitoring camera 400 in the order of shooting time, creates one file of moving image data for each storage unit (for example, 10 minutes), and stores it in the secondary video folder 112.
 以上のような第1の変形例によれば、たとえばクラウド環境にサーバ100を設置して遠隔地の監視カメラ400から映像データをサーバ100に転送する際に、後で元の映像を復元可能な容量が小さい単位に分割して、ネットワーク500の混雑を回避しつつ送信できるので、クラウド環境で安定的なカメラ映像の収集が可能となる。 According to the first modification as described above, when the server 100 is installed in a cloud environment and video data is transferred from the remote monitoring camera 400 to the server 100, the original video can be restored later. Since it can be divided into smaller units and transmitted while avoiding congestion of the network 500, it is possible to collect camera images stably in a cloud environment.
(第2の変形例)
 第2の変形例について、図13を参照して説明する。図13は、第2の変形例に係る監視カメラ400の概略的な構成を示す図である。
(Second modification)
A second modification will be described with reference to FIG. FIG. 13 is a diagram illustrating a schematic configuration of a monitoring camera 400 according to the second modification.
 図13に示すように、第2の変形例では、監視カメラ400のライブ映像生成部403Bは、未加工の映像データ(RAWデータ)を、双方向予測インターフレーム(Bフレーム)を含まないリアルタイムの動画データ、すなわち単独でデコーディング可能で一定時間ごとに作成されるイントラフレーム(Iフレーム)と、過去のフレームを参照してイントラフレームの間に複数作成される予測インターフレーム(Pフレーム)とから構成されるベースラインプロファイルの動画データに符号化して、ストリーム方式でサーバ100に送信する。 As shown in FIG. 13, in the second modified example, the live video generation unit 403B of the monitoring camera 400 converts the raw video data (RAW data) into real-time data that does not include a bidirectional prediction interframe (B frame). From moving image data, that is, an intra frame (I frame) that can be decoded independently and created at regular intervals, and a plurality of predicted inter frames (P frames) created between intra frames with reference to past frames The encoded video data of the baseline profile is encoded and transmitted to the server 100 in a stream format.
 第2の変形例によっても、上述した実施の形態と同様に、監視カメラ400のエンコード部402が未加工の映像データを、双方向予測インターフレーム(Bフレーム)を含む動画データに符号化してサーバ100に送信するため、サーバ100に負荷をかけることなく最適化したデータ保存が可能となるとともに、ライブ映像生成部403が、遅延のない未加工の映像データからリアルタイムの映像データを作成してサーバ100に送信するため、サーバ100はリアルタイムの映像を取得可能であり、リアルタイムでの閲覧という監視カメラシステムとしての必須の要件は何ら損なわれることがない。 Also according to the second modified example, as in the above-described embodiment, the encoding unit 402 of the surveillance camera 400 encodes raw video data into moving image data including a bi-directional prediction interframe (B frame). Because the data is transmitted to 100, optimized data storage can be performed without imposing a load on the server 100, and the live video generation unit 403 creates real-time video data from unprocessed video data without delay, and the server Therefore, the server 100 can acquire a real-time video, and the essential requirement as a surveillance camera system for real-time browsing is not impaired.
 なお、第2の変形例によれば、監視カメラ400がリアルタイムの映像をストリーム方式で送出するので、上述した実施の形態に比べてネットワーク500のトラフィックが増える可能性があるものの、ローカル環境であれば問題にならない。 According to the second modified example, since the monitoring camera 400 transmits real-time video in a stream format, traffic on the network 500 may increase as compared with the above-described embodiment, but it may be in a local environment. Will not be a problem.
(第3の変形例)
 第3の変形例について、図14を参照して説明する。図14は、第3の変形例に係る監視カメラ400の概略的な構成を示す図である。
(Third Modification)
A third modification will be described with reference to FIG. FIG. 14 is a diagram illustrating a schematic configuration of a monitoring camera 400 according to the third modification.
 図14に示すように、第3の変形例では、監視カメラ400のライブ映像生成部403Cは、タイマー409などに予め定められたスナップショット時間間隔(たとえば、1秒ごと)で、未加工の映像データ(RAWデータ)からリアルタイムの静止画像データを切り出してサーバ100に送信することを繰り返す。監視カメラ400からサーバ100への静止画像データの転送では、FTP、HTTPなどのTCP/IPベースの転送方式が用いられる。 As shown in FIG. 14, in the third modified example, the live video generation unit 403C of the monitoring camera 400 performs unprocessed video at a snapshot time interval (for example, every second) set in advance by the timer 409 or the like. It repeats that real-time still image data is cut out from data (RAW data) and transmitted to the server 100. In the transfer of still image data from the monitoring camera 400 to the server 100, a TCP / IP-based transfer method such as FTP or HTTP is used.
 第3の変形例によっても、上述した実施の形態と同様に、監視カメラ400のエンコード部402が未加工の映像データを、双方向予測インターフレーム(Bフレーム)を含む動画データに符号化してサーバ100に送信するため、サーバ100に負荷をかけることなく最適化したデータ保存が可能となるとともに、ライブ映像生成部403が、遅延のない未加工の映像データからリアルタイムの映像データを作成してサーバ100に送信するため、サーバ100はリアルタイムの映像を取得可能であり、リアルタイムでの閲覧という監視カメラシステムとしての必須の要件は何ら損なわれることがない。 Also according to the third modified example, as in the above-described embodiment, the encoding unit 402 of the surveillance camera 400 encodes raw video data into moving image data including a bi-directional prediction interframe (B frame). Because the data is transmitted to 100, optimized data storage can be performed without imposing a load on the server 100, and the live video generation unit 403 creates real-time video data from unprocessed video data without delay, and the server Therefore, the server 100 can acquire a real-time video, and the essential requirement as a surveillance camera system for real-time browsing is not impaired.
 なお、上述した実施の形態および個々の変形例の記載ならびに図面の開示は、特許請求の範囲に記載された発明を説明するための一例に過ぎず、上述した実施の形態および個々の変形例の記載または図面の開示によって特許請求の範囲に記載された発明が限定されることはない。上述した実施の形態および個々の変形例の構成要素は、発明の主旨を逸脱しない範囲で任意に組み合わせることが可能である。 The description of the above-described embodiments and individual modifications and the disclosure of the drawings are merely examples for explaining the invention described in the claims, and the description of the above-described embodiments and individual modifications. The invention described in the scope of claims is not limited by the description or the disclosure of the drawings. The components of the above-described embodiments and individual modifications can be arbitrarily combined without departing from the gist of the invention.

Claims (10)

  1.  監視カメラと、
     前記監視カメラにネットワークを介して接続されたサーバと、
    を備え、
     前記監視カメラは、
     受光した映像を撮像して未加工の映像データを生成する撮像部と、
     前記未加工の映像データを、過去のフレームと未来のフレームの両方を参照して作成される双方向予測インターフレームを含む動画データに符号化して、前記サーバに送信するエンコード部と、
     前記未加工の映像データからリアルタイムの映像データを作成して、前記サーバに送信するライブ映像生成部と、
    を有する、
    ことを特徴とする監視カメラシステム。
    A surveillance camera,
    A server connected to the surveillance camera via a network;
    With
    The surveillance camera is
    An imaging unit that captures the received video and generates raw video data;
    An encoding unit that encodes the raw video data into video data including a bi-directional prediction interframe created by referring to both past and future frames, and transmits the encoded video data to the server;
    Creating a real-time video data from the raw video data and transmitting it to the server;
    Having
    A surveillance camera system characterized by that.
  2.  前記エンコード部は、前記動画データをストリーム方式で前記サーバに送信する
    ことを特徴とする請求項1に記載の監視カメラシステム。
    The surveillance camera system according to claim 1, wherein the encoding unit transmits the moving image data to the server by a stream method.
  3.  前記エンコード部は、前記動画データからGOP単位で撮影時刻順に分割データを作成してメモリに保存し、前記サーバから分割データ要求を受け付ける毎に、前記メモリから前記サーバに1つずつ分割データを転送し、
     前記サーバは、前記監視カメラから受け取る分割データを撮影時刻順に結合して動画データを作成する
    ことを特徴とする請求項1に記載の監視カメラシステム。
    The encoding unit creates divided data in order of shooting time from the moving image data in GOP units, stores the divided data in a memory, and transfers the divided data from the memory to the server one by one each time a divided data request is received from the server. And
    The surveillance server system according to claim 1, wherein the server creates moving image data by combining divided data received from the surveillance camera in order of photographing time.
  4.  前記ライブ映像生成部は、前記サーバからライブ画像要求を受け付ける毎に、前記未加工の映像データからリアルタイムの静止画像データを切り出して前記サーバに送信する
    ことを特徴とする請求項1~3のいずれかに記載の監視カメラシステム。
    The live video generation unit cuts out real-time still image data from the raw video data and transmits the live video request to the server every time a live image request is received from the server. Surveillance camera system according to Crab.
  5.  前記ライブ映像生成部は、前記未加工の映像データを、前記双方向予測インターフレームを含まないリアルタイムの動画データに符号化して、ストリーム方式で前記サーバに送信する
    ことを特徴とする請求項1~3のいずれかに記載の監視カメラシステム。
    The live video generation unit encodes the raw video data into real-time video data not including the bidirectional prediction interframe, and transmits the encoded video data to the server in a stream format. 4. The surveillance camera system according to any one of 3.
  6.  前記ライブ映像生成部は、予め定められたスナップショット時間間隔で、前記未加工の映像データからリアルタイムの静止画像データを切り出して前記サーバに送信することを繰り返す
    ことを特徴とする請求項1~3のいずれかに記載の監視カメラシステム。
    The live video generation unit repeats cutting out real-time still image data from the raw video data and transmitting it to the server at predetermined snapshot time intervals. The surveillance camera system according to any one of the above.
  7.  前記サーバは、
     前記監視カメラから受け取る静止画像データを記憶装置に保存する一次映像保存部と、
     ネットワークを介して接続された端末装置に前記静止画像データをライブ画像として送信するライブ画像送信部と、
     前記端末装置から、ライブ画像からの巻き戻し要求を受け付けると、前回送信した静止画像データより一定時間分過去の時点の静止画像データを前記記憶装置から取得して、巻き戻しライブ画像として前記端末装置に送信する巻き戻しライブ画像送信部と、
     前記端末装置から、巻き戻しライブ画像からの早送り要求を受け付けると、現時点の静止画像データに達するまで、前回送信した静止画像データより一定時間分未来の時点の静止画像データを前記記憶装置から取得して、早送りライブ画像として前記端末装置に送信する早送りライブ画像送信部と、
    を有する、
    ことを特徴とする請求項4または6に記載の監視カメラシステム。
    The server
    A primary video storage unit for storing still image data received from the surveillance camera in a storage device;
    A live image transmission unit that transmits the still image data as a live image to a terminal device connected via a network;
    When receiving a rewind request from a live image from the terminal device, the terminal device acquires still image data at a time point that is a certain time past from the previously transmitted still image data from the storage device, and the terminal device as a rewind live image. A rewind live image transmitter to transmit to
    When a fast-forward request from a rewind live image is received from the terminal device, still image data at a time point that is a certain time later than the previously transmitted still image data is acquired from the storage device until the current still image data is reached. A fast-forward live image transmission unit for transmitting to the terminal device as a fast-forward live image;
    Having
    The surveillance camera system according to claim 4 or 6, wherein
  8.  前記サーバは、前記監視カメラから受け取る動画データを、前記エンコード部での符号化に係る遅延に応じて時間調整して記憶装置に保存する二次映像保存部を有する、
    ことを特徴とする請求項1~7のいずれかに記載の監視カメラシステム。
    The server includes a secondary video storage unit that adjusts time according to a delay related to encoding in the encoding unit and stores the video data received from the monitoring camera in a storage device.
    The surveillance camera system according to any one of claims 1 to 7, characterized in that:
  9.  サーバにネットワークを介して接続される監視カメラであって、
     受光した映像を撮像して未加工の映像データを生成する撮像部と、
     前記未加工の映像データを、過去のフレームと未来のフレームの両方を参照して作成される双方向予測インターフレームを含む動画データに符号化して、前記サーバに送信するエンコード部と、
     前記未加工の映像データからリアルタイムの映像データを作成して、前記サーバに送信するライブ映像生成部と、
    を有する、
    ことを特徴とする監視カメラ。
    A surveillance camera connected to a server via a network,
    An imaging unit that captures the received video and generates raw video data;
    An encoding unit that encodes the raw video data into video data including a bi-directional prediction interframe created by referring to both past and future frames, and transmits the encoded video data to the server;
    Creating a real-time video data from the raw video data and transmitting it to the server;
    Having
    A surveillance camera characterized by that.
  10.  受光した映像を撮像して未加工の映像データを生成する撮像部と、
     前記未加工の映像データを、過去のフレームと未来のフレームの両方を参照して作成される双方向予測インターフレームを含む動画データに符号化して、前記サーバに送信するエンコード部と、
     前記未加工の映像データからリアルタイムの映像データを作成して、前記サーバに送信するライブ映像生成部と、
    を有する監視カメラにネットワークを介して接続されるサーバであって、
     前記監視カメラから受け取るリアルタイムの映像データを、記憶装置に保存する一次映像保存部と、
     前記監視カメラから受け取る動画データを、前記エンコード部での符号化に係る遅延に応じて時間調整して前記記憶装置に保存する二次映像保存部と、
    を備えたことを特徴とするサーバ。
    An imaging unit that captures the received video and generates raw video data;
    An encoding unit that encodes the raw video data into video data including a bi-directional prediction interframe created by referring to both past and future frames, and transmits the encoded video data to the server;
    Creating a real-time video data from the raw video data and transmitting it to the server;
    A server connected to a surveillance camera via a network,
    A primary video storage unit that stores real-time video data received from the surveillance camera in a storage device;
    A secondary video storage unit that saves the video data received from the monitoring camera in the storage device by adjusting the time according to the delay related to the encoding in the encoding unit;
    A server characterized by comprising:
PCT/JP2017/006251 2016-12-27 2017-02-21 Monitoring camera system WO2018123078A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2016-252729 2016-12-27
JP2016252729A JP2018107655A (en) 2016-12-27 2016-12-27 Monitoring camera system

Publications (1)

Publication Number Publication Date
WO2018123078A1 true WO2018123078A1 (en) 2018-07-05

Family

ID=62707129

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2017/006251 WO2018123078A1 (en) 2016-12-27 2017-02-21 Monitoring camera system

Country Status (3)

Country Link
JP (1) JP2018107655A (en)
TW (1) TW201824850A (en)
WO (1) WO2018123078A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022148446A1 (en) * 2021-01-08 2022-07-14 华为技术有限公司 Image processing method and apparatus, device, and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2003158731A (en) * 2001-11-20 2003-05-30 Hitachi Ltd Image delivery system, camera terminal, and receiving terminal
JP2004320822A (en) * 2002-10-31 2004-11-11 Hitachi Kokusai Electric Inc Method and system for storing video data
JP2005333520A (en) * 2004-05-21 2005-12-02 Mitsubishi Electric Corp Image transmission apparatus, image transmission method, transmission system, and video monitoring system
JP2012090172A (en) * 2010-10-21 2012-05-10 Canon Inc Monitoring camera, monitoring camera control method, and network camera system
JP2013051534A (en) * 2011-08-31 2013-03-14 Nexpoint Co Ltd Monitoring camera system and monitoring method
WO2015118664A1 (en) * 2014-02-07 2015-08-13 日立マクセル株式会社 Image transmission device, image reception device, and surveillance camera system, teleconference system, and vehicle-mounted camera system using same

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2003158731A (en) * 2001-11-20 2003-05-30 Hitachi Ltd Image delivery system, camera terminal, and receiving terminal
JP2004320822A (en) * 2002-10-31 2004-11-11 Hitachi Kokusai Electric Inc Method and system for storing video data
JP2005333520A (en) * 2004-05-21 2005-12-02 Mitsubishi Electric Corp Image transmission apparatus, image transmission method, transmission system, and video monitoring system
JP2012090172A (en) * 2010-10-21 2012-05-10 Canon Inc Monitoring camera, monitoring camera control method, and network camera system
JP2013051534A (en) * 2011-08-31 2013-03-14 Nexpoint Co Ltd Monitoring camera system and monitoring method
WO2015118664A1 (en) * 2014-02-07 2015-08-13 日立マクセル株式会社 Image transmission device, image reception device, and surveillance camera system, teleconference system, and vehicle-mounted camera system using same

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022148446A1 (en) * 2021-01-08 2022-07-14 华为技术有限公司 Image processing method and apparatus, device, and storage medium

Also Published As

Publication number Publication date
TW201824850A (en) 2018-07-01
JP2018107655A (en) 2018-07-05

Similar Documents

Publication Publication Date Title
US10972519B2 (en) Real-time video streaming to client video element
US10009630B2 (en) System and method for encoding video content using virtual intra-frames
JP5854194B2 (en) Surveillance camera system and monitoring method
US20070188594A1 (en) Communication system, communication terminal and communication method
JP6238255B2 (en) Surveillance method and video segmentation apparatus using surveillance camera system
US10356302B2 (en) Transmission apparatus, reception apparatus, transmission and reception system, transmission apparatus control method, reception apparatus control method, transmission and reception system control method, and program
US20200037001A1 (en) Method and apparatus for generating a composite video stream from a plurality of video segments
US10306173B2 (en) Imaging apparatus, imaging method, and program
US10015395B2 (en) Communication system, communication apparatus, communication method and program
JP6319970B2 (en) Image communication apparatus and control method thereof
WO2018123078A1 (en) Monitoring camera system
JP2015119335A (en) Terminal, system, program and method to thin out frame of photographed moving image in accordance with movement change amount
JP2012137900A (en) Image output system, image output method and server device
WO2018003685A1 (en) Image compressing method, image reconstructing method, image compressing device, image reconstructing device, image compressing program product, and image reconstructing program product
JP6357188B2 (en) Surveillance camera system and surveillance camera data storage method
JP6363130B2 (en) Surveillance method, difference image creation method, image restoration method, and difference detection apparatus in surveillance camera system
KR102050416B1 (en) Network camera apparatus and method of providing image streaming thereof.
JP6468663B2 (en) Video content distribution device
JP5896438B2 (en) Surveillance camera system and monitoring method
JP6425389B2 (en) Imaging device and imaging system
CN116074584A (en) Video stream transmission method, device and storage medium
Bhattacharjee An adaptive video streaming solution over wireless networks

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17886544

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17886544

Country of ref document: EP

Kind code of ref document: A1