US20040213474A1 - Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system - Google Patents
Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system Download PDFInfo
- Publication number
- US20040213474A1 US20040213474A1 US10/852,159 US85215904A US2004213474A1 US 20040213474 A1 US20040213474 A1 US 20040213474A1 US 85215904 A US85215904 A US 85215904A US 2004213474 A1 US2004213474 A1 US 2004213474A1
- Authority
- US
- United States
- Prior art keywords
- terminals
- image data
- image
- connection device
- multipoint connection
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H—ELECTRICITY
- H03—ELECTRONIC CIRCUITRY
- H03M—CODING; DECODING; CODE CONVERSION IN GENERAL
- H03M13/00—Coding, decoding or code conversion, for error detection or error correction; Coding theory basic assumptions; Coding bounds; Error probability evaluation methods; Channel models; Simulation or testing of codes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T9/00—Image coding
- G06T9/005—Statistical coding, e.g. Huffman, run length coding
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B20/00—Signal processing not specific to the method of recording or reproducing; Circuits therefor
- G11B20/10—Digital recording or reproducing
- G11B20/18—Error detection or correction; Testing, e.g. of drop-outs
Definitions
- the present invention relates to a data communication control apparatus and its control method, an image processing apparatus and its method, and a data communication system which realize interactive video communication among a plurality of points.
- a multipoint video conference system widely used now enables video conference terminals based on the ITU-T Recommendation H. 320 and the like to perform interactive communication among a number of points via a multipoint connection device.
- the ITU-T Recommendation H. 231 defining the functional construction of the multipoint connection device and the format of multipoint communication
- the ITU-T Recommendation H. 243 determining an in-channel communication procedure and the like have been made.
- FIG. 1 shows the construction of a conventional video conference system.
- a multipoint connection device 22 interconnects video conference terminals 21 a to 21 n at three or more points, and performs audio mixing, video data delivery control or video data combining for split screen view, further, performs presidential control over the conference.
- a multipoint conference connecting a plurality of remote points can be realized by connecting terminals via a multipoint connection device as long as the terminals are based on the Recommendation H. 320 .
- FIG. 11 shows video data transfer in this system.
- numerals 41 a to 41 c denote client terminals realized by personal computers or the like having video and audio input/output functions; 42 , a multipoint connection server which delivers video and audio data from the respective client terminals 41 a to 41 c to the client terminals 41 a to 41 c without any processing.
- the multipoint connection server 42 delivers video and audio data from the respective client terminals 41 a to 41 c to the client terminals 41 a to 41 c without any processing.
- the client terminal 41 a transmits terminal information A to the multipoint connection server 42
- the multipoint connection server 42 transfers terminal information B and C from other client terminals 41 b and 41 c to the client terminal 41 a.
- the code amount of the image data delivered from the multipoint connection server 42 to the client terminals 41 a to 41 c increases in proportion to the number of conference participants (terminals).
- a data communication control apparatus comprising: compression means for compressing image data to be transmitted; communication means for transmitting the image data compressed by the compression means to the terminals; recognition means for recognizing a specific area in the image data; and compression control means for controlling a compression method in the compression means in accordance with whether or not an object area in the image data is the specific area.
- the foregoing object is attained by providing the data communication control apparatus wherein the specific area is an area showing a person's face.
- an image processing apparatus comprising: recognition means for recognizing a-specific area in an image represented by compressed image data; and compression control means for controlling a re-compression method for re-compressing the compressed image data based on recognition by the recognition means.
- FIG. 1 is a block diagram showing the construction of a multipoint video conference system according to a first embodiment of the present invention
- FIG. 2 is a block diagram showing the basic construction of a multipoint connection device
- FIG. 3 is an example of display screen of a video conference terminal
- FIG. 4 is a flowchart showing image processing by the multipoint connection device
- FIG. 5 is an example of image delivered in the first embodiment
- FIG. 6 is a block diagram showing the basic construction of the multipoint connection device according to a second embodiment of the present invention.
- FIG. 7 is a flowchart showing image processing by the multipoint connection device
- FIG. 8 is an example of a spatial filter
- FIG. 9 is a block diagram showing the construction of the multipoint connection device according to a third embodiment of the present invention.
- FIG. 10 is an example of image processing by the conventional multipoint connection device.
- FIG. 11 is a block diagram showing an Internet-type multipoint video conference system.
- FIG. 12 is a block diagram showing the another construction of a multipoint connection device.
- the basic construction of a video conference system according to a first embodiment of the present invention is similar to that of the above-described prior art as shown in FIG. 1.
- the video conference terminals 21 a to 21 n (hereinbelow, simply referred to as “terminals”) at three or more points are interconnected via the multipoint connection device 22 .
- Image data outputted from the respective terminals 21 a to 21 d are delivered to the other terminals via the multipoint connection device 22 .
- FIG. 2 is a block diagram showing the basic construction of the multipoint connection device 22 in the present embodiment.
- numerals 11 a to 11 n denote network interface units for accommodating the remote terminals 21 a to 21 n, as line interfaces for ISDN, PSTN and the like; and 12 a to 12 n, multiplexing/demultiplexing units which perform multiplexing/demultiplexing on various image, audio and control data in accordance with, e.g., the ITU-T Recommendation H.221.
- the number of the network interface units 11 a to 11 n and that of the multiplexing/demultiplexing units 12 a to 12 n correspond to the number of video conference terminals which can be directly connected to the multipoint connection device. Further, the multipoint connection device 22 can be connected to another multipoint connection device, a server or the like via the network interface units 11 a to 11 n and the multiplexing/demultiplexing units 12 a to 12 n.
- Numeral 13 denotes an audio processor which performs decompression/mixing processing on audio code data, transmitted from the respective terminals 21 a to 21 n and demultiplexed by the multiplexing/demultiplexing units 12 a to 12 n, then performs compression encoding on the data again, and delivers the data to the respective terminals 21 a to 21 d.
- Numeral 14 denotes an image decoder which decodes compressed image code data demultiplexed by the multiplexing/demultiplexing units 12 a to 12 n, thus generating uncompressed image data.
- Numeral 15 denotes a face-area recognition unit which recognizes a face area of a person from image data decoded by the image decoder 14 ; 16 , an image encoder which again compresses the image data decoded by the image decoder 14 ; and 17 , a quantization controller which performs quantization control upon re-compression by the image encoder 16 , in accordance with the result of face area recognition by the face-area recognition unit 15 . More specifically, a plurality of quantization tables are prepared, and if it is determined that a quantization object image is not a face area image, a quantization table is selected such that coefficients of the table are set for a quantization step greater than that used in decoding.
- Numeral 18 denotes a system controller which performs various communication protocol processing and the entire operation of the multipoint connection device 22 .
- FIG. 3 is an example of image displayed on the terminals 21 a to 21 n in the multipoint video conference system of the present embodiment.
- numeral 71 denotes a window to display a video image of a user of the terminal 21 d; 72 a to 72 c, windows to display video images of users of the other terminals (e.g., 21 a to 21 c ) as participants of the conference; and 73 , a common window in the conference to display, e.g., a white board.
- Multiplexed multimedia data sent via the communication interface units 11 a to 11 n accommodating the terminals 21 a to 21 n are frame-demultiplexed by the multiplexing/demultiplexing units 12 a to 12 n, and audio compressed data, video compressed data, data conference low-speed data, system control data and the like are extracted (step S 1001 ).
- audio compressed data, video compressed data, data conference low-speed data, system control data and the like are extracted (step S 1001 ).
- processing on the demultiplexed video compressed data will be described.
- the demultiplexed audio compressed data is decompressed by the audio processor 13 , and mixed with other decompressed audio data demultiplexed by the respective multiplexing/demultiplexing units 12 a to 12 n. Then, the mixed audio data is compressed, multiplexed with other data such as image data by the multiplexing/demultiplexing units 12 a to 12 n, and delivered to the respective terminals 21 a to 21 n.
- the system controller 18 determines whether or not code amount control is performed on the demultiplexed video compressed data in accordance with an instruction from a terminal as a delivery destination of the data (step S 1002 ). If the code amount control is not performed, the video compressed data is multiplexed again with other data by the multiplexing/demultiplexing units 12 a to 12 n (step S 1009 ). In this arrangement, regarding a video image from a specific terminal, specified by a user, if the user instructs not to perform the above-described code amount control, the video image can be delivered as a fine image with a clear background image.
- the video compressed data is decoded by the image decoder 14 (step S 1003 ), and temporarily stored in a RAM (not shown) or the like in the image decoder 14 .
- the face-area recognition unit 15 performs face area recognition processing on the decoded image data (step S 1004 ).
- the face area recognition by the face-area recognition unit 15 an area including a person's face can be recognized by, e.g., a method described in “Extraction of Plural Face Areas from a Natural Image by Using Genetic Algorithm” (Yokoo and Hagiwara, The Institute of Electrical Engineers of Japan Proceedings C, Vol. 117-C, No. 9, September, 1997) or the like.
- the face area recognition method is not limited to any particular method, but any method can be used.
- a higher level recognition method such as extraction of exact face area outline can be employed as long as the multipoint connection device has sufficiently high processing capability.
- the decoded image data is re-compressed by the image encoder 16 .
- the quantization coefficients are controlled based on the result of face area recognition.
- the result of face area recognition is outputted to the quantization controller 17 .
- the quantization controller 17 has quantization coefficients used in decoding by the image decoder 14 (or Q factors for quantization) and quantization coefficients multiplied by n (n>1). Note that-the former quantization coefficients-are set as default coefficients.
- step S 1005 As a non-face area, the quantization coefficients having the set default values are changed to the quantization coefficients multiplied by n (step S 1006 ). Note that regarding an area determined as a face area, the same quantization coefficients (default values) as those in the decoding are used.
- the image encoder 16 performs quantization based on the set quantization coefficients on the image data held in the image decoder 14 , thus performing compression processing (step S 1007 ).
- the multipoint connection device 22 performs the above processings independently on data transferred from the respective terminals 21 a to 21 n.
- the re-compressed image data are multiplexed with the audio compressed data by the multiplexing/demultiplexing units 12 a to 12 n (step S 1008 ), and delivered to the respective terminals 21 a to 21 n in accordance with a communication protocol.
- FIG. 5 is an example of displayed image, obtained by e.g. a camera (not shown) of the terminal 21 d, then processed by the multipoint connection device 22 as described above, and delivered to the respective terminals 21 a to 21 n, in the video conference system of the present embodiment.
- an area in a circle 51 is recognized as a face area; and an area 52 is a background area recognized as a non-face area.
- the background area 52 quantized by a greater quantization step, becomes a low-resolution image, while the face area 51 has image quality equivalent to that of an image obtained by the terminal 21 d.
- a significant portion of an image obtained by a terminal can be transmitted with high precision to the respective terminals of video conference system. Accordingly, even if the respective terminals are connected by a narrow band communication channel, the significant portion can be transmitted in high-quality image communication.
- a significant portion, face area, i.e., the facial expression of the user of the other terminal can be clearly displayed without reduction or frame thinning.
- an image from another arbitrary terminal may be received as an entirely high quality image at any time.
- the decoding processing and the encoding processing in the present embodiment may be performed in predetermined image data block units. In this case, the above-described processings at steps S 1003 to S 1007 are repeated in each block.
- an area where high image quality is maintained is not limited to a face area.
- the same advantage as that of the present embodiment can be obtained by setting an arbitrary significant object as an object of high-precision display in accordance with the content of conference, and transferring other image areas with degraded image quality.
- FIG. 6 is a block diagram showing the construction of the multipoint connection device 22 according to the second embodiment of the present invention.
- constituents corresponding to those in FIG. 2 have the same reference numerals and the explanations of these constituents will be omitted.
- numeral 86 denotes a spatial filter processor which performs low-frequency filter processing on decoded image data based on the result of recognition by the face-area recognition unit 15 . More specifically, the spatial filter processor 86 performs spatial filter processing on areas other than a face area by a low-pass filter, to attenuate high frequency components.
- the decoded image data is re-compressed by the image encoder 16 .
- spatial filter processing is performed based on the result of face area recognition.
- the spatial filter processor 86 inputs decoded image data and the result of face area recognition.
- a low-pass filter is prepared in advance, and spatial filter processing by the low-pass filter is performed (step S 1106 ) on an area determined at step S 1005 as a non-face area.
- the spatial filter processed image data is compressed by the image encoder 16 (step S 1008 ).
- FIG. 8 is an example of the low-pass filter in the spatial filter processor 86 .
- the filter processing expressed by the following equation (1) is performed by using the low-pass filter.
- P ′( x,y ) ( P ( x ⁇ 1 ,y ⁇ 1)+ P ( x,y ⁇ 1)+ P ( x+ 1 ,y ⁇ 1 )+ P ( x ⁇ 1 ,y )+ P ( x,y )+( x +1 ,y )+ P ( x,y +1)+ P ( x +1 ,y +1))/9 (1)
- P(x,y) is a pixel value of coordinates (x,y).
- the spatial filter in the second embodiment is not limited to that as shown in FIG. 8, but any filter may be employed as long as it is a low-pass filter.
- execution/non-execution of spatial filter processing is controlled in accordance with whether or not an object area is a face area (specific area), however, the present invention is not limited to this arrangement.
- spatial filter processing may be performed on a specific area and other areas.
- a low-pass filer used in the processing is selected in accordance with whether an object area is a specific area or not.
- a filter for the other areas than the face area is selected so as to remove high frequency components more than those in the face area removed by a filter for the face area.
- a specific area with great significance such as a face area is re-compressed without degrading image quality, and in areas other than the specific area with less significance such as background, high frequency components are attenuated. This reduces the total code amount, and accordingly, obtains the same advantage as that of the first embodiment.
- the camera of a terminal may be operated such that the user's face comes to the center of the obtained image.
- an undesirable portion such as background other than a person's face might be arbitrarily image-sensed.
- detecting a face area is rather advantageous than previously setting low compression rate area and high compression rate area within the obtained image.
- the face area detection unnecessitates operation of the orientation of the camera, and further ensures privacy protection for conference participants.
- the principal processings of the multipoint connection device 22 having the functions as described in the first embodiment are realized by software.
- FIG. 9 is a block diagram showing the construction of the multipoint connection device 22 according to the third embodiment of the present invention.
- Numeral 61 denotes a communication interface unit for communication with an ISDN, a PSTN and the like; and 62 , a media processor which performs principal signal processings.
- the media processor 62 performs various digital signal processings by a specialized processor for multimedia signal processing, a DSP (Digital Signal Processor) and the like.
- the video conference system can be realized with a simple and low-cost construction without preparing dedicated hardware, by processing various data by multitask processing by utilizing the media processor 62 .
- Numeral 63 denotes a RAM (Random Access Memory) which is used as a work memory necessary for the operation of the media processor 62 and as a buffer memory for various image and audio data to be processed.
- the RAM 63 is realized by, e.g., a high-speed large-capacity memory such as an SDRAM (Synchronous DRAM).
- Numeral 64 denotes a CPU which performs system control for the multipoint connection device 22 , communication protocol processing and the like; 65 , a RAM used as a work buffer necessary for the operation of the CPU 64 ; and 66 , a ROM (Read Only Memory) in which a control program necessary for the operation of the CPU 64 and a control program necessary for the operation of the media processor 62 are stored. The program necessary for the operation of the media processor 62 is transferred from the ROM 66 to the RAM 63 upon system initialization.
- a control program to realize the processing is stored in the ROM 66 in advance.
- the program is developed on the RAM 63 or the RAM 65 , and executed by the media processor 62 or the CPU 64 .
- processing related to data compression/decompression is performed by the media processor 62 , and other processings are controlled by the CPU 64 .
- the multipoint connection device 22 as described in the first embodiment is realized by software.
- the multipoint connection device 22 as described in the second embodiment may be realized by software.
- the compression method employed in the above-described first to third embodiments may be any method.
- an H. 261 coding method or the like based on DCT (Discrete Cosine Transformation) used in the Recommendation H.320 may be used.
- the control on the compression method (quantization coefficients) based on the result of specific-area recognition is performed in macroblock units.
- the compression control method of the present invention is not limited to the quantization control or spatial filter control as described in the first to third embodiments, but any control method may be performed as long as it controls the degree of image quality degradation due to re-compression in accordance with whether or not an object area is a specific area. In other words, any control method may be used as long as it controls the compression rate in re-compression.
- more flexible image quality control including further improvement in image quality of delivered image data can be realized by combining the quantization control described in the first embodiment and the spatial filter control described in the second embodiment.
- the video conference terminals are directly connected to a public line such as PSTN/ISDN line, however, the present invention is not limited to this example.
- the video conference terminals may be connected to the Internet, an Intranet or the like.
- communication is performed with a plurality of terminals by the Internet Protocol via a network access controller 121 having a LAN interface such as the Ethernet.
- the means for communication with the respective terminals is logically constructed for each terminal. Control on packetized data and data multiplexing/demultiplexing processing are performed by the network access controller 121 and a system controller 122 .
- the video conference terminals connected to the multipoint connection device of the present invention may be dedicated terminals or general-purpose terminals such as personal computers where a predetermined application (software) program is operated.
- the multipoint connection device of the present invention may have an automatic HTML (Hyper Text Markup Language) data generating function and an HTTP (Hyper Text Transfer Protocol) server function in addition to the function related to multipoint terminal connection.
- general-purpose communication terminals which display and visualize image and audio data stored in the HTTP server of the multipoint connection device, can be connected to the multipoint connection device.
- general-purpose communication terminals personal computers having a so-called WWW-browser and network computers can be employed. Further, by virtue of recent development of semiconductor technologies, these computers may be employed as quick and convenient terminals even in a mobile communication environment.
- These general-purpose communication terminals are IP (Internet Protocol)-connected to the multipoint connection terminal via a public network such as an ISDN or a PSTN or a LAN (Local Area Network). This enables interconnection among the existing video conference terminals and the general-purpose terminals such as personal computers and network computers.
- IP Internet Protocol
- the present invention can be applied to a system constituted by a plurality of devices (e.g., a host computer, an interface, a reader and a printer) or to an apparatus comprising a single device (e.g., a copy machine or a facsimile apparatus).
- the object of the present invention can be also achieved by providing a storage medium storing program code for performing the aforesaid processes to a system or an apparatus, reading the program code with a computer (e.g., CPU, MPU) of the system or apparatus from the storage medium, then executing the program.
- a computer e.g., CPU, MPU
- the program code read from the storage medium realize the functions according to the embodiments, and the storage medium storing the program code constitutes the invention.
- the storage medium such as a floppy disk, a hard disk, an optical disk, a magneto-optical disk, CD-ROM, CD-R, a magnetic tape, a non-volatile type memory card, and ROM can be used for providing the program code.
- the present invention includes a case where an OS (operating system) or the like working on the computer performs a part or entire processes in accordance with designations of the program code and realizes functions according to the above embodiments.
- a product obtained by the image processing method of the present invention e.g., printed matter, is included in the present invention.
- the present invention also includes a case where, after the program code read from the storage medium are written in a function expansion card which is inserted into the computer or in a memory provided in a function expansion unit which is connected to the computer, CPU or the like contained in the function expansion card or unit performs a part or entire process in accordance with designations of the program code and realizes functions of the above embodiments.
- the storage medium stores program code corresponding to the flowcharts shown in FIGS. 4 and 7 described in the embodiments.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Probability & Statistics with Applications (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Compression Of Band Width Or Redundancy In Fax (AREA)
- Data Exchanges In Wide-Area Networks (AREA)
- Communication Control (AREA)
- Telephonic Communication Services (AREA)
Abstract
In video conference system terminals, the conventional difficulty in high-quality display of facial expressions of a plurality of participants is solved. In a multipoint connection device connecting a plurality of video conference terminals, image data transmitted from a terminal is decoded, and a face area in the image data is recognized. Then, quantization coefficients for the face area are set to be greater coefficients than those for other areas than the face area, then the image data is compressed, and delivered to the respective terminals. By this arrangement, in image data, a face area with great significance can be re-compressed without degradation of image quality, while an area with less significance such as background can be compressed with high efficiency. Thus, the total code amount can be reduced. Accordingly, even if the system uses a narrow band communication channel, the significant face area can be clearly displayed at the respective terminals.
Description
- The present invention relates to a data communication control apparatus and its control method, an image processing apparatus and its method, and a data communication system which realize interactive video communication among a plurality of points.
- A multipoint video conference system widely used now enables video conference terminals based on the ITU-T Recommendation H. 320 and the like to perform interactive communication among a number of points via a multipoint connection device.
- As standards related to the multipoint connection device to control such multipoint connection among video conference terminals, the ITU-T Recommendation H. 231 defining the functional construction of the multipoint connection device and the format of multipoint communication, the ITU-T Recommendation H. 243 determining an in-channel communication procedure and the like have been made.
- FIG. 1 shows the construction of a conventional video conference system. In FIG. 1, a
multipoint connection device 22 interconnectsvideo conference terminals 21 a to 21 n at three or more points, and performs audio mixing, video data delivery control or video data combining for split screen view, further, performs presidential control over the conference. In this manner, a multipoint conference connecting a plurality of remote points can be realized by connecting terminals via a multipoint connection device as long as the terminals are based on the Recommendation H. 320. - To realize multipoint interactive video communication in a video conference system based on the Recommendation H. 320, to deliver image data from the
multipoint connection device 22 to thevideo conference terminals 21 a to 21 n as shown in FIG. 1, the following methods (1), (2) and the like have been put into practical use: - (1) A method to deliver a video image from a specific video conference terminal designated by a presidential terminal to the respective video conference terminals.
- (2) As shown in FIG. 10, a method to reduce video images from respective
video conference terminals 21 a to 21 n in themultipoint connection device 22, so as to simultaneously deliver images from respectivevideo conference terminals 21 a to 21 n within a limited band. - Note that in FIG. 10,
numerals 31 a to 31 d denote video data transmitted from the respectivevideo conference terminals 21 a to 21 d; 32, reduced and synthesized video data delivered from themultipoint connection device 22 to the respectivevideo conference terminals 21 a to 21 d. - Further, to realize a video conference system in an Internet environment, software programs such as CU-SeeMe by White Pine Software, Inc. have been provided. In the multipoint video conference system, image data transmitted from the respective client terminals are delivered via a multipoint connection server to the respective client terminals without any processing. FIG. 11 shows video data transfer in this system. In FIG. 11,
numerals 41 a to 41 c denote client terminals realized by personal computers or the like having video and audio input/output functions; 42, a multipoint connection server which delivers video and audio data from therespective client terminals 41 a to 41 c to theclient terminals 41 a to 41 c without any processing. - However, the conventional video conference system using the multipoint connection based on the Recommendation H. 320 or the like, in which communication is performed within a limited band to the respective terminals, has problems as follows.
- (1) Only the video image of a particular reference participant is delivered; or
- (2) As reduced images of plurality of participants are transmitted, the images of the respective speakers are unclear.
- In the video conference terminal, it is difficult to simultaneously and clearly display a plurality of conference participants.
- Further, in the video conference system utilizing the Internet as shown in FIG. 11, the
multipoint connection server 42 delivers video and audio data from therespective client terminals 41 a to 41 c to theclient terminals 41 a to 41 c without any processing. For example, between themultipoint connection server 42 and theclient terminal 41 a, theclient terminal 41 a transmits terminal information A to themultipoint connection server 42, and themultipoint connection server 42 transfers terminal information B and C fromother client terminals client terminal 41 a. Accordingly, the code amount of the image data delivered from themultipoint connection server 42 to theclient terminals 41 a to 41 c increases in proportion to the number of conference participants (terminals). If a sufficient band is not ensured as a communication channel, the frame rate of images received by theclient terminals 41 a to 41 c is extremely reduced. Accordingly, to display a received image, so-called frame thinning, i.e., omitting a frame at predetermined intervals, must be performed. - Accordingly, it is an object of the present invention to provide a data communication control apparatus and its control method and a data communication system which enable high-quality image communication. It is another object of the present invention to provide a data communication control apparatus and its control method and a data communication system which enables image communication without degration of the quality of the important portion of an image even if using a narrow band communication channel.
- According to the present invention, the foregoing objects are attained by providing a data communication control apparatus comprising: compression means for compressing image data to be transmitted; communication means for transmitting the image data compressed by the compression means to the terminals; recognition means for recognizing a specific area in the image data; and compression control means for controlling a compression method in the compression means in accordance with whether or not an object area in the image data is the specific area.
- It is another object of the present invention to provide a data communication control apparatus and its control method and a data communication system which enable clearly displayed facial expressions of participants in a multipoint video conference system.
- According to the present invention, the foregoing object is attained by providing the data communication control apparatus wherein the specific area is an area showing a person's face.
- It is further object of the present invention to provide an image processing apparatus and its method which mitigate degradation of image quality upon reuse of compressed image data.
- According to the present invention, the foregoing object is attained by providing an image processing apparatus comprising: recognition means for recognizing a-specific area in an image represented by compressed image data; and compression control means for controlling a re-compression method for re-compressing the compressed image data based on recognition by the recognition means.
- Other features and advantages of the present invention will be apparent from the following description taken in conjunction with the accompanying drawings, in which like reference characters designate the same name or similar parts throughout the figures thereof.
- The accompanying drawings, which are incorporated in and constitute a part of the specification, illustrate embodiments of the invention and, together with the description, serve to explain the principles of the invention.
- FIG. 1 is a block diagram showing the construction of a multipoint video conference system according to a first embodiment of the present invention;
- FIG. 2 is a block diagram showing the basic construction of a multipoint connection device;
- FIG. 3 is an example of display screen of a video conference terminal;
- FIG. 4 is a flowchart showing image processing by the multipoint connection device;
- FIG. 5 is an example of image delivered in the first embodiment;
- FIG. 6 is a block diagram showing the basic construction of the multipoint connection device according to a second embodiment of the present invention;
- FIG. 7 is a flowchart showing image processing by the multipoint connection device;
- FIG. 8 is an example of a spatial filter;
- FIG. 9 is a block diagram showing the construction of the multipoint connection device according to a third embodiment of the present invention;
- FIG. 10 is an example of image processing by the conventional multipoint connection device; and
- FIG. 11 is a block diagram showing an Internet-type multipoint video conference system.
- FIG. 12 is a block diagram showing the another construction of a multipoint connection device.
- Preferred embodiments of the present invention will now be described in detail in accordance with the accompanying drawings.
- The basic construction of a video conference system according to a first embodiment of the present invention is similar to that of the above-described prior art as shown in FIG. 1. The
video conference terminals 21 a to 21 n (hereinbelow, simply referred to as “terminals”) at three or more points are interconnected via themultipoint connection device 22. Image data outputted from therespective terminals 21 a to 21 d are delivered to the other terminals via themultipoint connection device 22. - FIG. 2 is a block diagram showing the basic construction of the
multipoint connection device 22 in the present embodiment. In FIG. 2,numerals 11 a to 11 n denote network interface units for accommodating theremote terminals 21 a to 21 n, as line interfaces for ISDN, PSTN and the like; and 12 a to 12 n, multiplexing/demultiplexing units which perform multiplexing/demultiplexing on various image, audio and control data in accordance with, e.g., the ITU-T Recommendation H.221. Note that in themultipoint connection device 22, the number of thenetwork interface units 11 a to 11 n and that of the multiplexing/demultiplexing units 12 a to 12 n correspond to the number of video conference terminals which can be directly connected to the multipoint connection device. Further, themultipoint connection device 22 can be connected to another multipoint connection device, a server or the like via thenetwork interface units 11 a to 11 n and the multiplexing/demultiplexing units 12 a to 12 n. -
Numeral 13 denotes an audio processor which performs decompression/mixing processing on audio code data, transmitted from therespective terminals 21 a to 21 n and demultiplexed by the multiplexing/demultiplexing units 12 a to 12 n, then performs compression encoding on the data again, and delivers the data to therespective terminals 21 a to 21 d. Numeral 14 denotes an image decoder which decodes compressed image code data demultiplexed by the multiplexing/demultiplexing units 12 a to 12 n, thus generating uncompressed image data. - Numeral15 denotes a face-area recognition unit which recognizes a face area of a person from image data decoded by the
image decoder 14; 16, an image encoder which again compresses the image data decoded by theimage decoder 14; and 17, a quantization controller which performs quantization control upon re-compression by theimage encoder 16, in accordance with the result of face area recognition by the face-area recognition unit 15. More specifically, a plurality of quantization tables are prepared, and if it is determined that a quantization object image is not a face area image, a quantization table is selected such that coefficients of the table are set for a quantization step greater than that used in decoding. - Numeral18 denotes a system controller which performs various communication protocol processing and the entire operation of the
multipoint connection device 22. - FIG. 3 is an example of image displayed on the
terminals 21 a to 21 n in the multipoint video conference system of the present embodiment. Assuming that the image in FIG. 5 is a displayed image on the terminal 21 d,numeral 71 denotes a window to display a video image of a user of the terminal 21 d; 72 a to 72 c, windows to display video images of users of the other terminals (e.g., 21 a to 21 c) as participants of the conference; and 73, a common window in the conference to display, e.g., a white board. - Next, image processing by the
multipoint connection device 22 will be described in detail with reference to the flowchart of FIG. 4. - Multiplexed multimedia data sent via the
communication interface units 11 a to 11 n accommodating theterminals 21 a to 21 n are frame-demultiplexed by the multiplexing/demultiplexing units 12 a to 12 n, and audio compressed data, video compressed data, data conference low-speed data, system control data and the like are extracted (step S1001). Hereinbelow, processing on the demultiplexed video compressed data will be described. - Note that the demultiplexed audio compressed data is decompressed by the
audio processor 13, and mixed with other decompressed audio data demultiplexed by the respective multiplexing/demultiplexing units 12 a to 12 n. Then, the mixed audio data is compressed, multiplexed with other data such as image data by the multiplexing/demultiplexing units 12 a to 12 n, and delivered to therespective terminals 21 a to 21 n. - The
system controller 18 determines whether or not code amount control is performed on the demultiplexed video compressed data in accordance with an instruction from a terminal as a delivery destination of the data (step S1002). If the code amount control is not performed, the video compressed data is multiplexed again with other data by the multiplexing/demultiplexing units 12 a to 12 n (step S1009). In this arrangement, regarding a video image from a specific terminal, specified by a user, if the user instructs not to perform the above-described code amount control, the video image can be delivered as a fine image with a clear background image. - On the other hand, if the code amount control is performed, the video compressed data is decoded by the image decoder14 (step S1003), and temporarily stored in a RAM (not shown) or the like in the
image decoder 14. Then the face-area recognition unit 15 performs face area recognition processing on the decoded image data (step S1004). Note that as the face area recognition by the face-area recognition unit 15, an area including a person's face can be recognized by, e.g., a method described in “Extraction of Plural Face Areas from a Natural Image by Using Genetic Algorithm” (Yokoo and Hagiwara, The Institute of Electrical Engineers of Japan Proceedings C, Vol. 117-C, No. 9, September, 1997) or the like. - Note that in the present embodiment, the face area recognition method is not limited to any particular method, but any method can be used. A higher level recognition method such as extraction of exact face area outline can be employed as long as the multipoint connection device has sufficiently high processing capability. Further, in some face area recognition method, it is not necessary for the
image decoder 14 to completely decode the image data. - After the face area recognition, the decoded image data is re-compressed by the
image encoder 16. In the present embodiment, upon compression processing, the quantization coefficients are controlled based on the result of face area recognition. - The result of face area recognition is outputted to the
quantization controller 17. Thequantization controller 17 has quantization coefficients used in decoding by the image decoder 14 (or Q factors for quantization) and quantization coefficients multiplied by n (n>1). Note that-the former quantization coefficients-are set as default coefficients. - Regarding an area determined at step S1005 as a non-face area, the quantization coefficients having the set default values are changed to the quantization coefficients multiplied by n (step S1006). Note that regarding an area determined as a face area, the same quantization coefficients (default values) as those in the decoding are used.
- Then, the
image encoder 16 performs quantization based on the set quantization coefficients on the image data held in theimage decoder 14, thus performing compression processing (step S1007). - In this arrangement, a face area with great significance is re-compressed without degrading image quality, while an area other than the face area such as background with less significance is compressed by a greater quantization step. Thus the total code amount can be reduced.
- The
multipoint connection device 22 performs the above processings independently on data transferred from therespective terminals 21 a to 21 n. Note that the re-compressed image data are multiplexed with the audio compressed data by the multiplexing/demultiplexing units 12 a to 12 n (step S1008), and delivered to therespective terminals 21 a to 21 n in accordance with a communication protocol. - FIG. 5 is an example of displayed image, obtained by e.g. a camera (not shown) of the terminal21 d, then processed by the
multipoint connection device 22 as described above, and delivered to therespective terminals 21 a to 21 n, in the video conference system of the present embodiment. In FIG. 5, an area in acircle 51 is recognized as a face area; and anarea 52 is a background area recognized as a non-face area. Thebackground area 52, quantized by a greater quantization step, becomes a low-resolution image, while theface area 51 has image quality equivalent to that of an image obtained by the terminal 21 d. - As described above, according to the present embodiment, in a video conference system, only a significant portion of an image obtained by a terminal can be transmitted with high precision to the respective terminals of video conference system. Accordingly, even if the respective terminals are connected by a narrow band communication channel, the significant portion can be transmitted in high-quality image communication. At the respective terminals, in an image transferred from another terminal, a significant portion, face area, i.e., the facial expression of the user of the other terminal can be clearly displayed without reduction or frame thinning.
- Further, in response to an instruction from a terminal, an image from another arbitrary terminal may be received as an entirely high quality image at any time.
- Note that the decoding processing and the encoding processing in the present embodiment may be performed in predetermined image data block units. In this case, the above-described processings at steps S1003 to S1007 are repeated in each block.
- Further, an area where high image quality is maintained is not limited to a face area. The same advantage as that of the present embodiment can be obtained by setting an arbitrary significant object as an object of high-precision display in accordance with the content of conference, and transferring other image areas with degraded image quality.
- Next, a second embodiment of the present invention will be described.
- As the basic construction of the video conference system of the second embodiment is similar to that of the first embodiment as shown in FIG. 1, the explanation of the construction will be omitted.
- FIG. 6 is a block diagram showing the construction of the
multipoint connection device 22 according to the second embodiment of the present invention. In FIG. 6, constituents corresponding to those in FIG. 2 have the same reference numerals and the explanations of these constituents will be omitted. - In FIG. 6, numeral86 denotes a spatial filter processor which performs low-frequency filter processing on decoded image data based on the result of recognition by the face-
area recognition unit 15. More specifically, thespatial filter processor 86 performs spatial filter processing on areas other than a face area by a low-pass filter, to attenuate high frequency components. - Next, the image processing by the
multipoint connection device 22 according to the second embodiment will be described in detail with reference to the flowchart of FIG. 7. Note that steps corresponding to those in FIG. 4 have the same step numbers, and the explanations of those steps will be omitted. - As in the case of the first embodiment, after the face area recognition at step S1004, the decoded image data is re-compressed by the
image encoder 16. In the present embodiment, upon this compression processing, spatial filter processing is performed based on the result of face area recognition. - The
spatial filter processor 86 inputs decoded image data and the result of face area recognition. In thespatial filter processor 86, a low-pass filter is prepared in advance, and spatial filter processing by the low-pass filter is performed (step S1106) on an area determined at step S1005 as a non-face area. Then, the spatial filter processed image data is compressed by the image encoder 16 (step S1008). - FIG. 8 is an example of the low-pass filter in the
spatial filter processor 86. The filter processing expressed by the following equation (1) is performed by using the low-pass filter. - P′(x,y)=(P(x−1,y−1)+P(x,y−1)+P(x+1,y−1 )+P(x−1,y)+P(x,y)+(x+1,y)+P(x−1,y+1)+P(x,y+1)+P(x+1,y+1))/9 (1)
- Note that P(x,y) is a pixel value of coordinates (x,y).
- Note that the spatial filter in the second embodiment is not limited to that as shown in FIG. 8, but any filter may be employed as long as it is a low-pass filter.
- Further, in the second embodiment, execution/non-execution of spatial filter processing is controlled in accordance with whether or not an object area is a face area (specific area), however, the present invention is not limited to this arrangement. For example, spatial filter processing may be performed on a specific area and other areas. In this case, a low-pass filer used in the processing is selected in accordance with whether an object area is a specific area or not. A filter for the other areas than the face area is selected so as to remove high frequency components more than those in the face area removed by a filter for the face area.
- As described above, according to the second embodiment, in image data, a specific area with great significance such as a face area is re-compressed without degrading image quality, and in areas other than the specific area with less significance such as background, high frequency components are attenuated. This reduces the total code amount, and accordingly, obtains the same advantage as that of the first embodiment.
- Further, in the second embodiment, as the resolution of image is controlled by the spatial filter processing, degradation of image quality is visually more inconspicuous in comparison with the first embodiment where the code amount is reduced by control on quantization coefficients.
- In the above-described first and second embodiments, in each of the images obtained by the respective video conference terminals, especially a face area is detected. In a case where the terminals respectively have a fixed camera which obtains an image with a fixed image sensing area, if the compression rate for a part around the center of the obtained image is set to a lower rate than the compression rate for other parts in advance, the object of the present invention can be attained without detecting a face area from the image.
- However, in a case where a plurality of persons are image-sensed by such fixed camera, in an image from a terminal, the user of the terminal is not always positioned around the center of the image. Accordingly, in this case, if the image transfer is performed such that only a part around the center of the obtained image has high image quality, the transferred image is unnatural. For example, if only the half of the face of the user is included in a preset low compression rate area but the other half is included in a high compression rate area, it is impossible to transfer the image such that the entire face area has high image quality.
- Accordingly, even if the terminals respectively have a fixed camera, to ensure high-quality face area transfer, a face area is extracted as a significant portion from an obtained image, as in the case of the above-described first and second embodiments.
- Further, in a case where the terminals respectively have a camera in which its image sensing direction can be operated from another terminal, the camera of a terminal may be operated such that the user's face comes to the center of the obtained image. However, if the orientation of the camera can be arbitrarily operated from the opposite-end terminal side, an undesirable portion such as background other than a person's face might be arbitrarily image-sensed.
- Accordingly, in case of image sensing by such remote-controllable camera, detecting a face area is rather advantageous than previously setting low compression rate area and high compression rate area within the obtained image. The face area detection unnecessitates operation of the orientation of the camera, and further ensures privacy protection for conference participants.
- Next, a third embodiment of the present invention will be described. The basic construction of the video conference system of the third embodiment is similar to that of the first embodiment as shown in FIG. 1, therefore, the explanation of the construction will be omitted.
- In the third embodiment, the principal processings of the
multipoint connection device 22 having the functions as described in the first embodiment are realized by software. - FIG. 9 is a block diagram showing the construction of the
multipoint connection device 22 according to the third embodiment of the present invention.Numeral 61 denotes a communication interface unit for communication with an ISDN, a PSTN and the like; and 62, a media processor which performs principal signal processings. Themedia processor 62 performs various digital signal processings by a specialized processor for multimedia signal processing, a DSP (Digital Signal Processor) and the like. In the third embodiment, the video conference system can be realized with a simple and low-cost construction without preparing dedicated hardware, by processing various data by multitask processing by utilizing themedia processor 62. -
Numeral 63 denotes a RAM (Random Access Memory) which is used as a work memory necessary for the operation of themedia processor 62 and as a buffer memory for various image and audio data to be processed. TheRAM 63 is realized by, e.g., a high-speed large-capacity memory such as an SDRAM (Synchronous DRAM).Numeral 64 denotes a CPU which performs system control for themultipoint connection device 22, communication protocol processing and the like; 65, a RAM used as a work buffer necessary for the operation of theCPU 64; and 66, a ROM (Read Only Memory) in which a control program necessary for the operation of theCPU 64 and a control program necessary for the operation of themedia processor 62 are stored. The program necessary for the operation of themedia processor 62 is transferred from theROM 66 to theRAM 63 upon system initialization. - In the third embodiment, the above construction realizes processing as shown in the flowchart of FIG. 4.
- That is, a control program to realize the processing is stored in the
ROM 66 in advance. The program is developed on theRAM 63 or theRAM 65, and executed by themedia processor 62 or theCPU 64. For example, processing related to data compression/decompression is performed by themedia processor 62, and other processings are controlled by theCPU 64. - According to the third embodiment, the same advantages as those in the first embodiment can be obtained by realizing the processings described in the first embodiment by software.
- Note that in the third embodiment, the
multipoint connection device 22 as described in the first embodiment is realized by software. Similarly, themultipoint connection device 22 as described in the second embodiment may be realized by software. - Note that the compression method employed in the above-described first to third embodiments may be any method. For example, an H. 261 coding method or the like based on DCT (Discrete Cosine Transformation) used in the Recommendation H.320 may be used. In such case, the control on the compression method (quantization coefficients) based on the result of specific-area recognition is performed in macroblock units.
- Further, the compression control method of the present invention is not limited to the quantization control or spatial filter control as described in the first to third embodiments, but any control method may be performed as long as it controls the degree of image quality degradation due to re-compression in accordance with whether or not an object area is a specific area. In other words, any control method may be used as long as it controls the compression rate in re-compression.
- In addition, more flexible image quality control including further improvement in image quality of delivered image data can be realized by combining the quantization control described in the first embodiment and the spatial filter control described in the second embodiment.
- Further, in the first and second embodiments, the video conference terminals are directly connected to a public line such as PSTN/ISDN line, however, the present invention is not limited to this example. The video conference terminals may be connected to the Internet, an Intranet or the like. In this case, as shown in FIG. 12, communication is performed with a plurality of terminals by the Internet Protocol via a
network access controller 121 having a LAN interface such as the Ethernet. The means for communication with the respective terminals is logically constructed for each terminal. Control on packetized data and data multiplexing/demultiplexing processing are performed by thenetwork access controller 121 and asystem controller 122. - Further, the video conference terminals connected to the multipoint connection device of the present invention may be dedicated terminals or general-purpose terminals such as personal computers where a predetermined application (software) program is operated.
- Further, the multipoint connection device of the present invention may have an automatic HTML (Hyper Text Markup Language) data generating function and an HTTP (Hyper Text Transfer Protocol) server function in addition to the function related to multipoint terminal connection. In this case, general-purpose communication terminals, which display and visualize image and audio data stored in the HTTP server of the multipoint connection device, can be connected to the multipoint connection device. As the general-purpose communication terminals, personal computers having a so-called WWW-browser and network computers can be employed. Further, by virtue of recent development of semiconductor technologies, these computers may be employed as quick and convenient terminals even in a mobile communication environment. These general-purpose communication terminals are IP (Internet Protocol)-connected to the multipoint connection terminal via a public network such as an ISDN or a PSTN or a LAN (Local Area Network). This enables interconnection among the existing video conference terminals and the general-purpose terminals such as personal computers and network computers.
- The present invention can be applied to a system constituted by a plurality of devices (e.g., a host computer, an interface, a reader and a printer) or to an apparatus comprising a single device (e.g., a copy machine or a facsimile apparatus).
- Further, the object of the present invention can be also achieved by providing a storage medium storing program code for performing the aforesaid processes to a system or an apparatus, reading the program code with a computer (e.g., CPU, MPU) of the system or apparatus from the storage medium, then executing the program.
- In this case, the program code read from the storage medium realize the functions according to the embodiments, and the storage medium storing the program code constitutes the invention.
- Further, the storage medium, such as a floppy disk, a hard disk, an optical disk, a magneto-optical disk, CD-ROM, CD-R, a magnetic tape, a non-volatile type memory card, and ROM can be used for providing the program code.
- Furthermore, besides aforesaid functions according to the above embodiments are realized by executing the program code which are read by a computer, the present invention includes a case where an OS (operating system) or the like working on the computer performs a part or entire processes in accordance with designations of the program code and realizes functions according to the above embodiments.
- Further, a product obtained by the image processing method of the present invention, e.g., printed matter, is included in the present invention.
- Furthermore, the present invention also includes a case where, after the program code read from the storage medium are written in a function expansion card which is inserted into the computer or in a memory provided in a function expansion unit which is connected to the computer, CPU or the like contained in the function expansion card or unit performs a part or entire process in accordance with designations of the program code and realizes functions of the above embodiments. In a case where the present invention is applied to the aforesaid storage medium, the storage medium stores program code corresponding to the flowcharts shown in FIGS. 4 and 7 described in the embodiments.
- As described above, according to the present invention, in the video conference system in a narrow band communication channel, facial expressions of the participants of a conference can be clearly displayed.
- Further, upon re-compression on decompressed image data, degradation of image quality can be reduced.
- As many apparently widely different embodiments of the present invention can be made without departing from the spirit and scope thereof, it is to be understood that the invention is not limited to the specific embodiments thereof except as defined in the appended claims.
Claims (8)
1-48. (canceled)
49. A data communication control method for controlling data communication among a plurality of terminals connected to a multipoint connection device, the method comprising the steps of:
transferring image data from one of the plurality of terminals;
recognizing a specific area in the transferred image data at the multipoint connection device;
controlling the compression method for compressing the received image data at the multipoint connection device to compress the specific area and not reduce a burden of the plurality of terminals; and
transmitting the compressed image data from the multipoint connection device to at least one of the plurality of terminals.
50. The method according to claim 49 , wherein the transferred image data is determined whether the specific area is compressed or not, according to a designation of at least one of the plurality of terminals.
51. The method according to claim 49 , wherein the specific area is an area in which a person's face is displayed.
52. A multipoint connection device for controlling data communication among a plurality of terminals, comprising:
an interface unit which receives image data from one of the plurality of terminals;
a recognition unit which recognizes a specific area in the received data; and
a controller which controls the compression method for compressing the received image data to compress the specific area and not reduce a burden of the plurality of terminals,
wherein the interface unit transmits the compressed image data to at least one of the plurality of terminals.
53. The device according to claim 52 , wherein the received image data is determined whether the specific area is compressed or not, according to a designation of at least one of the plurality of terminals.
54. The device according to claim 52 , wherein the specific area is an area in which a person's face is displayed.
55. A computer readable medium containing control program code in the multipoint connection device for controlling data communication among the plurality of terminals, the control program code being implemented by a method comprising the steps of:
receiving image data from one of the plurality of terminals;
recognizing a specific area in the received image data;
controlling the compression method for compressing the received image data to compress the specific area and not reduce a burden of the plurality of terminals; and
transmitting the compressed image data from the multipoint connection device to at least one of the plurality of terminals.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/852,159 US20040213474A1 (en) | 1998-12-25 | 2004-05-25 | Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP10-371478 | 1998-12-25 | ||
JP37147898A JP4208315B2 (en) | 1998-12-25 | 1998-12-25 | DATA COMMUNICATION CONTROL DEVICE AND ITS CONTROL METHOD, DATA COMMUNICATION SYSTEM, RECORDING MEDIUM |
US09/469,798 US6744927B1 (en) | 1998-12-25 | 1999-12-22 | Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system |
US10/852,159 US20040213474A1 (en) | 1998-12-25 | 2004-05-25 | Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/469,798 Division US6744927B1 (en) | 1998-12-25 | 1999-12-22 | Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system |
Publications (1)
Publication Number | Publication Date |
---|---|
US20040213474A1 true US20040213474A1 (en) | 2004-10-28 |
Family
ID=18498779
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/469,798 Expired - Fee Related US6744927B1 (en) | 1998-12-25 | 1999-12-22 | Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system |
US10/852,159 Abandoned US20040213474A1 (en) | 1998-12-25 | 2004-05-25 | Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/469,798 Expired - Fee Related US6744927B1 (en) | 1998-12-25 | 1999-12-22 | Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system |
Country Status (2)
Country | Link |
---|---|
US (2) | US6744927B1 (en) |
JP (1) | JP4208315B2 (en) |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050213738A1 (en) * | 2001-12-31 | 2005-09-29 | Polycom, Inc. | Conference endpoint requesting and receiving billing information from a conference bridge |
US20050213728A1 (en) * | 2001-12-31 | 2005-09-29 | Polycom, Inc. | Conference endpoint instructing a remote device to establish a new connection |
US20050213736A1 (en) * | 2001-12-31 | 2005-09-29 | Polycom, Inc. | Speakerphone establishing and using a second connection of graphics information |
US20060009254A1 (en) * | 2004-07-09 | 2006-01-12 | Masayuki Kanazawa | Radio communication system |
US20070047626A1 (en) * | 2005-06-08 | 2007-03-01 | Polycom, Inc | Mixed voice and spread spectrum data signaling with multiplexing multiple users with cdma |
US7787605B2 (en) * | 2001-12-31 | 2010-08-31 | Polycom, Inc. | Conference bridge which decodes and responds to control information embedded in audio information |
US7864938B2 (en) | 2000-12-26 | 2011-01-04 | Polycom, Inc. | Speakerphone transmitting URL information to a remote device |
US7978838B2 (en) | 2001-12-31 | 2011-07-12 | Polycom, Inc. | Conference endpoint instructing conference bridge to mute participants |
US8004556B2 (en) | 2004-04-16 | 2011-08-23 | Polycom, Inc. | Conference link between a speakerphone and a video conference unit |
US8102984B2 (en) | 2001-12-31 | 2012-01-24 | Polycom Inc. | Speakerphone and conference bridge which receive and provide participant monitoring information |
US8126029B2 (en) | 2005-06-08 | 2012-02-28 | Polycom, Inc. | Voice interference correction for mixed voice and spread spectrum data signaling |
US8144854B2 (en) | 2001-12-31 | 2012-03-27 | Polycom Inc. | Conference bridge which detects control information embedded in audio information to prioritize operations |
US8199791B2 (en) | 2005-06-08 | 2012-06-12 | Polycom, Inc. | Mixed voice and spread spectrum data signaling with enhanced concealment of data |
US8705719B2 (en) | 2001-12-31 | 2014-04-22 | Polycom, Inc. | Speakerphone and conference bridge which receive and provide participant monitoring information |
US8805928B2 (en) | 2001-05-10 | 2014-08-12 | Polycom, Inc. | Control unit for multipoint multimedia/audio system |
US8885523B2 (en) | 2001-12-31 | 2014-11-11 | Polycom, Inc. | Speakerphone transmitting control information embedded in audio information through a conference bridge |
US8934382B2 (en) | 2001-05-10 | 2015-01-13 | Polycom, Inc. | Conference endpoint controlling functions of a remote device |
US8947487B2 (en) | 2001-12-31 | 2015-02-03 | Polycom, Inc. | Method and apparatus for combining speakerphone and video conference unit operations |
US8948059B2 (en) | 2000-12-26 | 2015-02-03 | Polycom, Inc. | Conference endpoint controlling audio volume of a remote device |
US8964604B2 (en) | 2000-12-26 | 2015-02-24 | Polycom, Inc. | Conference endpoint instructing conference bridge to dial phone number |
US8977683B2 (en) | 2000-12-26 | 2015-03-10 | Polycom, Inc. | Speakerphone transmitting password information to a remote device |
US8976712B2 (en) | 2001-05-10 | 2015-03-10 | Polycom, Inc. | Speakerphone and conference bridge which request and perform polling operations |
US9001702B2 (en) | 2000-12-26 | 2015-04-07 | Polycom, Inc. | Speakerphone using a secure audio connection to initiate a second secure connection |
Families Citing this family (31)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE10138267A1 (en) * | 2001-08-03 | 2003-02-20 | Siemens Ag | Method and terminal device for multipoint communication |
JP2003208085A (en) * | 2002-01-10 | 2003-07-25 | Nova:Kk | Method for providing educational service providing free talking service, device and program for providing educational service |
JP4015934B2 (en) * | 2002-04-18 | 2007-11-28 | 株式会社東芝 | Video coding method and apparatus |
AU2002239208A1 (en) * | 2002-03-22 | 2003-10-08 | Telefonaktiebolaget Lm Ericsson (Publ) | Reducing transmission time for data packets controlled by a link layer protocol comprising a fragmenting/defragmenting capability |
JP3716225B2 (en) * | 2002-04-05 | 2005-11-16 | 松下電器産業株式会社 | Code amount control apparatus and code amount control method |
JP3999091B2 (en) * | 2002-09-25 | 2007-10-31 | 富士フイルム株式会社 | Image correction processing apparatus and program |
KR100539923B1 (en) * | 2003-02-10 | 2005-12-28 | 삼성전자주식회사 | A video encoder capable of encoding deferentially as distinguishing image of user and method for compressing a video signal using that |
JP3787633B2 (en) * | 2003-04-21 | 2006-06-21 | 独立行政法人情報通信研究機構 | Real-time content editing system |
US20050008240A1 (en) * | 2003-05-02 | 2005-01-13 | Ashish Banerji | Stitching of video for continuous presence multipoint video conferencing |
JP4346966B2 (en) * | 2003-06-13 | 2009-10-21 | キヤノン株式会社 | Imaging device |
CN100423539C (en) * | 2004-02-23 | 2008-10-01 | Toa株式会社 | Image compression method, image compression device, image transmission system, data compression pre-processing device, and computer program |
JP2005346202A (en) * | 2004-05-31 | 2005-12-15 | Toshiba Corp | Electronic apparatus |
US7539341B2 (en) * | 2004-07-29 | 2009-05-26 | Xerox Corporation | Systems and methods for processing image data prior to compression |
US7620892B2 (en) * | 2004-07-29 | 2009-11-17 | Xerox Corporation | Server based image processing for client display of documents |
US20060106929A1 (en) * | 2004-10-15 | 2006-05-18 | Kenoyer Michael L | Network conference communications |
US7706576B1 (en) * | 2004-12-28 | 2010-04-27 | Avaya Inc. | Dynamic video equalization of images using face-tracking |
US7460150B1 (en) | 2005-03-14 | 2008-12-02 | Avaya Inc. | Using gaze detection to determine an area of interest within a scene |
WO2006106032A1 (en) * | 2005-04-05 | 2006-10-12 | Thomson Licensing | Method for locally adjusting a quantization step |
US7564476B1 (en) | 2005-05-13 | 2009-07-21 | Avaya Inc. | Prevent video calls based on appearance |
WO2007063922A1 (en) * | 2005-11-29 | 2007-06-07 | Kyocera Corporation | Communication terminal and communication system, and display method of communication terminal |
US8812978B2 (en) * | 2005-12-22 | 2014-08-19 | Xerox Corporation | System and method for dynamic zoom to view documents on small displays |
US7489772B2 (en) | 2005-12-30 | 2009-02-10 | Nokia Corporation | Network entity, method and computer program product for effectuating a conference session |
US8165282B1 (en) | 2006-05-25 | 2012-04-24 | Avaya Inc. | Exploiting facial characteristics for improved agent selection |
JP5181294B2 (en) * | 2008-03-31 | 2013-04-10 | 富士フイルム株式会社 | Imaging system, imaging method, and program |
JP4961582B2 (en) * | 2008-04-07 | 2012-06-27 | 富士フイルム株式会社 | Image processing system, image processing method, and program |
US8913668B2 (en) * | 2008-09-29 | 2014-12-16 | Microsoft Corporation | Perceptual mechanism for the selection of residues in video coders |
US8457194B2 (en) * | 2008-09-29 | 2013-06-04 | Microsoft Corporation | Processing real-time video |
CN101742324A (en) * | 2008-11-14 | 2010-06-16 | 北京中星微电子有限公司 | Video encoding and decoding methods, video encoding and decoding systems and encoder-decoder |
US20100188476A1 (en) * | 2009-01-29 | 2010-07-29 | Optical Fusion Inc. | Image Quality of Video Conferences |
JP2014060577A (en) * | 2012-09-18 | 2014-04-03 | Sony Corp | Communication terminal, program and information processing device |
US11798204B2 (en) * | 2022-03-02 | 2023-10-24 | Qualcomm Incorporated | Systems and methods of image processing based on gaze detection |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4951140A (en) * | 1988-02-22 | 1990-08-21 | Kabushiki Kaisha Toshiba | Image encoding apparatus |
US5675374A (en) * | 1993-11-26 | 1997-10-07 | Fujitsu Limited | Video teleconferencing system |
US5991445A (en) * | 1994-06-27 | 1999-11-23 | Canon Kabushiki Kaisha | Image processing apparatus |
US6112233A (en) * | 1998-02-04 | 2000-08-29 | Intel Corporation | Saving and retrieving a message via a network server using an index page |
US6181711B1 (en) * | 1997-06-26 | 2001-01-30 | Cisco Systems, Inc. | System and method for transporting a compressed video and data bit stream over a communication channel |
US6285661B1 (en) * | 1998-01-28 | 2001-09-04 | Picturetel Corporation | Low delay real time digital video mixing for multipoint video conferencing |
US6343313B1 (en) * | 1996-03-26 | 2002-01-29 | Pixion, Inc. | Computer conferencing system with real-time multipoint, multi-speed, multi-stream scalability |
US6351568B1 (en) * | 1998-03-20 | 2002-02-26 | Canon Kabushiki Kaisha | Image transform and significance bit-plane compression and decompression |
US6456335B1 (en) * | 1998-02-19 | 2002-09-24 | Fujitsu Limited | Multiple picture composing method and multiple picture composing apparatus |
-
1998
- 1998-12-25 JP JP37147898A patent/JP4208315B2/en not_active Expired - Fee Related
-
1999
- 1999-12-22 US US09/469,798 patent/US6744927B1/en not_active Expired - Fee Related
-
2004
- 2004-05-25 US US10/852,159 patent/US20040213474A1/en not_active Abandoned
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4951140A (en) * | 1988-02-22 | 1990-08-21 | Kabushiki Kaisha Toshiba | Image encoding apparatus |
US5675374A (en) * | 1993-11-26 | 1997-10-07 | Fujitsu Limited | Video teleconferencing system |
US5991445A (en) * | 1994-06-27 | 1999-11-23 | Canon Kabushiki Kaisha | Image processing apparatus |
US6343313B1 (en) * | 1996-03-26 | 2002-01-29 | Pixion, Inc. | Computer conferencing system with real-time multipoint, multi-speed, multi-stream scalability |
US6181711B1 (en) * | 1997-06-26 | 2001-01-30 | Cisco Systems, Inc. | System and method for transporting a compressed video and data bit stream over a communication channel |
US6285661B1 (en) * | 1998-01-28 | 2001-09-04 | Picturetel Corporation | Low delay real time digital video mixing for multipoint video conferencing |
US6112233A (en) * | 1998-02-04 | 2000-08-29 | Intel Corporation | Saving and retrieving a message via a network server using an index page |
US6456335B1 (en) * | 1998-02-19 | 2002-09-24 | Fujitsu Limited | Multiple picture composing method and multiple picture composing apparatus |
US6351568B1 (en) * | 1998-03-20 | 2002-02-26 | Canon Kabushiki Kaisha | Image transform and significance bit-plane compression and decompression |
Cited By (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7864938B2 (en) | 2000-12-26 | 2011-01-04 | Polycom, Inc. | Speakerphone transmitting URL information to a remote device |
US9001702B2 (en) | 2000-12-26 | 2015-04-07 | Polycom, Inc. | Speakerphone using a secure audio connection to initiate a second secure connection |
US8977683B2 (en) | 2000-12-26 | 2015-03-10 | Polycom, Inc. | Speakerphone transmitting password information to a remote device |
US8964604B2 (en) | 2000-12-26 | 2015-02-24 | Polycom, Inc. | Conference endpoint instructing conference bridge to dial phone number |
US8948059B2 (en) | 2000-12-26 | 2015-02-03 | Polycom, Inc. | Conference endpoint controlling audio volume of a remote device |
US8976712B2 (en) | 2001-05-10 | 2015-03-10 | Polycom, Inc. | Speakerphone and conference bridge which request and perform polling operations |
US8934382B2 (en) | 2001-05-10 | 2015-01-13 | Polycom, Inc. | Conference endpoint controlling functions of a remote device |
US8805928B2 (en) | 2001-05-10 | 2014-08-12 | Polycom, Inc. | Control unit for multipoint multimedia/audio system |
US7787605B2 (en) * | 2001-12-31 | 2010-08-31 | Polycom, Inc. | Conference bridge which decodes and responds to control information embedded in audio information |
US7742588B2 (en) | 2001-12-31 | 2010-06-22 | Polycom, Inc. | Speakerphone establishing and using a second connection of graphics information |
US7978838B2 (en) | 2001-12-31 | 2011-07-12 | Polycom, Inc. | Conference endpoint instructing conference bridge to mute participants |
US20050213728A1 (en) * | 2001-12-31 | 2005-09-29 | Polycom, Inc. | Conference endpoint instructing a remote device to establish a new connection |
US8102984B2 (en) | 2001-12-31 | 2012-01-24 | Polycom Inc. | Speakerphone and conference bridge which receive and provide participant monitoring information |
US20050213736A1 (en) * | 2001-12-31 | 2005-09-29 | Polycom, Inc. | Speakerphone establishing and using a second connection of graphics information |
US8144854B2 (en) | 2001-12-31 | 2012-03-27 | Polycom Inc. | Conference bridge which detects control information embedded in audio information to prioritize operations |
US8947487B2 (en) | 2001-12-31 | 2015-02-03 | Polycom, Inc. | Method and apparatus for combining speakerphone and video conference unit operations |
US8223942B2 (en) | 2001-12-31 | 2012-07-17 | Polycom, Inc. | Conference endpoint requesting and receiving billing information from a conference bridge |
US8705719B2 (en) | 2001-12-31 | 2014-04-22 | Polycom, Inc. | Speakerphone and conference bridge which receive and provide participant monitoring information |
US20050213738A1 (en) * | 2001-12-31 | 2005-09-29 | Polycom, Inc. | Conference endpoint requesting and receiving billing information from a conference bridge |
US8885523B2 (en) | 2001-12-31 | 2014-11-11 | Polycom, Inc. | Speakerphone transmitting control information embedded in audio information through a conference bridge |
US8934381B2 (en) | 2001-12-31 | 2015-01-13 | Polycom, Inc. | Conference endpoint instructing a remote device to establish a new connection |
US8004556B2 (en) | 2004-04-16 | 2011-08-23 | Polycom, Inc. | Conference link between a speakerphone and a video conference unit |
US7580685B2 (en) * | 2004-07-09 | 2009-08-25 | Hitachi Kokusai Electric, Inc. | Radio communication system |
US20060009254A1 (en) * | 2004-07-09 | 2006-01-12 | Masayuki Kanazawa | Radio communication system |
US7796565B2 (en) | 2005-06-08 | 2010-09-14 | Polycom, Inc. | Mixed voice and spread spectrum data signaling with multiplexing multiple users with CDMA |
US8199791B2 (en) | 2005-06-08 | 2012-06-12 | Polycom, Inc. | Mixed voice and spread spectrum data signaling with enhanced concealment of data |
US20070047626A1 (en) * | 2005-06-08 | 2007-03-01 | Polycom, Inc | Mixed voice and spread spectrum data signaling with multiplexing multiple users with cdma |
US8126029B2 (en) | 2005-06-08 | 2012-02-28 | Polycom, Inc. | Voice interference correction for mixed voice and spread spectrum data signaling |
Also Published As
Publication number | Publication date |
---|---|
JP2000197043A (en) | 2000-07-14 |
JP4208315B2 (en) | 2009-01-14 |
US6744927B1 (en) | 2004-06-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6744927B1 (en) | Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system | |
US7039675B1 (en) | Data communication control apparatus and method adapted to control distribution of data corresponding to various types of a plurality of terminals | |
US6989868B2 (en) | Method of converting format of encoded video data and apparatus therefor | |
US6593955B1 (en) | Video telephony system | |
US5634040A (en) | Data communication apparatus and method having concurrent image overlay function | |
US7006575B2 (en) | Picture and sound decoding apparatus picture and sound encoding apparatus and information transmission system | |
US6091857A (en) | System for producing a quantized signal | |
US6909810B2 (en) | Image processing apparatus and method, and its computer program and storage medium | |
US5577191A (en) | System and method for digital video editing and publishing, using intraframe-only video data in intermediate steps | |
US20030220971A1 (en) | Method and apparatus for video conferencing with audio redirection within a 360 degree view | |
US6084911A (en) | Transmission of coded and compressed voice and image data in fixed bit length data packets | |
JP2001514826A (en) | Method and apparatus for transmitting and displaying still images | |
JP2007166625A (en) | Video data encoder, video data encoding method, video data decoder, and video data decoding method | |
JPH10150663A (en) | Method and device for minimizing blocking effect | |
US20040001091A1 (en) | Method and apparatus for video conferencing system with 360 degree view | |
EP0805600A2 (en) | Compressed video text overlay | |
US6859557B1 (en) | System and method for selective decoding and decompression | |
US20010012397A1 (en) | Image processing apparatus and method | |
JP2002262288A (en) | Image data transmitter and image data receiver | |
KR100312421B1 (en) | A conversion method of the compressed moving video on the video communication system | |
US8086056B2 (en) | Encoding device and method, decoding device and method, and program | |
JPH08186814A (en) | Image compressor | |
US6462783B1 (en) | Picture encoding method and apparatus | |
JPH09116759A (en) | Image decoder and image coding decoding system | |
KR20020070721A (en) | Streaming device for moving picture |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |