US20040213474A1 - Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system - Google Patents

Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system Download PDF

Info

Publication number
US20040213474A1
US20040213474A1 US10/852,159 US85215904A US2004213474A1 US 20040213474 A1 US20040213474 A1 US 20040213474A1 US 85215904 A US85215904 A US 85215904A US 2004213474 A1 US2004213474 A1 US 2004213474A1
Authority
US
United States
Prior art keywords
terminals
image data
image
connection device
multipoint connection
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US10/852,159
Inventor
Masami Kato
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Canon Inc
Original Assignee
Canon Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Canon Inc filed Critical Canon Inc
Priority to US10/852,159 priority Critical patent/US20040213474A1/en
Publication of US20040213474A1 publication Critical patent/US20040213474A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H03ELECTRONIC CIRCUITRY
    • H03MCODING; DECODING; CODE CONVERSION IN GENERAL
    • H03M13/00Coding, decoding or code conversion, for error detection or error correction; Coding theory basic assumptions; Coding bounds; Error probability evaluation methods; Channel models; Simulation or testing of codes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • G06T9/005Statistical coding, e.g. Huffman, run length coding
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B20/00Signal processing not specific to the method of recording or reproducing; Circuits therefor
    • G11B20/10Digital recording or reproducing
    • G11B20/18Error detection or correction; Testing, e.g. of drop-outs

Definitions

  • the present invention relates to a data communication control apparatus and its control method, an image processing apparatus and its method, and a data communication system which realize interactive video communication among a plurality of points.
  • a multipoint video conference system widely used now enables video conference terminals based on the ITU-T Recommendation H. 320 and the like to perform interactive communication among a number of points via a multipoint connection device.
  • the ITU-T Recommendation H. 231 defining the functional construction of the multipoint connection device and the format of multipoint communication
  • the ITU-T Recommendation H. 243 determining an in-channel communication procedure and the like have been made.
  • FIG. 1 shows the construction of a conventional video conference system.
  • a multipoint connection device 22 interconnects video conference terminals 21 a to 21 n at three or more points, and performs audio mixing, video data delivery control or video data combining for split screen view, further, performs presidential control over the conference.
  • a multipoint conference connecting a plurality of remote points can be realized by connecting terminals via a multipoint connection device as long as the terminals are based on the Recommendation H. 320 .
  • FIG. 11 shows video data transfer in this system.
  • numerals 41 a to 41 c denote client terminals realized by personal computers or the like having video and audio input/output functions; 42 , a multipoint connection server which delivers video and audio data from the respective client terminals 41 a to 41 c to the client terminals 41 a to 41 c without any processing.
  • the multipoint connection server 42 delivers video and audio data from the respective client terminals 41 a to 41 c to the client terminals 41 a to 41 c without any processing.
  • the client terminal 41 a transmits terminal information A to the multipoint connection server 42
  • the multipoint connection server 42 transfers terminal information B and C from other client terminals 41 b and 41 c to the client terminal 41 a.
  • the code amount of the image data delivered from the multipoint connection server 42 to the client terminals 41 a to 41 c increases in proportion to the number of conference participants (terminals).
  • a data communication control apparatus comprising: compression means for compressing image data to be transmitted; communication means for transmitting the image data compressed by the compression means to the terminals; recognition means for recognizing a specific area in the image data; and compression control means for controlling a compression method in the compression means in accordance with whether or not an object area in the image data is the specific area.
  • the foregoing object is attained by providing the data communication control apparatus wherein the specific area is an area showing a person's face.
  • an image processing apparatus comprising: recognition means for recognizing a-specific area in an image represented by compressed image data; and compression control means for controlling a re-compression method for re-compressing the compressed image data based on recognition by the recognition means.
  • FIG. 1 is a block diagram showing the construction of a multipoint video conference system according to a first embodiment of the present invention
  • FIG. 2 is a block diagram showing the basic construction of a multipoint connection device
  • FIG. 3 is an example of display screen of a video conference terminal
  • FIG. 4 is a flowchart showing image processing by the multipoint connection device
  • FIG. 5 is an example of image delivered in the first embodiment
  • FIG. 6 is a block diagram showing the basic construction of the multipoint connection device according to a second embodiment of the present invention.
  • FIG. 7 is a flowchart showing image processing by the multipoint connection device
  • FIG. 8 is an example of a spatial filter
  • FIG. 9 is a block diagram showing the construction of the multipoint connection device according to a third embodiment of the present invention.
  • FIG. 10 is an example of image processing by the conventional multipoint connection device.
  • FIG. 11 is a block diagram showing an Internet-type multipoint video conference system.
  • FIG. 12 is a block diagram showing the another construction of a multipoint connection device.
  • the basic construction of a video conference system according to a first embodiment of the present invention is similar to that of the above-described prior art as shown in FIG. 1.
  • the video conference terminals 21 a to 21 n (hereinbelow, simply referred to as “terminals”) at three or more points are interconnected via the multipoint connection device 22 .
  • Image data outputted from the respective terminals 21 a to 21 d are delivered to the other terminals via the multipoint connection device 22 .
  • FIG. 2 is a block diagram showing the basic construction of the multipoint connection device 22 in the present embodiment.
  • numerals 11 a to 11 n denote network interface units for accommodating the remote terminals 21 a to 21 n, as line interfaces for ISDN, PSTN and the like; and 12 a to 12 n, multiplexing/demultiplexing units which perform multiplexing/demultiplexing on various image, audio and control data in accordance with, e.g., the ITU-T Recommendation H.221.
  • the number of the network interface units 11 a to 11 n and that of the multiplexing/demultiplexing units 12 a to 12 n correspond to the number of video conference terminals which can be directly connected to the multipoint connection device. Further, the multipoint connection device 22 can be connected to another multipoint connection device, a server or the like via the network interface units 11 a to 11 n and the multiplexing/demultiplexing units 12 a to 12 n.
  • Numeral 13 denotes an audio processor which performs decompression/mixing processing on audio code data, transmitted from the respective terminals 21 a to 21 n and demultiplexed by the multiplexing/demultiplexing units 12 a to 12 n, then performs compression encoding on the data again, and delivers the data to the respective terminals 21 a to 21 d.
  • Numeral 14 denotes an image decoder which decodes compressed image code data demultiplexed by the multiplexing/demultiplexing units 12 a to 12 n, thus generating uncompressed image data.
  • Numeral 15 denotes a face-area recognition unit which recognizes a face area of a person from image data decoded by the image decoder 14 ; 16 , an image encoder which again compresses the image data decoded by the image decoder 14 ; and 17 , a quantization controller which performs quantization control upon re-compression by the image encoder 16 , in accordance with the result of face area recognition by the face-area recognition unit 15 . More specifically, a plurality of quantization tables are prepared, and if it is determined that a quantization object image is not a face area image, a quantization table is selected such that coefficients of the table are set for a quantization step greater than that used in decoding.
  • Numeral 18 denotes a system controller which performs various communication protocol processing and the entire operation of the multipoint connection device 22 .
  • FIG. 3 is an example of image displayed on the terminals 21 a to 21 n in the multipoint video conference system of the present embodiment.
  • numeral 71 denotes a window to display a video image of a user of the terminal 21 d; 72 a to 72 c, windows to display video images of users of the other terminals (e.g., 21 a to 21 c ) as participants of the conference; and 73 , a common window in the conference to display, e.g., a white board.
  • Multiplexed multimedia data sent via the communication interface units 11 a to 11 n accommodating the terminals 21 a to 21 n are frame-demultiplexed by the multiplexing/demultiplexing units 12 a to 12 n, and audio compressed data, video compressed data, data conference low-speed data, system control data and the like are extracted (step S 1001 ).
  • audio compressed data, video compressed data, data conference low-speed data, system control data and the like are extracted (step S 1001 ).
  • processing on the demultiplexed video compressed data will be described.
  • the demultiplexed audio compressed data is decompressed by the audio processor 13 , and mixed with other decompressed audio data demultiplexed by the respective multiplexing/demultiplexing units 12 a to 12 n. Then, the mixed audio data is compressed, multiplexed with other data such as image data by the multiplexing/demultiplexing units 12 a to 12 n, and delivered to the respective terminals 21 a to 21 n.
  • the system controller 18 determines whether or not code amount control is performed on the demultiplexed video compressed data in accordance with an instruction from a terminal as a delivery destination of the data (step S 1002 ). If the code amount control is not performed, the video compressed data is multiplexed again with other data by the multiplexing/demultiplexing units 12 a to 12 n (step S 1009 ). In this arrangement, regarding a video image from a specific terminal, specified by a user, if the user instructs not to perform the above-described code amount control, the video image can be delivered as a fine image with a clear background image.
  • the video compressed data is decoded by the image decoder 14 (step S 1003 ), and temporarily stored in a RAM (not shown) or the like in the image decoder 14 .
  • the face-area recognition unit 15 performs face area recognition processing on the decoded image data (step S 1004 ).
  • the face area recognition by the face-area recognition unit 15 an area including a person's face can be recognized by, e.g., a method described in “Extraction of Plural Face Areas from a Natural Image by Using Genetic Algorithm” (Yokoo and Hagiwara, The Institute of Electrical Engineers of Japan Proceedings C, Vol. 117-C, No. 9, September, 1997) or the like.
  • the face area recognition method is not limited to any particular method, but any method can be used.
  • a higher level recognition method such as extraction of exact face area outline can be employed as long as the multipoint connection device has sufficiently high processing capability.
  • the decoded image data is re-compressed by the image encoder 16 .
  • the quantization coefficients are controlled based on the result of face area recognition.
  • the result of face area recognition is outputted to the quantization controller 17 .
  • the quantization controller 17 has quantization coefficients used in decoding by the image decoder 14 (or Q factors for quantization) and quantization coefficients multiplied by n (n>1). Note that-the former quantization coefficients-are set as default coefficients.
  • step S 1005 As a non-face area, the quantization coefficients having the set default values are changed to the quantization coefficients multiplied by n (step S 1006 ). Note that regarding an area determined as a face area, the same quantization coefficients (default values) as those in the decoding are used.
  • the image encoder 16 performs quantization based on the set quantization coefficients on the image data held in the image decoder 14 , thus performing compression processing (step S 1007 ).
  • the multipoint connection device 22 performs the above processings independently on data transferred from the respective terminals 21 a to 21 n.
  • the re-compressed image data are multiplexed with the audio compressed data by the multiplexing/demultiplexing units 12 a to 12 n (step S 1008 ), and delivered to the respective terminals 21 a to 21 n in accordance with a communication protocol.
  • FIG. 5 is an example of displayed image, obtained by e.g. a camera (not shown) of the terminal 21 d, then processed by the multipoint connection device 22 as described above, and delivered to the respective terminals 21 a to 21 n, in the video conference system of the present embodiment.
  • an area in a circle 51 is recognized as a face area; and an area 52 is a background area recognized as a non-face area.
  • the background area 52 quantized by a greater quantization step, becomes a low-resolution image, while the face area 51 has image quality equivalent to that of an image obtained by the terminal 21 d.
  • a significant portion of an image obtained by a terminal can be transmitted with high precision to the respective terminals of video conference system. Accordingly, even if the respective terminals are connected by a narrow band communication channel, the significant portion can be transmitted in high-quality image communication.
  • a significant portion, face area, i.e., the facial expression of the user of the other terminal can be clearly displayed without reduction or frame thinning.
  • an image from another arbitrary terminal may be received as an entirely high quality image at any time.
  • the decoding processing and the encoding processing in the present embodiment may be performed in predetermined image data block units. In this case, the above-described processings at steps S 1003 to S 1007 are repeated in each block.
  • an area where high image quality is maintained is not limited to a face area.
  • the same advantage as that of the present embodiment can be obtained by setting an arbitrary significant object as an object of high-precision display in accordance with the content of conference, and transferring other image areas with degraded image quality.
  • FIG. 6 is a block diagram showing the construction of the multipoint connection device 22 according to the second embodiment of the present invention.
  • constituents corresponding to those in FIG. 2 have the same reference numerals and the explanations of these constituents will be omitted.
  • numeral 86 denotes a spatial filter processor which performs low-frequency filter processing on decoded image data based on the result of recognition by the face-area recognition unit 15 . More specifically, the spatial filter processor 86 performs spatial filter processing on areas other than a face area by a low-pass filter, to attenuate high frequency components.
  • the decoded image data is re-compressed by the image encoder 16 .
  • spatial filter processing is performed based on the result of face area recognition.
  • the spatial filter processor 86 inputs decoded image data and the result of face area recognition.
  • a low-pass filter is prepared in advance, and spatial filter processing by the low-pass filter is performed (step S 1106 ) on an area determined at step S 1005 as a non-face area.
  • the spatial filter processed image data is compressed by the image encoder 16 (step S 1008 ).
  • FIG. 8 is an example of the low-pass filter in the spatial filter processor 86 .
  • the filter processing expressed by the following equation (1) is performed by using the low-pass filter.
  • P ′( x,y ) ( P ( x ⁇ 1 ,y ⁇ 1)+ P ( x,y ⁇ 1)+ P ( x+ 1 ,y ⁇ 1 )+ P ( x ⁇ 1 ,y )+ P ( x,y )+( x +1 ,y )+ P ( x,y +1)+ P ( x +1 ,y +1))/9 (1)
  • P(x,y) is a pixel value of coordinates (x,y).
  • the spatial filter in the second embodiment is not limited to that as shown in FIG. 8, but any filter may be employed as long as it is a low-pass filter.
  • execution/non-execution of spatial filter processing is controlled in accordance with whether or not an object area is a face area (specific area), however, the present invention is not limited to this arrangement.
  • spatial filter processing may be performed on a specific area and other areas.
  • a low-pass filer used in the processing is selected in accordance with whether an object area is a specific area or not.
  • a filter for the other areas than the face area is selected so as to remove high frequency components more than those in the face area removed by a filter for the face area.
  • a specific area with great significance such as a face area is re-compressed without degrading image quality, and in areas other than the specific area with less significance such as background, high frequency components are attenuated. This reduces the total code amount, and accordingly, obtains the same advantage as that of the first embodiment.
  • the camera of a terminal may be operated such that the user's face comes to the center of the obtained image.
  • an undesirable portion such as background other than a person's face might be arbitrarily image-sensed.
  • detecting a face area is rather advantageous than previously setting low compression rate area and high compression rate area within the obtained image.
  • the face area detection unnecessitates operation of the orientation of the camera, and further ensures privacy protection for conference participants.
  • the principal processings of the multipoint connection device 22 having the functions as described in the first embodiment are realized by software.
  • FIG. 9 is a block diagram showing the construction of the multipoint connection device 22 according to the third embodiment of the present invention.
  • Numeral 61 denotes a communication interface unit for communication with an ISDN, a PSTN and the like; and 62 , a media processor which performs principal signal processings.
  • the media processor 62 performs various digital signal processings by a specialized processor for multimedia signal processing, a DSP (Digital Signal Processor) and the like.
  • the video conference system can be realized with a simple and low-cost construction without preparing dedicated hardware, by processing various data by multitask processing by utilizing the media processor 62 .
  • Numeral 63 denotes a RAM (Random Access Memory) which is used as a work memory necessary for the operation of the media processor 62 and as a buffer memory for various image and audio data to be processed.
  • the RAM 63 is realized by, e.g., a high-speed large-capacity memory such as an SDRAM (Synchronous DRAM).
  • Numeral 64 denotes a CPU which performs system control for the multipoint connection device 22 , communication protocol processing and the like; 65 , a RAM used as a work buffer necessary for the operation of the CPU 64 ; and 66 , a ROM (Read Only Memory) in which a control program necessary for the operation of the CPU 64 and a control program necessary for the operation of the media processor 62 are stored. The program necessary for the operation of the media processor 62 is transferred from the ROM 66 to the RAM 63 upon system initialization.
  • a control program to realize the processing is stored in the ROM 66 in advance.
  • the program is developed on the RAM 63 or the RAM 65 , and executed by the media processor 62 or the CPU 64 .
  • processing related to data compression/decompression is performed by the media processor 62 , and other processings are controlled by the CPU 64 .
  • the multipoint connection device 22 as described in the first embodiment is realized by software.
  • the multipoint connection device 22 as described in the second embodiment may be realized by software.
  • the compression method employed in the above-described first to third embodiments may be any method.
  • an H. 261 coding method or the like based on DCT (Discrete Cosine Transformation) used in the Recommendation H.320 may be used.
  • the control on the compression method (quantization coefficients) based on the result of specific-area recognition is performed in macroblock units.
  • the compression control method of the present invention is not limited to the quantization control or spatial filter control as described in the first to third embodiments, but any control method may be performed as long as it controls the degree of image quality degradation due to re-compression in accordance with whether or not an object area is a specific area. In other words, any control method may be used as long as it controls the compression rate in re-compression.
  • more flexible image quality control including further improvement in image quality of delivered image data can be realized by combining the quantization control described in the first embodiment and the spatial filter control described in the second embodiment.
  • the video conference terminals are directly connected to a public line such as PSTN/ISDN line, however, the present invention is not limited to this example.
  • the video conference terminals may be connected to the Internet, an Intranet or the like.
  • communication is performed with a plurality of terminals by the Internet Protocol via a network access controller 121 having a LAN interface such as the Ethernet.
  • the means for communication with the respective terminals is logically constructed for each terminal. Control on packetized data and data multiplexing/demultiplexing processing are performed by the network access controller 121 and a system controller 122 .
  • the video conference terminals connected to the multipoint connection device of the present invention may be dedicated terminals or general-purpose terminals such as personal computers where a predetermined application (software) program is operated.
  • the multipoint connection device of the present invention may have an automatic HTML (Hyper Text Markup Language) data generating function and an HTTP (Hyper Text Transfer Protocol) server function in addition to the function related to multipoint terminal connection.
  • general-purpose communication terminals which display and visualize image and audio data stored in the HTTP server of the multipoint connection device, can be connected to the multipoint connection device.
  • general-purpose communication terminals personal computers having a so-called WWW-browser and network computers can be employed. Further, by virtue of recent development of semiconductor technologies, these computers may be employed as quick and convenient terminals even in a mobile communication environment.
  • These general-purpose communication terminals are IP (Internet Protocol)-connected to the multipoint connection terminal via a public network such as an ISDN or a PSTN or a LAN (Local Area Network). This enables interconnection among the existing video conference terminals and the general-purpose terminals such as personal computers and network computers.
  • IP Internet Protocol
  • the present invention can be applied to a system constituted by a plurality of devices (e.g., a host computer, an interface, a reader and a printer) or to an apparatus comprising a single device (e.g., a copy machine or a facsimile apparatus).
  • the object of the present invention can be also achieved by providing a storage medium storing program code for performing the aforesaid processes to a system or an apparatus, reading the program code with a computer (e.g., CPU, MPU) of the system or apparatus from the storage medium, then executing the program.
  • a computer e.g., CPU, MPU
  • the program code read from the storage medium realize the functions according to the embodiments, and the storage medium storing the program code constitutes the invention.
  • the storage medium such as a floppy disk, a hard disk, an optical disk, a magneto-optical disk, CD-ROM, CD-R, a magnetic tape, a non-volatile type memory card, and ROM can be used for providing the program code.
  • the present invention includes a case where an OS (operating system) or the like working on the computer performs a part or entire processes in accordance with designations of the program code and realizes functions according to the above embodiments.
  • a product obtained by the image processing method of the present invention e.g., printed matter, is included in the present invention.
  • the present invention also includes a case where, after the program code read from the storage medium are written in a function expansion card which is inserted into the computer or in a memory provided in a function expansion unit which is connected to the computer, CPU or the like contained in the function expansion card or unit performs a part or entire process in accordance with designations of the program code and realizes functions of the above embodiments.
  • the storage medium stores program code corresponding to the flowcharts shown in FIGS. 4 and 7 described in the embodiments.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Probability & Statistics with Applications (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Compression Of Band Width Or Redundancy In Fax (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)
  • Communication Control (AREA)
  • Telephonic Communication Services (AREA)

Abstract

In video conference system terminals, the conventional difficulty in high-quality display of facial expressions of a plurality of participants is solved. In a multipoint connection device connecting a plurality of video conference terminals, image data transmitted from a terminal is decoded, and a face area in the image data is recognized. Then, quantization coefficients for the face area are set to be greater coefficients than those for other areas than the face area, then the image data is compressed, and delivered to the respective terminals. By this arrangement, in image data, a face area with great significance can be re-compressed without degradation of image quality, while an area with less significance such as background can be compressed with high efficiency. Thus, the total code amount can be reduced. Accordingly, even if the system uses a narrow band communication channel, the significant face area can be clearly displayed at the respective terminals.

Description

    BACKGROUND OF THE INVENTION
  • The present invention relates to a data communication control apparatus and its control method, an image processing apparatus and its method, and a data communication system which realize interactive video communication among a plurality of points. [0001]
  • A multipoint video conference system widely used now enables video conference terminals based on the ITU-T Recommendation H. 320 and the like to perform interactive communication among a number of points via a multipoint connection device. [0002]
  • As standards related to the multipoint connection device to control such multipoint connection among video conference terminals, the ITU-T Recommendation H. 231 defining the functional construction of the multipoint connection device and the format of multipoint communication, the ITU-T Recommendation H. 243 determining an in-channel communication procedure and the like have been made. [0003]
  • FIG. 1 shows the construction of a conventional video conference system. In FIG. 1, a [0004] multipoint connection device 22 interconnects video conference terminals 21 a to 21 n at three or more points, and performs audio mixing, video data delivery control or video data combining for split screen view, further, performs presidential control over the conference. In this manner, a multipoint conference connecting a plurality of remote points can be realized by connecting terminals via a multipoint connection device as long as the terminals are based on the Recommendation H. 320.
  • To realize multipoint interactive video communication in a video conference system based on the Recommendation H. 320, to deliver image data from the [0005] multipoint connection device 22 to the video conference terminals 21 a to 21 n as shown in FIG. 1, the following methods (1), (2) and the like have been put into practical use:
  • (1) A method to deliver a video image from a specific video conference terminal designated by a presidential terminal to the respective video conference terminals. [0006]
  • (2) As shown in FIG. 10, a method to reduce video images from respective [0007] video conference terminals 21 a to 21 n in the multipoint connection device 22, so as to simultaneously deliver images from respective video conference terminals 21 a to 21 n within a limited band.
  • Note that in FIG. 10, [0008] numerals 31 a to 31 d denote video data transmitted from the respective video conference terminals 21 a to 21 d; 32, reduced and synthesized video data delivered from the multipoint connection device 22 to the respective video conference terminals 21 a to 21 d.
  • Further, to realize a video conference system in an Internet environment, software programs such as CU-SeeMe by White Pine Software, Inc. have been provided. In the multipoint video conference system, image data transmitted from the respective client terminals are delivered via a multipoint connection server to the respective client terminals without any processing. FIG. 11 shows video data transfer in this system. In FIG. 11, [0009] numerals 41 a to 41 c denote client terminals realized by personal computers or the like having video and audio input/output functions; 42, a multipoint connection server which delivers video and audio data from the respective client terminals 41 a to 41 c to the client terminals 41 a to 41 c without any processing.
  • However, the conventional video conference system using the multipoint connection based on the Recommendation H. 320 or the like, in which communication is performed within a limited band to the respective terminals, has problems as follows. [0010]
  • (1) Only the video image of a particular reference participant is delivered; or [0011]
  • (2) As reduced images of plurality of participants are transmitted, the images of the respective speakers are unclear. [0012]
  • In the video conference terminal, it is difficult to simultaneously and clearly display a plurality of conference participants. [0013]
  • Further, in the video conference system utilizing the Internet as shown in FIG. 11, the [0014] multipoint connection server 42 delivers video and audio data from the respective client terminals 41 a to 41 c to the client terminals 41 a to 41 c without any processing. For example, between the multipoint connection server 42 and the client terminal 41 a, the client terminal 41 a transmits terminal information A to the multipoint connection server 42, and the multipoint connection server 42 transfers terminal information B and C from other client terminals 41 b and 41 c to the client terminal 41 a. Accordingly, the code amount of the image data delivered from the multipoint connection server 42 to the client terminals 41 a to 41 c increases in proportion to the number of conference participants (terminals). If a sufficient band is not ensured as a communication channel, the frame rate of images received by the client terminals 41 a to 41 c is extremely reduced. Accordingly, to display a received image, so-called frame thinning, i.e., omitting a frame at predetermined intervals, must be performed.
  • SUMMARY OF THE INVENTION
  • Accordingly, it is an object of the present invention to provide a data communication control apparatus and its control method and a data communication system which enable high-quality image communication. It is another object of the present invention to provide a data communication control apparatus and its control method and a data communication system which enables image communication without degration of the quality of the important portion of an image even if using a narrow band communication channel. [0015]
  • According to the present invention, the foregoing objects are attained by providing a data communication control apparatus comprising: compression means for compressing image data to be transmitted; communication means for transmitting the image data compressed by the compression means to the terminals; recognition means for recognizing a specific area in the image data; and compression control means for controlling a compression method in the compression means in accordance with whether or not an object area in the image data is the specific area. [0016]
  • It is another object of the present invention to provide a data communication control apparatus and its control method and a data communication system which enable clearly displayed facial expressions of participants in a multipoint video conference system. [0017]
  • According to the present invention, the foregoing object is attained by providing the data communication control apparatus wherein the specific area is an area showing a person's face. [0018]
  • It is further object of the present invention to provide an image processing apparatus and its method which mitigate degradation of image quality upon reuse of compressed image data. [0019]
  • According to the present invention, the foregoing object is attained by providing an image processing apparatus comprising: recognition means for recognizing a-specific area in an image represented by compressed image data; and compression control means for controlling a re-compression method for re-compressing the compressed image data based on recognition by the recognition means. [0020]
  • Other features and advantages of the present invention will be apparent from the following description taken in conjunction with the accompanying drawings, in which like reference characters designate the same name or similar parts throughout the figures thereof.[0021]
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The accompanying drawings, which are incorporated in and constitute a part of the specification, illustrate embodiments of the invention and, together with the description, serve to explain the principles of the invention. [0022]
  • FIG. 1 is a block diagram showing the construction of a multipoint video conference system according to a first embodiment of the present invention; [0023]
  • FIG. 2 is a block diagram showing the basic construction of a multipoint connection device; [0024]
  • FIG. 3 is an example of display screen of a video conference terminal; [0025]
  • FIG. 4 is a flowchart showing image processing by the multipoint connection device; [0026]
  • FIG. 5 is an example of image delivered in the first embodiment; [0027]
  • FIG. 6 is a block diagram showing the basic construction of the multipoint connection device according to a second embodiment of the present invention; [0028]
  • FIG. 7 is a flowchart showing image processing by the multipoint connection device; [0029]
  • FIG. 8 is an example of a spatial filter; [0030]
  • FIG. 9 is a block diagram showing the construction of the multipoint connection device according to a third embodiment of the present invention; [0031]
  • FIG. 10 is an example of image processing by the conventional multipoint connection device; and [0032]
  • FIG. 11 is a block diagram showing an Internet-type multipoint video conference system. [0033]
  • FIG. 12 is a block diagram showing the another construction of a multipoint connection device.[0034]
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • Preferred embodiments of the present invention will now be described in detail in accordance with the accompanying drawings. [0035]
  • First Embodiment
  • The basic construction of a video conference system according to a first embodiment of the present invention is similar to that of the above-described prior art as shown in FIG. 1. The [0036] video conference terminals 21 a to 21 n (hereinbelow, simply referred to as “terminals”) at three or more points are interconnected via the multipoint connection device 22. Image data outputted from the respective terminals 21 a to 21 d are delivered to the other terminals via the multipoint connection device 22.
  • FIG. 2 is a block diagram showing the basic construction of the [0037] multipoint connection device 22 in the present embodiment. In FIG. 2, numerals 11 a to 11 n denote network interface units for accommodating the remote terminals 21 a to 21 n, as line interfaces for ISDN, PSTN and the like; and 12 a to 12 n, multiplexing/demultiplexing units which perform multiplexing/demultiplexing on various image, audio and control data in accordance with, e.g., the ITU-T Recommendation H.221. Note that in the multipoint connection device 22, the number of the network interface units 11 a to 11 n and that of the multiplexing/demultiplexing units 12 a to 12 n correspond to the number of video conference terminals which can be directly connected to the multipoint connection device. Further, the multipoint connection device 22 can be connected to another multipoint connection device, a server or the like via the network interface units 11 a to 11 n and the multiplexing/demultiplexing units 12 a to 12 n.
  • [0038] Numeral 13 denotes an audio processor which performs decompression/mixing processing on audio code data, transmitted from the respective terminals 21 a to 21 n and demultiplexed by the multiplexing/demultiplexing units 12 a to 12 n, then performs compression encoding on the data again, and delivers the data to the respective terminals 21 a to 21 d. Numeral 14 denotes an image decoder which decodes compressed image code data demultiplexed by the multiplexing/demultiplexing units 12 a to 12 n, thus generating uncompressed image data.
  • Numeral [0039] 15 denotes a face-area recognition unit which recognizes a face area of a person from image data decoded by the image decoder 14; 16, an image encoder which again compresses the image data decoded by the image decoder 14; and 17, a quantization controller which performs quantization control upon re-compression by the image encoder 16, in accordance with the result of face area recognition by the face-area recognition unit 15. More specifically, a plurality of quantization tables are prepared, and if it is determined that a quantization object image is not a face area image, a quantization table is selected such that coefficients of the table are set for a quantization step greater than that used in decoding.
  • Numeral [0040] 18 denotes a system controller which performs various communication protocol processing and the entire operation of the multipoint connection device 22.
  • FIG. 3 is an example of image displayed on the [0041] terminals 21 a to 21 n in the multipoint video conference system of the present embodiment. Assuming that the image in FIG. 5 is a displayed image on the terminal 21 d, numeral 71 denotes a window to display a video image of a user of the terminal 21 d; 72 a to 72 c, windows to display video images of users of the other terminals (e.g., 21 a to 21 c) as participants of the conference; and 73, a common window in the conference to display, e.g., a white board.
  • Next, image processing by the [0042] multipoint connection device 22 will be described in detail with reference to the flowchart of FIG. 4.
  • Multiplexed multimedia data sent via the [0043] communication interface units 11 a to 11 n accommodating the terminals 21 a to 21 n are frame-demultiplexed by the multiplexing/demultiplexing units 12 a to 12 n, and audio compressed data, video compressed data, data conference low-speed data, system control data and the like are extracted (step S1001). Hereinbelow, processing on the demultiplexed video compressed data will be described.
  • Note that the demultiplexed audio compressed data is decompressed by the [0044] audio processor 13, and mixed with other decompressed audio data demultiplexed by the respective multiplexing/demultiplexing units 12 a to 12 n. Then, the mixed audio data is compressed, multiplexed with other data such as image data by the multiplexing/demultiplexing units 12 a to 12 n, and delivered to the respective terminals 21 a to 21 n.
  • The [0045] system controller 18 determines whether or not code amount control is performed on the demultiplexed video compressed data in accordance with an instruction from a terminal as a delivery destination of the data (step S1002). If the code amount control is not performed, the video compressed data is multiplexed again with other data by the multiplexing/demultiplexing units 12 a to 12 n (step S1009). In this arrangement, regarding a video image from a specific terminal, specified by a user, if the user instructs not to perform the above-described code amount control, the video image can be delivered as a fine image with a clear background image.
  • On the other hand, if the code amount control is performed, the video compressed data is decoded by the image decoder [0046] 14 (step S1003), and temporarily stored in a RAM (not shown) or the like in the image decoder 14. Then the face-area recognition unit 15 performs face area recognition processing on the decoded image data (step S1004). Note that as the face area recognition by the face-area recognition unit 15, an area including a person's face can be recognized by, e.g., a method described in “Extraction of Plural Face Areas from a Natural Image by Using Genetic Algorithm” (Yokoo and Hagiwara, The Institute of Electrical Engineers of Japan Proceedings C, Vol. 117-C, No. 9, September, 1997) or the like.
  • Note that in the present embodiment, the face area recognition method is not limited to any particular method, but any method can be used. A higher level recognition method such as extraction of exact face area outline can be employed as long as the multipoint connection device has sufficiently high processing capability. Further, in some face area recognition method, it is not necessary for the [0047] image decoder 14 to completely decode the image data.
  • After the face area recognition, the decoded image data is re-compressed by the [0048] image encoder 16. In the present embodiment, upon compression processing, the quantization coefficients are controlled based on the result of face area recognition.
  • The result of face area recognition is outputted to the [0049] quantization controller 17. The quantization controller 17 has quantization coefficients used in decoding by the image decoder 14 (or Q factors for quantization) and quantization coefficients multiplied by n (n>1). Note that-the former quantization coefficients-are set as default coefficients.
  • Regarding an area determined at step S[0050] 1005 as a non-face area, the quantization coefficients having the set default values are changed to the quantization coefficients multiplied by n (step S1006). Note that regarding an area determined as a face area, the same quantization coefficients (default values) as those in the decoding are used.
  • Then, the [0051] image encoder 16 performs quantization based on the set quantization coefficients on the image data held in the image decoder 14, thus performing compression processing (step S1007).
  • In this arrangement, a face area with great significance is re-compressed without degrading image quality, while an area other than the face area such as background with less significance is compressed by a greater quantization step. Thus the total code amount can be reduced. [0052]
  • The [0053] multipoint connection device 22 performs the above processings independently on data transferred from the respective terminals 21 a to 21 n. Note that the re-compressed image data are multiplexed with the audio compressed data by the multiplexing/demultiplexing units 12 a to 12 n (step S1008), and delivered to the respective terminals 21 a to 21 n in accordance with a communication protocol.
  • FIG. 5 is an example of displayed image, obtained by e.g. a camera (not shown) of the terminal [0054] 21 d, then processed by the multipoint connection device 22 as described above, and delivered to the respective terminals 21 a to 21 n, in the video conference system of the present embodiment. In FIG. 5, an area in a circle 51 is recognized as a face area; and an area 52 is a background area recognized as a non-face area. The background area 52, quantized by a greater quantization step, becomes a low-resolution image, while the face area 51 has image quality equivalent to that of an image obtained by the terminal 21 d.
  • As described above, according to the present embodiment, in a video conference system, only a significant portion of an image obtained by a terminal can be transmitted with high precision to the respective terminals of video conference system. Accordingly, even if the respective terminals are connected by a narrow band communication channel, the significant portion can be transmitted in high-quality image communication. At the respective terminals, in an image transferred from another terminal, a significant portion, face area, i.e., the facial expression of the user of the other terminal can be clearly displayed without reduction or frame thinning. [0055]
  • Further, in response to an instruction from a terminal, an image from another arbitrary terminal may be received as an entirely high quality image at any time. [0056]
  • Note that the decoding processing and the encoding processing in the present embodiment may be performed in predetermined image data block units. In this case, the above-described processings at steps S[0057] 1003 to S1007 are repeated in each block.
  • Further, an area where high image quality is maintained is not limited to a face area. The same advantage as that of the present embodiment can be obtained by setting an arbitrary significant object as an object of high-precision display in accordance with the content of conference, and transferring other image areas with degraded image quality. [0058]
  • Second Embodiment
  • Next, a second embodiment of the present invention will be described. [0059]
  • As the basic construction of the video conference system of the second embodiment is similar to that of the first embodiment as shown in FIG. 1, the explanation of the construction will be omitted. [0060]
  • FIG. 6 is a block diagram showing the construction of the [0061] multipoint connection device 22 according to the second embodiment of the present invention. In FIG. 6, constituents corresponding to those in FIG. 2 have the same reference numerals and the explanations of these constituents will be omitted.
  • In FIG. 6, numeral [0062] 86 denotes a spatial filter processor which performs low-frequency filter processing on decoded image data based on the result of recognition by the face-area recognition unit 15. More specifically, the spatial filter processor 86 performs spatial filter processing on areas other than a face area by a low-pass filter, to attenuate high frequency components.
  • Next, the image processing by the [0063] multipoint connection device 22 according to the second embodiment will be described in detail with reference to the flowchart of FIG. 7. Note that steps corresponding to those in FIG. 4 have the same step numbers, and the explanations of those steps will be omitted.
  • As in the case of the first embodiment, after the face area recognition at step S[0064] 1004, the decoded image data is re-compressed by the image encoder 16. In the present embodiment, upon this compression processing, spatial filter processing is performed based on the result of face area recognition.
  • The [0065] spatial filter processor 86 inputs decoded image data and the result of face area recognition. In the spatial filter processor 86, a low-pass filter is prepared in advance, and spatial filter processing by the low-pass filter is performed (step S1106) on an area determined at step S1005 as a non-face area. Then, the spatial filter processed image data is compressed by the image encoder 16 (step S1008).
  • FIG. 8 is an example of the low-pass filter in the [0066] spatial filter processor 86. The filter processing expressed by the following equation (1) is performed by using the low-pass filter.
  • P′(x,y)=(P(x−1,y−1)+P(x,y−1)+P(x+1,y−1 )+P(x−1,y)+P(x,y)+(x+1,y)+P(x−1,y+1)+P(x,y+1)+P(x+1,y+1))/9   (1)
  • Note that P(x,y) is a pixel value of coordinates (x,y). [0067]
  • Note that the spatial filter in the second embodiment is not limited to that as shown in FIG. 8, but any filter may be employed as long as it is a low-pass filter. [0068]
  • Further, in the second embodiment, execution/non-execution of spatial filter processing is controlled in accordance with whether or not an object area is a face area (specific area), however, the present invention is not limited to this arrangement. For example, spatial filter processing may be performed on a specific area and other areas. In this case, a low-pass filer used in the processing is selected in accordance with whether an object area is a specific area or not. A filter for the other areas than the face area is selected so as to remove high frequency components more than those in the face area removed by a filter for the face area. [0069]
  • As described above, according to the second embodiment, in image data, a specific area with great significance such as a face area is re-compressed without degrading image quality, and in areas other than the specific area with less significance such as background, high frequency components are attenuated. This reduces the total code amount, and accordingly, obtains the same advantage as that of the first embodiment. [0070]
  • Further, in the second embodiment, as the resolution of image is controlled by the spatial filter processing, degradation of image quality is visually more inconspicuous in comparison with the first embodiment where the code amount is reduced by control on quantization coefficients. [0071]
  • In the above-described first and second embodiments, in each of the images obtained by the respective video conference terminals, especially a face area is detected. In a case where the terminals respectively have a fixed camera which obtains an image with a fixed image sensing area, if the compression rate for a part around the center of the obtained image is set to a lower rate than the compression rate for other parts in advance, the object of the present invention can be attained without detecting a face area from the image. [0072]
  • However, in a case where a plurality of persons are image-sensed by such fixed camera, in an image from a terminal, the user of the terminal is not always positioned around the center of the image. Accordingly, in this case, if the image transfer is performed such that only a part around the center of the obtained image has high image quality, the transferred image is unnatural. For example, if only the half of the face of the user is included in a preset low compression rate area but the other half is included in a high compression rate area, it is impossible to transfer the image such that the entire face area has high image quality. [0073]
  • Accordingly, even if the terminals respectively have a fixed camera, to ensure high-quality face area transfer, a face area is extracted as a significant portion from an obtained image, as in the case of the above-described first and second embodiments. [0074]
  • Further, in a case where the terminals respectively have a camera in which its image sensing direction can be operated from another terminal, the camera of a terminal may be operated such that the user's face comes to the center of the obtained image. However, if the orientation of the camera can be arbitrarily operated from the opposite-end terminal side, an undesirable portion such as background other than a person's face might be arbitrarily image-sensed. [0075]
  • Accordingly, in case of image sensing by such remote-controllable camera, detecting a face area is rather advantageous than previously setting low compression rate area and high compression rate area within the obtained image. The face area detection unnecessitates operation of the orientation of the camera, and further ensures privacy protection for conference participants. [0076]
  • Third Embodiment
  • Next, a third embodiment of the present invention will be described. The basic construction of the video conference system of the third embodiment is similar to that of the first embodiment as shown in FIG. 1, therefore, the explanation of the construction will be omitted. [0077]
  • In the third embodiment, the principal processings of the [0078] multipoint connection device 22 having the functions as described in the first embodiment are realized by software.
  • FIG. 9 is a block diagram showing the construction of the [0079] multipoint connection device 22 according to the third embodiment of the present invention. Numeral 61 denotes a communication interface unit for communication with an ISDN, a PSTN and the like; and 62, a media processor which performs principal signal processings. The media processor 62 performs various digital signal processings by a specialized processor for multimedia signal processing, a DSP (Digital Signal Processor) and the like. In the third embodiment, the video conference system can be realized with a simple and low-cost construction without preparing dedicated hardware, by processing various data by multitask processing by utilizing the media processor 62.
  • [0080] Numeral 63 denotes a RAM (Random Access Memory) which is used as a work memory necessary for the operation of the media processor 62 and as a buffer memory for various image and audio data to be processed. The RAM 63 is realized by, e.g., a high-speed large-capacity memory such as an SDRAM (Synchronous DRAM). Numeral 64 denotes a CPU which performs system control for the multipoint connection device 22, communication protocol processing and the like; 65, a RAM used as a work buffer necessary for the operation of the CPU 64; and 66, a ROM (Read Only Memory) in which a control program necessary for the operation of the CPU 64 and a control program necessary for the operation of the media processor 62 are stored. The program necessary for the operation of the media processor 62 is transferred from the ROM 66 to the RAM 63 upon system initialization.
  • In the third embodiment, the above construction realizes processing as shown in the flowchart of FIG. 4. [0081]
  • That is, a control program to realize the processing is stored in the [0082] ROM 66 in advance. The program is developed on the RAM 63 or the RAM 65, and executed by the media processor 62 or the CPU 64. For example, processing related to data compression/decompression is performed by the media processor 62, and other processings are controlled by the CPU 64.
  • According to the third embodiment, the same advantages as those in the first embodiment can be obtained by realizing the processings described in the first embodiment by software. [0083]
  • Note that in the third embodiment, the [0084] multipoint connection device 22 as described in the first embodiment is realized by software. Similarly, the multipoint connection device 22 as described in the second embodiment may be realized by software.
  • Note that the compression method employed in the above-described first to third embodiments may be any method. For example, an H. 261 coding method or the like based on DCT (Discrete Cosine Transformation) used in the Recommendation H.320 may be used. In such case, the control on the compression method (quantization coefficients) based on the result of specific-area recognition is performed in macroblock units. [0085]
  • Further, the compression control method of the present invention is not limited to the quantization control or spatial filter control as described in the first to third embodiments, but any control method may be performed as long as it controls the degree of image quality degradation due to re-compression in accordance with whether or not an object area is a specific area. In other words, any control method may be used as long as it controls the compression rate in re-compression. [0086]
  • In addition, more flexible image quality control including further improvement in image quality of delivered image data can be realized by combining the quantization control described in the first embodiment and the spatial filter control described in the second embodiment. [0087]
  • Further, in the first and second embodiments, the video conference terminals are directly connected to a public line such as PSTN/ISDN line, however, the present invention is not limited to this example. The video conference terminals may be connected to the Internet, an Intranet or the like. In this case, as shown in FIG. 12, communication is performed with a plurality of terminals by the Internet Protocol via a [0088] network access controller 121 having a LAN interface such as the Ethernet. The means for communication with the respective terminals is logically constructed for each terminal. Control on packetized data and data multiplexing/demultiplexing processing are performed by the network access controller 121 and a system controller 122.
  • Further, the video conference terminals connected to the multipoint connection device of the present invention may be dedicated terminals or general-purpose terminals such as personal computers where a predetermined application (software) program is operated. [0089]
  • Further, the multipoint connection device of the present invention may have an automatic HTML (Hyper Text Markup Language) data generating function and an HTTP (Hyper Text Transfer Protocol) server function in addition to the function related to multipoint terminal connection. In this case, general-purpose communication terminals, which display and visualize image and audio data stored in the HTTP server of the multipoint connection device, can be connected to the multipoint connection device. As the general-purpose communication terminals, personal computers having a so-called WWW-browser and network computers can be employed. Further, by virtue of recent development of semiconductor technologies, these computers may be employed as quick and convenient terminals even in a mobile communication environment. These general-purpose communication terminals are IP (Internet Protocol)-connected to the multipoint connection terminal via a public network such as an ISDN or a PSTN or a LAN (Local Area Network). This enables interconnection among the existing video conference terminals and the general-purpose terminals such as personal computers and network computers. [0090]
  • Other Embodiment
  • The present invention can be applied to a system constituted by a plurality of devices (e.g., a host computer, an interface, a reader and a printer) or to an apparatus comprising a single device (e.g., a copy machine or a facsimile apparatus). [0091]
  • Further, the object of the present invention can be also achieved by providing a storage medium storing program code for performing the aforesaid processes to a system or an apparatus, reading the program code with a computer (e.g., CPU, MPU) of the system or apparatus from the storage medium, then executing the program. [0092]
  • In this case, the program code read from the storage medium realize the functions according to the embodiments, and the storage medium storing the program code constitutes the invention. [0093]
  • Further, the storage medium, such as a floppy disk, a hard disk, an optical disk, a magneto-optical disk, CD-ROM, CD-R, a magnetic tape, a non-volatile type memory card, and ROM can be used for providing the program code. [0094]
  • Furthermore, besides aforesaid functions according to the above embodiments are realized by executing the program code which are read by a computer, the present invention includes a case where an OS (operating system) or the like working on the computer performs a part or entire processes in accordance with designations of the program code and realizes functions according to the above embodiments. [0095]
  • Further, a product obtained by the image processing method of the present invention, e.g., printed matter, is included in the present invention. [0096]
  • Furthermore, the present invention also includes a case where, after the program code read from the storage medium are written in a function expansion card which is inserted into the computer or in a memory provided in a function expansion unit which is connected to the computer, CPU or the like contained in the function expansion card or unit performs a part or entire process in accordance with designations of the program code and realizes functions of the above embodiments. In a case where the present invention is applied to the aforesaid storage medium, the storage medium stores program code corresponding to the flowcharts shown in FIGS. 4 and 7 described in the embodiments. [0097]
  • As described above, according to the present invention, in the video conference system in a narrow band communication channel, facial expressions of the participants of a conference can be clearly displayed. [0098]
  • Further, upon re-compression on decompressed image data, degradation of image quality can be reduced. [0099]
  • As many apparently widely different embodiments of the present invention can be made without departing from the spirit and scope thereof, it is to be understood that the invention is not limited to the specific embodiments thereof except as defined in the appended claims. [0100]

Claims (8)

1-48. (canceled)
49. A data communication control method for controlling data communication among a plurality of terminals connected to a multipoint connection device, the method comprising the steps of:
transferring image data from one of the plurality of terminals;
recognizing a specific area in the transferred image data at the multipoint connection device;
controlling the compression method for compressing the received image data at the multipoint connection device to compress the specific area and not reduce a burden of the plurality of terminals; and
transmitting the compressed image data from the multipoint connection device to at least one of the plurality of terminals.
50. The method according to claim 49, wherein the transferred image data is determined whether the specific area is compressed or not, according to a designation of at least one of the plurality of terminals.
51. The method according to claim 49, wherein the specific area is an area in which a person's face is displayed.
52. A multipoint connection device for controlling data communication among a plurality of terminals, comprising:
an interface unit which receives image data from one of the plurality of terminals;
a recognition unit which recognizes a specific area in the received data; and
a controller which controls the compression method for compressing the received image data to compress the specific area and not reduce a burden of the plurality of terminals,
wherein the interface unit transmits the compressed image data to at least one of the plurality of terminals.
53. The device according to claim 52, wherein the received image data is determined whether the specific area is compressed or not, according to a designation of at least one of the plurality of terminals.
54. The device according to claim 52, wherein the specific area is an area in which a person's face is displayed.
55. A computer readable medium containing control program code in the multipoint connection device for controlling data communication among the plurality of terminals, the control program code being implemented by a method comprising the steps of:
receiving image data from one of the plurality of terminals;
recognizing a specific area in the received image data;
controlling the compression method for compressing the received image data to compress the specific area and not reduce a burden of the plurality of terminals; and
transmitting the compressed image data from the multipoint connection device to at least one of the plurality of terminals.
US10/852,159 1998-12-25 2004-05-25 Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system Abandoned US20040213474A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US10/852,159 US20040213474A1 (en) 1998-12-25 2004-05-25 Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP10-371478 1998-12-25
JP37147898A JP4208315B2 (en) 1998-12-25 1998-12-25 DATA COMMUNICATION CONTROL DEVICE AND ITS CONTROL METHOD, DATA COMMUNICATION SYSTEM, RECORDING MEDIUM
US09/469,798 US6744927B1 (en) 1998-12-25 1999-12-22 Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system
US10/852,159 US20040213474A1 (en) 1998-12-25 2004-05-25 Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US09/469,798 Division US6744927B1 (en) 1998-12-25 1999-12-22 Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system

Publications (1)

Publication Number Publication Date
US20040213474A1 true US20040213474A1 (en) 2004-10-28

Family

ID=18498779

Family Applications (2)

Application Number Title Priority Date Filing Date
US09/469,798 Expired - Fee Related US6744927B1 (en) 1998-12-25 1999-12-22 Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system
US10/852,159 Abandoned US20040213474A1 (en) 1998-12-25 2004-05-25 Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US09/469,798 Expired - Fee Related US6744927B1 (en) 1998-12-25 1999-12-22 Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system

Country Status (2)

Country Link
US (2) US6744927B1 (en)
JP (1) JP4208315B2 (en)

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050213738A1 (en) * 2001-12-31 2005-09-29 Polycom, Inc. Conference endpoint requesting and receiving billing information from a conference bridge
US20050213728A1 (en) * 2001-12-31 2005-09-29 Polycom, Inc. Conference endpoint instructing a remote device to establish a new connection
US20050213736A1 (en) * 2001-12-31 2005-09-29 Polycom, Inc. Speakerphone establishing and using a second connection of graphics information
US20060009254A1 (en) * 2004-07-09 2006-01-12 Masayuki Kanazawa Radio communication system
US20070047626A1 (en) * 2005-06-08 2007-03-01 Polycom, Inc Mixed voice and spread spectrum data signaling with multiplexing multiple users with cdma
US7787605B2 (en) * 2001-12-31 2010-08-31 Polycom, Inc. Conference bridge which decodes and responds to control information embedded in audio information
US7864938B2 (en) 2000-12-26 2011-01-04 Polycom, Inc. Speakerphone transmitting URL information to a remote device
US7978838B2 (en) 2001-12-31 2011-07-12 Polycom, Inc. Conference endpoint instructing conference bridge to mute participants
US8004556B2 (en) 2004-04-16 2011-08-23 Polycom, Inc. Conference link between a speakerphone and a video conference unit
US8102984B2 (en) 2001-12-31 2012-01-24 Polycom Inc. Speakerphone and conference bridge which receive and provide participant monitoring information
US8126029B2 (en) 2005-06-08 2012-02-28 Polycom, Inc. Voice interference correction for mixed voice and spread spectrum data signaling
US8144854B2 (en) 2001-12-31 2012-03-27 Polycom Inc. Conference bridge which detects control information embedded in audio information to prioritize operations
US8199791B2 (en) 2005-06-08 2012-06-12 Polycom, Inc. Mixed voice and spread spectrum data signaling with enhanced concealment of data
US8705719B2 (en) 2001-12-31 2014-04-22 Polycom, Inc. Speakerphone and conference bridge which receive and provide participant monitoring information
US8805928B2 (en) 2001-05-10 2014-08-12 Polycom, Inc. Control unit for multipoint multimedia/audio system
US8885523B2 (en) 2001-12-31 2014-11-11 Polycom, Inc. Speakerphone transmitting control information embedded in audio information through a conference bridge
US8934382B2 (en) 2001-05-10 2015-01-13 Polycom, Inc. Conference endpoint controlling functions of a remote device
US8947487B2 (en) 2001-12-31 2015-02-03 Polycom, Inc. Method and apparatus for combining speakerphone and video conference unit operations
US8948059B2 (en) 2000-12-26 2015-02-03 Polycom, Inc. Conference endpoint controlling audio volume of a remote device
US8964604B2 (en) 2000-12-26 2015-02-24 Polycom, Inc. Conference endpoint instructing conference bridge to dial phone number
US8977683B2 (en) 2000-12-26 2015-03-10 Polycom, Inc. Speakerphone transmitting password information to a remote device
US8976712B2 (en) 2001-05-10 2015-03-10 Polycom, Inc. Speakerphone and conference bridge which request and perform polling operations
US9001702B2 (en) 2000-12-26 2015-04-07 Polycom, Inc. Speakerphone using a secure audio connection to initiate a second secure connection

Families Citing this family (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE10138267A1 (en) * 2001-08-03 2003-02-20 Siemens Ag Method and terminal device for multipoint communication
JP2003208085A (en) * 2002-01-10 2003-07-25 Nova:Kk Method for providing educational service providing free talking service, device and program for providing educational service
JP4015934B2 (en) * 2002-04-18 2007-11-28 株式会社東芝 Video coding method and apparatus
AU2002239208A1 (en) * 2002-03-22 2003-10-08 Telefonaktiebolaget Lm Ericsson (Publ) Reducing transmission time for data packets controlled by a link layer protocol comprising a fragmenting/defragmenting capability
JP3716225B2 (en) * 2002-04-05 2005-11-16 松下電器産業株式会社 Code amount control apparatus and code amount control method
JP3999091B2 (en) * 2002-09-25 2007-10-31 富士フイルム株式会社 Image correction processing apparatus and program
KR100539923B1 (en) * 2003-02-10 2005-12-28 삼성전자주식회사 A video encoder capable of encoding deferentially as distinguishing image of user and method for compressing a video signal using that
JP3787633B2 (en) * 2003-04-21 2006-06-21 独立行政法人情報通信研究機構 Real-time content editing system
US20050008240A1 (en) * 2003-05-02 2005-01-13 Ashish Banerji Stitching of video for continuous presence multipoint video conferencing
JP4346966B2 (en) * 2003-06-13 2009-10-21 キヤノン株式会社 Imaging device
CN100423539C (en) * 2004-02-23 2008-10-01 Toa株式会社 Image compression method, image compression device, image transmission system, data compression pre-processing device, and computer program
JP2005346202A (en) * 2004-05-31 2005-12-15 Toshiba Corp Electronic apparatus
US7539341B2 (en) * 2004-07-29 2009-05-26 Xerox Corporation Systems and methods for processing image data prior to compression
US7620892B2 (en) * 2004-07-29 2009-11-17 Xerox Corporation Server based image processing for client display of documents
US20060106929A1 (en) * 2004-10-15 2006-05-18 Kenoyer Michael L Network conference communications
US7706576B1 (en) * 2004-12-28 2010-04-27 Avaya Inc. Dynamic video equalization of images using face-tracking
US7460150B1 (en) 2005-03-14 2008-12-02 Avaya Inc. Using gaze detection to determine an area of interest within a scene
WO2006106032A1 (en) * 2005-04-05 2006-10-12 Thomson Licensing Method for locally adjusting a quantization step
US7564476B1 (en) 2005-05-13 2009-07-21 Avaya Inc. Prevent video calls based on appearance
WO2007063922A1 (en) * 2005-11-29 2007-06-07 Kyocera Corporation Communication terminal and communication system, and display method of communication terminal
US8812978B2 (en) * 2005-12-22 2014-08-19 Xerox Corporation System and method for dynamic zoom to view documents on small displays
US7489772B2 (en) 2005-12-30 2009-02-10 Nokia Corporation Network entity, method and computer program product for effectuating a conference session
US8165282B1 (en) 2006-05-25 2012-04-24 Avaya Inc. Exploiting facial characteristics for improved agent selection
JP5181294B2 (en) * 2008-03-31 2013-04-10 富士フイルム株式会社 Imaging system, imaging method, and program
JP4961582B2 (en) * 2008-04-07 2012-06-27 富士フイルム株式会社 Image processing system, image processing method, and program
US8913668B2 (en) * 2008-09-29 2014-12-16 Microsoft Corporation Perceptual mechanism for the selection of residues in video coders
US8457194B2 (en) * 2008-09-29 2013-06-04 Microsoft Corporation Processing real-time video
CN101742324A (en) * 2008-11-14 2010-06-16 北京中星微电子有限公司 Video encoding and decoding methods, video encoding and decoding systems and encoder-decoder
US20100188476A1 (en) * 2009-01-29 2010-07-29 Optical Fusion Inc. Image Quality of Video Conferences
JP2014060577A (en) * 2012-09-18 2014-04-03 Sony Corp Communication terminal, program and information processing device
US11798204B2 (en) * 2022-03-02 2023-10-24 Qualcomm Incorporated Systems and methods of image processing based on gaze detection

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4951140A (en) * 1988-02-22 1990-08-21 Kabushiki Kaisha Toshiba Image encoding apparatus
US5675374A (en) * 1993-11-26 1997-10-07 Fujitsu Limited Video teleconferencing system
US5991445A (en) * 1994-06-27 1999-11-23 Canon Kabushiki Kaisha Image processing apparatus
US6112233A (en) * 1998-02-04 2000-08-29 Intel Corporation Saving and retrieving a message via a network server using an index page
US6181711B1 (en) * 1997-06-26 2001-01-30 Cisco Systems, Inc. System and method for transporting a compressed video and data bit stream over a communication channel
US6285661B1 (en) * 1998-01-28 2001-09-04 Picturetel Corporation Low delay real time digital video mixing for multipoint video conferencing
US6343313B1 (en) * 1996-03-26 2002-01-29 Pixion, Inc. Computer conferencing system with real-time multipoint, multi-speed, multi-stream scalability
US6351568B1 (en) * 1998-03-20 2002-02-26 Canon Kabushiki Kaisha Image transform and significance bit-plane compression and decompression
US6456335B1 (en) * 1998-02-19 2002-09-24 Fujitsu Limited Multiple picture composing method and multiple picture composing apparatus

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4951140A (en) * 1988-02-22 1990-08-21 Kabushiki Kaisha Toshiba Image encoding apparatus
US5675374A (en) * 1993-11-26 1997-10-07 Fujitsu Limited Video teleconferencing system
US5991445A (en) * 1994-06-27 1999-11-23 Canon Kabushiki Kaisha Image processing apparatus
US6343313B1 (en) * 1996-03-26 2002-01-29 Pixion, Inc. Computer conferencing system with real-time multipoint, multi-speed, multi-stream scalability
US6181711B1 (en) * 1997-06-26 2001-01-30 Cisco Systems, Inc. System and method for transporting a compressed video and data bit stream over a communication channel
US6285661B1 (en) * 1998-01-28 2001-09-04 Picturetel Corporation Low delay real time digital video mixing for multipoint video conferencing
US6112233A (en) * 1998-02-04 2000-08-29 Intel Corporation Saving and retrieving a message via a network server using an index page
US6456335B1 (en) * 1998-02-19 2002-09-24 Fujitsu Limited Multiple picture composing method and multiple picture composing apparatus
US6351568B1 (en) * 1998-03-20 2002-02-26 Canon Kabushiki Kaisha Image transform and significance bit-plane compression and decompression

Cited By (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7864938B2 (en) 2000-12-26 2011-01-04 Polycom, Inc. Speakerphone transmitting URL information to a remote device
US9001702B2 (en) 2000-12-26 2015-04-07 Polycom, Inc. Speakerphone using a secure audio connection to initiate a second secure connection
US8977683B2 (en) 2000-12-26 2015-03-10 Polycom, Inc. Speakerphone transmitting password information to a remote device
US8964604B2 (en) 2000-12-26 2015-02-24 Polycom, Inc. Conference endpoint instructing conference bridge to dial phone number
US8948059B2 (en) 2000-12-26 2015-02-03 Polycom, Inc. Conference endpoint controlling audio volume of a remote device
US8976712B2 (en) 2001-05-10 2015-03-10 Polycom, Inc. Speakerphone and conference bridge which request and perform polling operations
US8934382B2 (en) 2001-05-10 2015-01-13 Polycom, Inc. Conference endpoint controlling functions of a remote device
US8805928B2 (en) 2001-05-10 2014-08-12 Polycom, Inc. Control unit for multipoint multimedia/audio system
US7787605B2 (en) * 2001-12-31 2010-08-31 Polycom, Inc. Conference bridge which decodes and responds to control information embedded in audio information
US7742588B2 (en) 2001-12-31 2010-06-22 Polycom, Inc. Speakerphone establishing and using a second connection of graphics information
US7978838B2 (en) 2001-12-31 2011-07-12 Polycom, Inc. Conference endpoint instructing conference bridge to mute participants
US20050213728A1 (en) * 2001-12-31 2005-09-29 Polycom, Inc. Conference endpoint instructing a remote device to establish a new connection
US8102984B2 (en) 2001-12-31 2012-01-24 Polycom Inc. Speakerphone and conference bridge which receive and provide participant monitoring information
US20050213736A1 (en) * 2001-12-31 2005-09-29 Polycom, Inc. Speakerphone establishing and using a second connection of graphics information
US8144854B2 (en) 2001-12-31 2012-03-27 Polycom Inc. Conference bridge which detects control information embedded in audio information to prioritize operations
US8947487B2 (en) 2001-12-31 2015-02-03 Polycom, Inc. Method and apparatus for combining speakerphone and video conference unit operations
US8223942B2 (en) 2001-12-31 2012-07-17 Polycom, Inc. Conference endpoint requesting and receiving billing information from a conference bridge
US8705719B2 (en) 2001-12-31 2014-04-22 Polycom, Inc. Speakerphone and conference bridge which receive and provide participant monitoring information
US20050213738A1 (en) * 2001-12-31 2005-09-29 Polycom, Inc. Conference endpoint requesting and receiving billing information from a conference bridge
US8885523B2 (en) 2001-12-31 2014-11-11 Polycom, Inc. Speakerphone transmitting control information embedded in audio information through a conference bridge
US8934381B2 (en) 2001-12-31 2015-01-13 Polycom, Inc. Conference endpoint instructing a remote device to establish a new connection
US8004556B2 (en) 2004-04-16 2011-08-23 Polycom, Inc. Conference link between a speakerphone and a video conference unit
US7580685B2 (en) * 2004-07-09 2009-08-25 Hitachi Kokusai Electric, Inc. Radio communication system
US20060009254A1 (en) * 2004-07-09 2006-01-12 Masayuki Kanazawa Radio communication system
US7796565B2 (en) 2005-06-08 2010-09-14 Polycom, Inc. Mixed voice and spread spectrum data signaling with multiplexing multiple users with CDMA
US8199791B2 (en) 2005-06-08 2012-06-12 Polycom, Inc. Mixed voice and spread spectrum data signaling with enhanced concealment of data
US20070047626A1 (en) * 2005-06-08 2007-03-01 Polycom, Inc Mixed voice and spread spectrum data signaling with multiplexing multiple users with cdma
US8126029B2 (en) 2005-06-08 2012-02-28 Polycom, Inc. Voice interference correction for mixed voice and spread spectrum data signaling

Also Published As

Publication number Publication date
JP2000197043A (en) 2000-07-14
JP4208315B2 (en) 2009-01-14
US6744927B1 (en) 2004-06-01

Similar Documents

Publication Publication Date Title
US6744927B1 (en) Data communication control apparatus and its control method, image processing apparatus and its method, and data communication system
US7039675B1 (en) Data communication control apparatus and method adapted to control distribution of data corresponding to various types of a plurality of terminals
US6989868B2 (en) Method of converting format of encoded video data and apparatus therefor
US6593955B1 (en) Video telephony system
US5634040A (en) Data communication apparatus and method having concurrent image overlay function
US7006575B2 (en) Picture and sound decoding apparatus picture and sound encoding apparatus and information transmission system
US6091857A (en) System for producing a quantized signal
US6909810B2 (en) Image processing apparatus and method, and its computer program and storage medium
US5577191A (en) System and method for digital video editing and publishing, using intraframe-only video data in intermediate steps
US20030220971A1 (en) Method and apparatus for video conferencing with audio redirection within a 360 degree view
US6084911A (en) Transmission of coded and compressed voice and image data in fixed bit length data packets
JP2001514826A (en) Method and apparatus for transmitting and displaying still images
JP2007166625A (en) Video data encoder, video data encoding method, video data decoder, and video data decoding method
JPH10150663A (en) Method and device for minimizing blocking effect
US20040001091A1 (en) Method and apparatus for video conferencing system with 360 degree view
EP0805600A2 (en) Compressed video text overlay
US6859557B1 (en) System and method for selective decoding and decompression
US20010012397A1 (en) Image processing apparatus and method
JP2002262288A (en) Image data transmitter and image data receiver
KR100312421B1 (en) A conversion method of the compressed moving video on the video communication system
US8086056B2 (en) Encoding device and method, decoding device and method, and program
JPH08186814A (en) Image compressor
US6462783B1 (en) Picture encoding method and apparatus
JPH09116759A (en) Image decoder and image coding decoding system
KR20020070721A (en) Streaming device for moving picture

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION