CA2565755A1 - Information processing apparatus and control method thereof - Google Patents

Information processing apparatus and control method thereof Download PDF

Info

Publication number
CA2565755A1
CA2565755A1 CA002565755A CA2565755A CA2565755A1 CA 2565755 A1 CA2565755 A1 CA 2565755A1 CA 002565755 A CA002565755 A CA 002565755A CA 2565755 A CA2565755 A CA 2565755A CA 2565755 A1 CA2565755 A1 CA 2565755A1
Authority
CA
Canada
Prior art keywords
speech
output
image data
moving image
unit
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
CA002565755A
Other languages
French (fr)
Inventor
Takeshi Makita
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Toshiba Corp
Original Assignee
Toshiba Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Toshiba Corp filed Critical Toshiba Corp
Publication of CA2565755A1 publication Critical patent/CA2565755A1/en
Abandoned legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/44Receiver circuitry for the reception of television signals according to analogue transmission standards
    • H04N5/60Receiver circuitry for the reception of television signals according to analogue transmission standards for the sound signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • H04N21/4316Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations for displaying supplemental content in a region of the screen, e.g. an advertisement in a separate window
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/44Receiver circuitry for the reception of television signals according to analogue transmission standards
    • H04N5/445Receiver circuitry for the reception of television signals according to analogue transmission standards for displaying additional information
    • H04N5/45Picture in picture, e.g. displaying simultaneously another television channel in a region of the screen

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Business, Economics & Management (AREA)
  • Marketing (AREA)
  • General Engineering & Computer Science (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Telephonic Communication Services (AREA)

Abstract

When moving image data and speech data received via a communications device are displayed on a display unit, speech data corresponding to the moving image data are appropriately distributed to output units and loudspeakers and then output in accordance with a position of the displayed moving image data.

Description

TITLE OF THE INVENTION

INFORMATION PROCESSING APPARATUS AND CONTROL METHOD
THEREOF

BACKGROUND OF THE INVENTION

This invention relates to a video conference system and, more particularly, to an information processing apparatus and a control method thereof, capable of improving a sense of realism in speech of a speaker by emphasizing speech from a loudspeaker on a monitor side on which a speaker is displayed.

In a video conference system as disclosed in Jpn.
Pat. Appln. KOKAI Publication No. 9-307869, for example, a main participant, of plural participants, is displayed and emphasized.

According to this technique, however, the speaker's speech is not considered, and it is often difficult to discriminate which speaker has made the speech output from a loudspeaker.

BRIEF SUMMARY OF THE INVENTION

The present invention has been accomplished to solve the above-described problem. The object of the present invention is to provide an information processing apparatus and a control method thereof, capable of emphasizing a speech output from a loudspeaker on the basis of a displayed position of the speaker.

To achieve this object, an aspect of the present invention is an information processing apparatus comprising communications means, display means, a plurality of speech output means, acquisition means, and distribution means. The acquisition means acquires a plurality of moving image data items and speech data items received via the communications means. When the plurality of moving image data items acquired by the acquisition means are displayed by the display means, the distribution means appropriately distributes the speech data item corresponding to each of the displayed moving image data items to the plurality of speech output means and allows the speech output means to output the speech data item in accordance with a position of the displayed moving image data item.

According to an aspect of the present invention, there is provided an information processing apparatus and a control method thereof, capable of emphasizing the speech output from the loudspeaker on the basis of the displayed position of the speaker.

Additional objects and advantages of the invention will be set forth in the description which follows, and in part will be obvious from the description, or may be learned by practice of the invention. The objects and advantages of the invention may be realized and obtained by means of the instrumentalities and combinations particularly pointed out hereinafter.
BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWING
The accompanying drawings, which are incorporated in and constitute a part of the specification, illustrate embodiments of the invention, and together with the general description given above and the detailed description of the embodiments given below, serve to explain the principles of the invention.

FIG. 1 shows an illustration of a configuration of a video conference system to which an information processing apparatus according to a first embodiment of the present invention is applied;

FIG. 2 shows an illustration of displaying image data on a display;

FIG. 3 shows an illustration of displaying image data on a display;

FIG. 4 shows a flowchart of steps in a control method to which the information processing apparatus of the present invention is applied;

FIG. 5 shows an illustration of displaying image data on a display, according to a modified example of the first embodiment;

FIG. 6 shows an illustration of displaying image data on a display, according to a modified example of the first embodiment; and FIG. 7 shows an illustration of a configuration of a video conference system to which an information processing apparatus according to a second embodiment of the present invention is applied.

DETAILED DESCRIPTION OF THE INVENTION
Embodiments of the present invention will be explained below with reference to the accompanying drawings.

(First Embodiment) FIG. 1 shows an illustration of a configuration of a video conference system to which an information processing apparatus according to a first embodiment of the present invention is applied.

The video conference system comprises terminal apparatuses 12a to 12d, a WAN/LAN 11, and a server 10 which synthesizes data received from the terminal apparatuses 12a to 12d and distributes the synthesized data to each of the terminal apparatuses 12a to 12d via the WAN/LAN 11.

The terminal apparatuses 12a to 12d have the same structure. For example, the terminal apparatus 12a comprises a camera 23 which inputs images, a microphone 24 which inputs speech, a data controller 22 which receives data from the camera 23 and the microphone 24 and converts the received data into communications data or processes data received from the server 10, a display unit 26 which reproduces image data (moving image data and audio data), a loudspeaker 25 which reproduces audio data, and a communications device 21 which receives communications data from the server 10.
FIG. 2 and FIG. 3 show illustrations of displaying image data items 26a to 26d on the display unit 26.
FIG. 4 shows a flowchart of steps in a control method to which the information processing apparatus of the present invention is applied.

First, the terminal apparatus 12a acquires the image data (moving image data and audio data) received via the communications device 21 and displays the image data 26a to 26d on the display unit 26.

The terminal apparatus 12a discriminates whether or not the speaker is on the left side of the display screen (step S1). Since the display screen 26a of the speaker is on the left side (YES of step S1) as shown in FIG. 2, the terminal apparatus 12a discriminates whether or not the speaker is on the lower side of the display screen (step S2). As the display screen 26a of the speaker is on the upper side (NO of step S2), speech of, for example, 90dB SPL is output from an upper output unit of a left speaker 25a and is not output from the other speaker output units (step S3).
Next, the terminal apparatus 12a discriminates whether or not the speaker has changed (step S5). If the speaker is on the display screen 26d as shown in, for example, FIG. 3, the terminal apparatus 12a discriminates that the speaker has changed (YES of step S5) and the operation returns to step S1 to output the speech from an appropriate output unit of the loudspeaker.

On the other hand, if the terminal apparatus 12a discriminates that the speaker has not changed (NO of step S5), the terminal apparatus 12a discriminates that speaking has not been further conducted and the video conference is ended (step S6).

If it is discriminated at step S2 that the display screen of the speaker is on the lower side (YES of step S2), speech of, for example, 90dB SPL is output from a lower output unit of the left speaker 25a and is not output from the other speaker output units (step S4).
If it is discriminated at step Si that the display screen of the speaker is on the right side (NO of step S1), it is discriminated whether or not the speaker is on the lower side of the display screen (step S7). If it is discriminated that the speaker is on the lower side of the display screen (YES of step S7), speech of, for example, 90dB SPL is output from a lower output unit of a right speaker 25b and is not output from the other speaker output units (step S9).

On the other hand, if it is discriminated that the speaker is on the upper side of the display screen (NO
of step S7), speech of, for example, 90dB SPL is output from an upper output unit of a right speaker 25b and is not output from the other speaker output units (step S8).

As for the speech output value distribution of the loudspeaker 25, an output of, for example, 10dB SPL
that is clearly smaller than the output of 90dB SPL
from the output unit of the main speaker which outputs the speech may be output from the output units of the other speaker.

Thus, the video conference system rich in a sense of realism, capable of executing the processing of emphasizing the speech output from the loudspeaker can be executed on the basis of the display position of the speaker, and capable of outputting the speech in accordance with the displayed position of the speaker, can be constructed.

(Modified Example of the First Embodiment) Next, a modified example of the first embodiment will be described with reference to FIG. 5 and FIG. 6.
The modified example of the first embodiment has a characteristic of setting, for example, nine display screens of the speaker on the display unit.

The display screens of the speaker synchronize with the speaker output units, similarly to the first embodiment. For example, as shown in FIG. 5, if the display screen of the speaker is a display screen 26g, speech of, for example, 90dB SPL is output from the lower output unit of the left speaker 25a and is not output from the other speaker output units, since the display screen 26g is on the lower left side of the display unit 26.
In addition, for example, as shown in FIG. 6, if the display screen of the speaker is a display screen 26f, speech of, for example, 90dB SPL is output from both the output units of the right speaker 25b and is not output from the other speaker output units, since the display screen 26f is on the central right side of the display unit 26.

The number of display screens to be displayed on the display unit 26 is not limited to the above-described embodiments if the output speech appropriately synchronizes with the display screens of the speaker.

Therefore, even if the number of display screens to be displayed on the display unit is increased, the output speech can appropriately synchronize with the display screens of the speaker.

(Second Embodiment) FIG. 7 shows an illustration of display screens in a video conference system to which an information processing apparatus according to the second embodiment of the present invention is applied.

In the second embodiment, the speech is also output appropriately in a case where the display screen of the speaker is moved by an input device such as a mouse, remote controller, etc.

For example, movement of the display screen 26a to the lower right side as shown in FIG. 7 will be described. It can be understood that the moved display screen 26a is moved by 31 to the right side and by 32 to the lower side from the initial position.

The rate of lateral movement of the display screen 26a and the output distribution from the output units of the speakers can be obtained by calculating the balance ratio in the lateral direction.

Since the lateral distance between the display screen 26a and the display screen 26b is, for example, al, the moved display screen 26a is located at a position of 0 1 : a 1- 0 1 in the lateral direction.
The output distribution of the speech output of the left loudspeaker 25a and the right loudspeaker 25b is thereby set at 81 : a 1- (3 1.

The rate of longitudinal movement of the display screen 26a can be obtained by calculating the longitudinal balance ratio. Since the longitudinal distance between the display screen 26a and the display screen 26c is, for example, a2, the moved display screen 26a is located at a position of /3 2: a 2- /3 2 in the longitudinal direction. The output distribution of the speech output of the upper and lower output units in each of the loudspeaker 25a and the loudspeaker 25b is thereby set at 82 : a 2- a 2.

Then, the output distribution is determined in the following manner.

If the display unit 26 is shaped in a square, al = a2. In addition, the numerical values are assumed as follows.

al = a2 = 100 cm (31 = 40 cm (32 = 30 cm Thus, the distribution of the left and right speech outputs is 81 : al - 31 = 40 : 60 and the distribution of the upper and lower speech outputs is ~2 : a2 - 82 = 30 : 70 Therefore, the output distribution of the output units of the loudspeakers is Upper output unit of the right loudspeaker 25a = about 12 dB SPL

Lower output unit of the right loudspeaker 25a = about 28 dB SPL

Upper output unit of the left loudspeaker 25b = about 18 dB SPL

Lower output unit of the left loudspeaker 25b = about 42 dB SPL

In the above-described embodiments, the number of loudspeakers is two and the number of output units in each loudspeaker is two. However, the number of loudspeakers and the number of output units in each loudspeaker are not limited to those if the output speech appropriately synchronizes with the display screen of the speaker.

As a result, if the display screen of the speaker is moved on the display unit, the speech can be output in synchronization with the moved display screen.

The present invention is not limited to the embodiments described above but the constituent elements of the invention can be modified in various manners without departing from the spirit and scope of the invention. Various aspects of the invention can also be extracted from any appropriate combination of a plurality of constituent elements disclosed in the embodiments. For example, some constituent elements may be deleted in all of the constituent elements disclosed in the embodiments. The constituent elements described in different embodiments may be combined arbitrarily.

Additional advantages and modifications will readily occur to those skilled in the art. Therefore, the invention in its broader aspects is not limited to the specific details and representative embodiments shown and described herein. Accordingly, various modifications may be made without departing from the spirit or scope of the general inventive concept as defined by the appended claims and their equivalents.

Claims (10)

1. An information processing apparatus, comprising:

communications unit;
display unit;

a plurality of speech output unit; and acquiring a plurality of moving image data items and speech data items received via the communications unit and when the plurality of moving image data items acquired by the acquisition unit are displayed by the display unit, appropriately distributing the speech data item corresponding to each of the displayed moving image data items to the plurality of speech output unit and allowing the speech output unit to output the speech data item in accordance with a position of the displayed moving image data item.
2. The apparatus according to claim 1, wherein the plurality of speech output unit are loudspeakers arranged at predetermined positions and speech is output with emphasis from at least one of the loudspeakers close to the position of the displayed moving image data item.
3. The apparatus according to claim 2, wherein each of the speakers has a plurality of speech output units and the speech is output with emphasis from at least one of the speech output units of the loudspeakers close to the position of the displayed moving image data item.
4. The apparatus according to claim 1, wherein the plurality of speech output unit are loudspeakers arranged at predetermined positions and speech is output only from at least one of the loudspeakers close to the position of the displayed moving image data item.
5. The apparatus according to claim 4, wherein each of the speakers has a plurality of speech output units and the speech is output only from at least one of the speech output units of the loudspeakers close to the position of the displayed moving image data item.
6. The apparatus according to claim 1, wherein when a display range of the moving image data item displayed on the display unit is moved, the speech data item is appropriately redistributed to the plurality of speech output unit and output in response to the moved position in the display unit.
7. The apparatus according to claim 1, wherein when the speech data items corresponding to not less than two moving image data items, of the plurality of moving image data items, exist simultaneously, the speech data items are output simultaneously.
8. A method of controlling an information processing apparatus comprising communications unit, display unit, and a plurality of speech output unit, acquiring a plurality of moving image data items and speech data items received via the communications unit and when the plurality of moving image data items acquired by the acquisition unit are displayed by the display unit, appropriately distributing the speech data item corresponding to each of the displayed moving image data items to the plurality of speech output unit and allowing the speech output unit to output the speech data item in accordance with a position of the displayed moving image data item.
9. The method according to claim 8, wherein the plurality of speech output unit are loudspeakers arranged at predetermined positions and speech is output with emphasis from at least one of the loudspeakers close to the position of the displayed moving image data item.
10. The method according to claim 9, wherein each of the speakers has a plurality of speech output units and the speech is output with emphasis from at least one of the speech output units of the loudspeakers close to the position of the displayed moving image data item.
CA002565755A 2005-10-27 2006-10-26 Information processing apparatus and control method thereof Abandoned CA2565755A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2005-313299 2005-10-27
JP2005313299A JP2007124253A (en) 2005-10-27 2005-10-27 Information processor and control method

Publications (1)

Publication Number Publication Date
CA2565755A1 true CA2565755A1 (en) 2007-04-27

Family

ID=37965238

Family Applications (1)

Application Number Title Priority Date Filing Date
CA002565755A Abandoned CA2565755A1 (en) 2005-10-27 2006-10-26 Information processing apparatus and control method thereof

Country Status (3)

Country Link
US (1) US20070097222A1 (en)
JP (1) JP2007124253A (en)
CA (1) CA2565755A1 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TW200822745A (en) * 2006-11-01 2008-05-16 Inventec Corp Electronic device
CN102209225B (en) * 2010-03-30 2013-04-17 华为终端有限公司 Method and device for realizing video communication
JP6062803B2 (en) * 2013-05-29 2017-01-18 京セラ株式会社 Communication device and audio output change method

Also Published As

Publication number Publication date
US20070097222A1 (en) 2007-05-03
JP2007124253A (en) 2007-05-17

Similar Documents

Publication Publication Date Title
US5548346A (en) Apparatus for integrally controlling audio and video signals in real time and multi-site communication control method
US5808663A (en) Multimedia carousel for video conferencing and multimedia presentation applications
CN101459797B (en) Sound positioning method and system
EP2342894A1 (en) Controlling a video window position relative to a video camera position
JP2006041887A (en) Information processing apparatus and method, recording medium, and program
US20090094375A1 (en) Method And System For Presenting An Event Using An Electronic Device
JP3036088B2 (en) Sound signal output method for displaying multiple image windows
JP2013062640A (en) Signal processor, signal processing method, and program
JPH08163522A (en) Video conference system and terminal equipment
JP4644555B2 (en) Video / audio synthesizer and remote experience sharing type video viewing system
CN102119531A (en) Audio/video system
CA2565755A1 (en) Information processing apparatus and control method thereof
JP4720321B2 (en) Information processing apparatus and method, and program
JP4839782B2 (en) Output control system and method, output control apparatus and method, and program
WO2018198790A1 (en) Communication device, communication method, program, and telepresence system
JP2001268078A (en) Communication controller, its method, providing medium and communication equipment
JP2897627B2 (en) Conference environment control device
JP2023505986A (en) Multiple output control based on user input
JPH0758859A (en) Information transmitter and information receiver for conference
KR101488936B1 (en) Apparatus and method for adjusting middle layer
JP2011066745A (en) Terminal apparatus, communication method and communication system
JPH0463084A (en) Television conference system among scattered sites
US10020903B2 (en) Method, device, and non-transitory computer-readable recording medium for supporting relay broadcasting using mobile device
JPH01206765A (en) Video conference system
JPH08125760A (en) Information processor

Legal Events

Date Code Title Description
EEER Examination request
FZDE Discontinued