WO2006106671A1 - Image processing device, image display device, reception device, transmission device, communication system, image processing method, image processing program, and recording medium containing the image processing program - Google Patents

Image processing device, image display device, reception device, transmission device, communication system, image processing method, image processing program, and recording medium containing the image processing program Download PDF

Info

Publication number
WO2006106671A1
WO2006106671A1 PCT/JP2006/306297 JP2006306297W WO2006106671A1 WO 2006106671 A1 WO2006106671 A1 WO 2006106671A1 JP 2006306297 W JP2006306297 W JP 2006306297W WO 2006106671 A1 WO2006106671 A1 WO 2006106671A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
information
voice
image processing
display
Prior art date
Application number
PCT/JP2006/306297
Other languages
French (fr)
Japanese (ja)
Inventor
Toshiharu Baba
Original Assignee
Pioneer Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Pioneer Corporation filed Critical Pioneer Corporation
Publication of WO2006106671A1 publication Critical patent/WO2006106671A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • G06T13/203D [Three Dimensional] animation
    • G06T13/2053D [Three Dimensional] animation driven by audio data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • G06T13/203D [Three Dimensional] animation
    • G06T13/403D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/06Transformation of speech into a non-audible representation, e.g. speech visualisation or speech processing for tactile aids

Definitions

  • Image processing apparatus image display apparatus, receiving apparatus, transmitting apparatus, communication system, image processing method, image processing program, and recording medium recording image processing program
  • the present invention relates to an image processing device that displays an image, an image display device, a receiving device, a transmitting device, a communication system, an image processing method, an image processing program, and a recording medium that records the image processing program.
  • Patent Document 1 Japanese Patent Application Laid-Open No. 2004-15158 (refer to pages 4 to 5 and FIGS. 3 to 4) Disclosure of the Invention
  • the present invention relates to an image processing device, an image display device, a receiving device, a transmitting device, a communication system, an image processing method, an image processing program, and an image processing program capable of displaying an appropriate image.
  • One objective is to provide a recorded recording medium.
  • the image processing apparatus of the present invention is an image processing apparatus that processes an image displayed on a display unit according to audio information obtained by reception, and recognizes a change in an audio state in the audio information.
  • a display control unit configured to display the image on the display unit and to change the displayed image in accordance with the change in the audio state.
  • the image display device of the present invention comprises storage means for storing an image, display means for displaying the image when the audio information is received, and the image processing device of the present invention described above. It is characterized by that.
  • a receiving device of the present invention includes the above-described image display device of the present invention and a receiving unit capable of receiving the audio information, and the display control unit receives the audio information by the receiving unit. In this case, an image corresponding to the transmission source of the audio information is displayed.
  • a transmission device of the present invention includes the above-described image display device of the present invention and a transmission unit capable of transmitting and receiving the audio signal, wherein the transmission unit is for a transmission destination that transmits the audio information.
  • the display unit displays an image corresponding to the transmission destination in response to a call made by the calling unit or a response of the transmission destination to the call, and the image processing
  • the apparatus is characterized in that the displayed image is changed in accordance with a change in a voice state in voice information from a transmission destination received in response to the call.
  • the communication system of the present invention is a communication system including a transmission / reception terminal capable of transmitting and receiving audio information to and from each other, and each of the transmission / reception terminals transmits storage means for storing images and audio information.
  • the image processing method of the present invention provides a sound obtained by receiving an image displayed on the display means.
  • An image processing method for processing according to information comprising: recognizing a change in sound state in the sound information, displaying the image on the display means, and displaying the displayed image according to the change in sound state. It is characterized by being changed.
  • An image processing program of the present invention is characterized in that a computing means functions as the above-described image processing apparatus of the present invention.
  • An image processing program of the present invention is characterized by causing a calculation means to perform the above-described image processing method of the present invention.
  • a recording medium on which the image processing program of the present invention is recorded is characterized in that the above-described image processing program of the present invention is recorded so as to be readable by an arithmetic means.
  • FIG. 1 is an overall perspective view of a portable telephone device according to an embodiment of the present invention.
  • FIG. 2 is a block diagram schematically showing the internal configuration of the portable telephone device.
  • FIG. 3 is a schematic diagram showing an outline of a registration information table recorded in storage means.
  • FIG. 4 is a schematic diagram showing an image of registered image information recorded in a storage means.
  • FIG. 5 is a schematic diagram showing a schematic configuration of a first image template table recorded in a storage means.
  • FIG. 6 is a schematic diagram showing a schematic configuration of a second image template table recorded in a storage means.
  • FIG. 7 is a schematic diagram showing a schematic configuration of an emotion recognition information table recorded in a storage means.
  • FIG. 8 is a block diagram schematically showing a configuration of a processing unit.
  • FIG. 9A is a schematic diagram showing an emotion deformation image transformed by the image processing means.
  • FIG. 9B is a schematic diagram showing another emotion deformation image transformed by the image processing means.
  • FIG. 9C is a schematic diagram showing another emotion deformation image transformed by the image processing means.
  • FIG. 10 is a flowchart showing a photographing process of a mobile phone.
  • FIG. 11 is a flowchart of incoming call processing when a mobile phone is incoming.
  • FIG. 12 is a flowchart of a calling process when a mobile phone is called.
  • FIG. 13A Eyebrow image of the first image basic information in a modification of the mobile phone of the present embodiment It is a schematic diagram which shows an example of the eyebrow image image
  • FIG. 13B is a schematic diagram showing an example of an eyebrow image taken when a person in the eyebrow image of the first image basic information is “angry” in a modification of the mobile phone of the present embodiment.
  • FIG. 13C is a schematic diagram showing an example of an eyebrow image captured when a person in the eyebrow image of the first image basic information is in a sad state in a modification of the mobile phone according to the present embodiment.
  • FIG. 14A is a schematic diagram showing an example of a mouth image taken when a person in the mouth image of the second image basic information is in a smiling state in a modification of the mobile phone according to the present embodiment.
  • FIG. 14B is a schematic diagram showing an example of a mouth image taken when a person in the mouth image of the second image basic information is “angry”.
  • FIG. 14C is a schematic diagram showing an example of a mouth image taken when a person in the mouth image of the second image basic information is in a sad state.
  • Image that also functions as 185 image information recognition means and image deformation processing means
  • FIG. 1 is an overall perspective view of a portable telephone device according to an embodiment of the present invention.
  • FIG. 2 is a block diagram schematically showing the internal configuration of the portable telephone device.
  • FIG. 3 is a schematic diagram showing an outline of a registration information table recorded in the storage device.
  • FIG. 4 is a schematic diagram showing an image of registered image information recorded in the storage means.
  • FIG. 5 is a schematic diagram showing a schematic configuration of the first image template table recorded in the storage means.
  • FIG. 6 is a schematic diagram showing a schematic configuration of the second image template table recorded in the storage means.
  • FIG. 7 is a schematic diagram showing a schematic configuration of an emotion recognition information table recorded in the storage means.
  • FIG. 8 is a block diagram schematically showing the configuration of the processing unit.
  • FIG. 1 is an overall perspective view of a portable telephone device according to an embodiment of the present invention.
  • FIG. 2 is a block diagram schematically showing the internal configuration of the portable telephone device.
  • FIG. 3 is a schematic diagram showing an outline
  • FIG. 9A is a schematic diagram showing an emotion-transformed image transformed by the image processing means.
  • FIG. 9B is a schematic diagram showing another emotion deformation image transformed by the image processing means.
  • FIG. 9C is a schematic diagram showing another emotion deformation image transformed by the image processing means.
  • reference numeral 100 denotes a portable telephone device (hereinafter referred to as a mobile phone) that also functions as an image display device, a receiving device, a transmitting device, and a transmitting / receiving terminal.
  • the 100 communicates with a communication device (not shown) as a transmitting device such as another mobile phone or a general telephone device via a network such as a telephone line or an Internet line, and is connected so as to be able to make a call with this communication device. To do. Then, a predetermined image is displayed on the display 110 according to the communication partner.
  • the mobile phone 100 is used as the image display device.
  • the present invention is not limited to this.
  • an image display unit such as a monitor that processes an image with a personal computer. It may be a device that is displayed on the screen.
  • the cellular phone 100 is configured to include an upper casing 100A and a lower casing 100B that can accommodate a circuit board and the like. Further, the upper casing 100A is rotatably provided to the lower casing 100B by a rotating portion 100C. Note that FIG. 1 illustrates an example in which the upper casing 100A is pivotably attached to the lower casing 100B. For example, the upper casing 100A and the lower casing 100B are combined. It may be something that is formed.
  • the upper casing 100A is provided with a display 110, an audio output unit 120, a receiving unit having an antenna 130A (see Fig. 1) and a transmitting / receiving unit 130 (see Fig. 2) as a transmitting unit.
  • the housing 100B is provided with an operation unit 140, an audio input unit 150, a storage unit 160, a memory 170, a processing unit 180 that also functions as an image processing device, and the like.
  • the circuit board provided in the upper casing 100A and connected to the display 110, the audio output unit 120, and the transmission / reception unit 130 is connected to the circuit board to which the processing unit 180 in the lower casing 100B is connected. They are electrically connected by a flexible board inserted through the inside of the rotating part 100C.
  • the upper casing 100A is partially formed with a window portion that communicates the inside and the outside, and the camera portion 101 for taking a picture is attached to the window portion so as to face the outside.
  • the camera unit 101 may be provided on the surface of the upper housing 100A opposite to the side on which the display 110 faces, or on another position.
  • the display 110 includes a display area in which various information such as predetermined image information and text information is displayed under the control of the processing unit 180.
  • This image information includes, for example, image information recorded in the storage means 160, TV image data received by a TV receiver (not shown), recording media such as an external device such as an optical disk, a magnetic disk, and a memory card. Image data recorded and read by a drive or driver, or image data from the memory 170.
  • Examples of the display 110 include a liquid crystal display panel, an organic EL (Electro Luminescence), a PDP (Plasma Display Panel), a CRT (and atnode-Ray Tube), an FED (Field Emission Display), and an electrophoretic display panel. Can be illustrated
  • the sound output unit 120 converts predetermined sound information into sound and outputs the sound under the control of the processing unit 180.
  • voice information for example, voice information related to the voice of the communication partner speaker, music information recorded in the storage means 160 or the like, a warning sound stored in the memory 170 or the like can be exemplified.
  • the audio output unit 120 shown in FIG. 1 has a speaker that can be switched between directivity and omnidirectionality by the user's operation. For example, when the user uses the mobile phone 100 without using it, In addition, the sound output from the sound output unit 120 is configured to be audible. For example, a separate omnidirectional speaker may be provided on the side surface of the lower housing 100B opposite to the surface on which the operation unit 140 is provided. Good.
  • the transceiver 130 includes an antenna 130A. Further, the transmission / reception unit 130 controls the voice information of the user input from the voice input unit 150 through the antenna 130A, for example, by the control of the processing unit 180. Send to a communication device such as a telephone device. Also, audio information transmitted from the other party's communication device or the like is received via the antenna 130A and output to the processing unit 180. Further, the transmission / reception unit 130 receives information from, for example, another sano on the network and outputs the information to the processing unit 180 under the control of the processing unit 180. Further, the transmission / reception unit 130 transmits information input from the processing unit 180 to a predetermined server or terminal device under the control of the processing unit 180.
  • the operation unit 140 has various operation buttons and operation knobs.
  • the contents of the input operation of these operation buttons and control knobs include, for example, input of the other party's number when making a call to the other party's mobile phone, operation for browsing the registered information of the other party, and start / end operation of the call A setting for acquiring information from another server can be exemplified.
  • the operation unit 140 appropriately outputs a predetermined signal to the processing unit 180 by a user input operation.
  • the audio input unit 150 is provided on the lower housing 100B side and includes a microphone capable of inputting audio.
  • the voice input unit 150 can switch between directivity and non-directionality by operating the operation unit 140.
  • the voice input unit 150 can be switched from the voice input unit 150 when the user uses the cellular phone 100 without using the mobile phone 100. It is configured to be able to input voice.
  • the storage means 160 includes, for example, the registration information table 10 as shown in FIG. 3, the first image template table 20 as shown in FIG. 5, the second image template table 30 as shown in FIG.
  • the emotion recognition information table 40 shown in FIG. 7 is stored, that is, stored so as to be readable.
  • the storage means 160 stores a registration information storage area in which the registration information table 10 is recorded, a first image storage area in which the first image template table 20 is recorded, and a second image template table 30 in which the second image template table 30 is stored. 2 image storage areas and an emotion recognition information storage area in which the emotion recognition information table 40 is stored.
  • the storage means 160 may be configured to include a recording area for recording other information in addition to these four storage areas.
  • a registration information storage area, a first image storage area, a second image storage area, an emotion recognition information storage area, and the like are stored in the memory 170 in the storage means. It is good also as a structure.
  • the storage unit 160 a configuration including a drive or a driver that can be read and stored in a recording medium such as an HD (Hard Disk) or a memory card can be illustrated.
  • a configuration including a DVD (Digital Versatile Disc), an optical disc, or the like may be used.
  • the registration information table 10 is information used when, for example, the user confirms the information of the other speaker when the user makes a call with the other speaker.
  • This registration information table 10 includes registration ID information 12, registrant name information 13, telephone number information 14, registration image information 15 as image information, registration detailed information 16, and the like, in association with one piece of data. It is configured in a table structure in which a plurality of registration information 11 configured as follows is recorded.
  • the registered HD information 12 is unique information for specifying the registered information 11 and is set for each registered information 11. For example, when the user registers the registration information 11 in the registration information table 10, the registration I blueprint 12 is automatically given as a serial number.
  • the registrant name information 13 is information relating to the name of the registrant registered as the registration information 11 specified by the registration HD information 12.
  • the registrant name information 13 is recorded in a text format, for example.
  • the telephone number information 14 is information related to the telephone number of the registrant's mobile phone registered as the registration information 11 specified by the registration HD information 12, or the telephone number of a general telephone.
  • the telephone number information 14 is information that is referred to when, for example, an incoming call is received from the registrant to the user's mobile phone 100 or when the user makes a call to the registrant.
  • the registered image information 15 is registered as the registered information 11 specified by the registered HD information 12.
  • the registered image information 15 is transmitted to the registrant.
  • This is image information related to an image to be displayed on the display 110 at this time.
  • the image of the registered image information 15 is an image obtained by capturing a front face image of a registrant, for example, so as to be stored in a predetermined frame.
  • a registered image 50 of the registered image information 15 is image information taken in a state where the registrant's face is housed in a frame 51.
  • the frame 51 is divided by five substantially parallel frame lines 52, and the registrant's face image is arranged so that the eyebrows, eyes, nose, and mouth are arranged in the divided areas.
  • the center line force of frame 51 in frame area A where the eyebrows are placed is also in the horizontal direction.
  • a region surrounded by a position where the central line force of the frame 51 in the frame region D where the mouth is disposed is also moved by a predetermined dimension in the left-right direction is a mouth region 55 where the mouth is disposed.
  • the frame 51 and the frame line 52 are illustrated, but these are supplementarily displayed when the registrant is photographed by the camera unit 101 of the mobile phone 100, for example. Therefore, the frame 51 and the frame line 52 are not displayed in the registered image 50 of the registered image information 15 that is actually registered.
  • a registered image 50 may be, for example, a photograph taken with a camera unit 101 for photographing provided in the mobile phone 100 and registered with a commercially available digital camera device. Even if you have registered! /.
  • the registration detail information 16 is information relating to the detailed items of the registrant registered as the registration information 11 specified by the registration HD information 12.
  • the detailed information recorded in the registration details information 16 includes, for example, information such as the registrant's address and work place, gender of the registrant, information on the action when there is an incoming call from the registrant, and the email address of the registrant. And basic voice information on the registrant's voice.
  • the first image template table 20 records the deformation patterns of the eyebrow images in the right eyebrow region 53 and the left eyebrow region 54 when the processing unit 180 deforms and edits the registered image 50 recorded in the registered image information 15. Information.
  • This first image template table 20 corresponds to the first image HD information 22, the first image basic information 23, the first image pattern information 24 as deformation amount information and deformation direction information, as shown in FIG. It is constructed in a table structure in which a plurality of first image template information 21 in which emotion information 25 and the like are associated are recorded.
  • the first image HD information 22 is unique information that identifies the first image template information 21.
  • the first image HD information 22 is unique information different from each first image template information 21, and for example, a serial number is recorded.
  • the first image basic information 23 is image information that is the basis of the first image template information 21 specified by the first image HD information 22.
  • a basic image of normal eyebrows is recorded.
  • the basic image of the first image basic information is a rectangular shape that is the shape of the right eyebrow region 53 and the left eyebrow region 54 of the registered image 50 as shown in FIG.
  • a rectangular shape having the same size as the shape or a rectangular shape having a similar relationship is formed, and an image of a normal eyebrow, for example, is arranged in the center of the rectangle.
  • the first image pattern information 24 is based on the basic image of the first image basic information 23 of the first image template information 21 specified by the first image information 22, and is in a predetermined emotional state.
  • a deformation direction and a deformation rate as a deformation amount indicating how much and how much each part of the eyebrow image of the first image basic information 23 is deformed are recorded.
  • the first image pattern information 24 moves the approximate center position of the eyebrows upward, for example, 10 dots, and both ends of the eyebrows downward, for example, 20 Information such as dot movement is recorded.
  • the ratio of moving each part of the eyebrow upward and the ratio moving downward with respect to the rectangular vertical dimension of the right eyebrow region 53 and the left eyebrow region 54 is recorded.
  • the amount of deformation and the direction of deformation may be recorded in vector format, and may be.
  • Corresponding emotion information 25 is information indicating a change rate indicating what kind of emotion the first image pattern information 24 of the first image template information 21 specified by the first image HD information 22 is. is there.
  • the corresponding emotion information 25 includes, for example, “0” indicating “normal” or “no expression”, “1” indicating “smile”, “2” indicating anger, etc. It is recorded as a numerical value. Information about these emotions may be recorded in a text format such as “normal”, “smile”, “anger”, etc.! /, Etc.
  • the second image template table 30 is information in which a deformation pattern of the mouth image of the mouth region 55 when the registered image 50 recorded in the registered image information 15 is modified and edited by the processing unit 180 is recorded. As shown in FIG. 6, the second image template table 30 includes second image information 32, second image basic information 33, second image notation information 34 as deformation amount information and deformation direction information, and corresponding emotions. It is constructed in a table structure that records a plurality of second image template information 31 that associates information 35 and the like.
  • the second image HD information 32 is unique information that identifies the second image template information 31.
  • the second image HD information 32 is unique information different from each of the second image template information 31, and for example, a serial number is recorded.
  • the second image basic information 33 is a second image template specified by the second image HD information 32. This is image information that is the basis of information 31.
  • a basic image of the mouth at normal times is recorded. This basic image is formed in a rectangular shape that is substantially the same size as the rectangular shape of the mouth area 55 of the registered image shown in FIG. 4 or a similar rectangular shape.
  • the mouth image is arranged in the approximate center.
  • the second image pattern information 34 is based on the second image basic information 33 of the second image template information 31 specified by the second image information 32, and the second image pattern information 34 in a predetermined emotional state.
  • the deformation direction and the rate of change as the amount of deformation indicating how much each part of the mouth image of the basic information 33 is moved in which direction are recorded. For example, in the case of a smiling face, if both ends of the mouth image are moved upward, for example, 20 dots, the information is recorded.
  • the rate of change may be, for example, the rate of moving each part of the mouth upward, the rate of moving downward, the rate of moving left or right, etc. May be recorded! For example, it may be recorded in vector format! /.
  • Corresponding emotion information 35 is similar to the first image template information 21, and the second image pattern information 34 of the second image template information 31 specified by the second image HD information 32 indicates what kind of emotion. It is the information which shows whether it is the change rate to show.
  • the corresponding emotion information 35 includes, for example, “0” indicating “normal” or “no expression”, “1” indicating “smile”, “2” indicating anger, etc. It may be recorded in a text format which may be recorded as a numerical value.
  • the emotion recognition information table 40 is used when the processing unit 180 recognizes the emotion of the speaker who generated the voice from the state of the voice information of the other speaker, for example, the strength or tempo of the voice. Data group.
  • This emotion recognition information table 40 records a plurality of emotion recognition information 41 constructed as one piece of data by associating emotion information 42, voice pattern information 43 as voice state change information, emotion information 44, etc.
  • the table structure is configured.
  • the emotion information 42 is unique information that identifies the emotion recognition information 41. This emotion information
  • the voice pattern information 43 is information relating to the voice state of the voice. Specifically, this The voice pattern information 43 includes phrase information related to emotional phrases such as “Kora” and “Hahaha”, strength information related to the strength of the voice, height information related to the voice high / low state, and tempo related to the voice tempo. Information is recorded.
  • the emotion information 44 is information related to the emotion corresponding to the voice pattern information 43 of the emotion recognition information 41 specified by the emotion I and the blueprint 42.
  • this emotion information 44 for example, if the phrase “hahaha” is recorded as the phrase information in which the voice strength is strong in the voice pattern information 43, for example, HD indicating that the voice is laughing loudly. Information such as “1” is recorded.
  • the emotion information 44 may be recorded in a text format such as “smile” or “anger”.
  • the emotion information 44 is associated with the corresponding emotion information 25 of the first image template information 21 and the corresponding emotion information 35 of the second image template information 31 described above.
  • the memory 170 stores setting items input and operated by the operation unit 140 so that they can be read as appropriate.
  • the memory 170 stores various programs developed on an OS (Operating System) that controls the operation of the entire mobile phone 100.
  • the memory 170 may be configured to include a drive and a driver that are readable and stored in a recording medium such as an HD (Hard Disk) or a magneto-optical disk.
  • the processing unit 180 includes various input / output ports (not shown), such as a display control port to which the display 110 is connected, an audio output control port to which the audio output unit 120 is connected, a transmission / reception port to which the transmission / reception unit 130 is connected, An input port to which the operation unit 140 is connected, a voice input control port to which the voice input unit 150 is connected, a storage port to which the storage unit 160 is connected, a memory port to which the memory 170 is connected, and the like. Then, the processing unit 180, as various programs, as shown in FIG. 8, an incoming / outgoing recognition means 181 that also functions as a calling means, a partner speaker recognition means 182 and a voice recognition means as voice information recognition means. 183, sound change determination means 184 as sound state recognition means, image information recognition means, image processing means 185 that also functions as image deformation processing means, display control means 186, photographing means 187, etc. It is equipped with.
  • the arrival / departure recognition means 181 recognizes incoming calls from other communication devices to the mobile phone 100 and arrivals / departures from the mobile phone 100 to other communication devices. Specifically, the arrival and departure recognition means 181 Controls the transmission / reception unit 130 to receive incoming call information requesting a call from another communication device. In addition, when the arrival / departure recognition means 181 recognizes the incoming call information, it controls the audio output unit 120 to output the incoming sound such as voice, warning sound, notification sound, etc. to notify the user that the incoming call information has been received. .
  • the arrival / departure recognition means 181 recognizes request information indicating that a call is to be made to a predetermined destination communication device by an input operation of the operation unit 140 of the user, the transmission information indicating that a call is requested to the destination communication device. Send.
  • the called party recognition means 182 recognizes the incoming call partner and the outgoing call partner.
  • the partner speaker recognition means 182 recognizes the telephone number of the called party who has sent the incoming information from the incoming information received by the arrival / reception recognition means 181. Then, the registration information 11 having the telephone number information 14 that matches the recognized telephone number of the called party is recognized from the registration information table 10. Furthermore, the other speaker recognition means 182 recognizes the destination telephone number of the transmission information transmitted by the arrival / reception recognition means 181. Then, the registration information 11 having the telephone number information 14 that matches the destination telephone number is recognized from the registration information table 10. Then, the partner speaker recognition means 182 stores the recognized registration information 11 in the memory 170 so that it can be read out appropriately.
  • the voice recognition means 183 also recognizes the voice status of the other speaker as received by the transmitting / receiving unit 130, and recognizes the voice status of the other speaker as voice basic information as standard voice status information of the other speaker. . Specifically, the voice recognition means 183 recognizes the voice information of the partner speaker input via the transmission / reception unit 130. Then, the voice pattern of the other speaker is determined from the voice state of the voice information, that is, the voice level, the voice strength, the tempo at which the voice is emitted, and the like are recognized as the voice basic information. The voice recognition means 183 may be configured to recognize the other speaker's gender, basic voice information, etc. as voice basic information based on the contents described in the registration details information 16 of the registration information 11. .
  • the voice change determination means 184 changes the voice when the voice of the voice information received by the transmission / reception unit 130 changes from the voice basic information of the other speaker recognized by the voice recognition means 183.
  • the voice change determination unit 184 determines the voice parameter of the voice state. Detect turn. For example, if the detected voice pattern is different in height, strength, tempo, etc. compared to the voice pattern of the basic voice information, the voice pattern of the received voice is changed based on the emotion recognition information table 40.
  • the voice pattern information 42 that substantially matches is recognized, and the emotion recognition information 41 corresponding to the voice pattern information 42 is recognized.
  • the voice change determination means 184 recognizes that the voice of the received voice information includes phrase information representing the emotion, such as “Kora”, “Hahaha”, etc., these phrases information is displayed.
  • the emotion recognition information 41 included in the voice pattern information 43 is recognized.
  • the voice change determination means 184 stores the emotion recognition information 41 including the recognized emotion information 44 in the memory 170 so that it can be read out as appropriate.
  • the image processing means 185 recognizes the registered image information 15 of the registered information 11 corresponding to the partner speaker. Then, based on the emotion information 44 of the emotion recognition information 41 recognized by the voice change determination means 184, the image of the registered image information 15 is changed. Specifically, the image processing means 185 recognizes the right eyebrow region 53 and the left eyebrow region 54 of the registered image 50 recorded in the registered image information 15. Then, based on the first image template table 20, the image processing means 185 has a first image having first image basic information 23 of an image that substantially matches the eyebrow images of the right eyebrow region 53 and the left eyebrow region 54. Search for template information 21.
  • the eyebrow image recorded in the first image basic information 23 is overlaid on the right eyebrow region 53 and the left eyebrow region 54 of the registered image information 15, for example, the first image basics with the highest degree of overlap of the eyebrow portions.
  • Search for information 23 the image processing means 185 includes a first image having corresponding emotion information 25 corresponding to the emotion information 44 recognized by the voice change determination means 184 from the first image template information 21 narrowed down by searching. Recognizes template information 21.
  • the shape of the eyebrows of the right eyebrow region 53 and the left eyebrow region 54 of the registered image 50 of the registered image information 15 according to the change rate recorded in the first image pattern information 24 of the first image template information 21 Transform
  • the image processing means 185 also deforms the mouth image in the same manner as the deformation process of the eyebrow image in the right eyebrow region 53 and the left eyebrow region 54. That is, the image processing means 185 recognizes the mouth area 55 of the registered image 50. Then, based on the second image template table 30, the second image template having the second image basic information 33 of the image that substantially matches the mouth image in the region 55. Search for port information 31. To do this, for example, the mouth image recorded in the second image basic information 33 is overlaid on the mouth area 55 of the registered image information 15, and for example, the second image basic information 33 with the largest degree of overlap of the mouth portion is searched. To do.
  • the image processing means 185 includes a second image having corresponding emotion information 35 corresponding to the emotion information 44 recognized by the voice change determination means 184 from the second image template information 31 searched and narrowed down. Recognize template information 31. Then, the mouth shape of the mouth region 55 of the registered image 50 of the registered image information 15 is changed according to the change rate recorded in the second image pattern information 34 of the second image template information 31. In addition, the image processing means 185 stores the registered image 50 in which the eyebrows and the mouth are deformed in the memory 170 so as to be appropriately readable as an emotion deformed image.
  • the display control means 186 controls the display 110 to display the image of the registered image information 15 of the registration information 11 in the display area. Specifically, the display control means 186, when the arrival / reception recognition means 181 recognizes the call request information indicating that the call is sent to the communication apparatus of the other speaker by the operation of the operation unit 140 by the user, Control is performed to display the registered image 50 recorded in the registered image information 15 on the display 110. In addition, the display control means 186 recognizes that the mobile phone 100 has an incoming communication device such as another mobile phone or a general telephone at the arrival / departure recognition means 181 and the destination speaker recognition means 182 is the transmission destination. When the registered image information 15 of the registered information 11 of the other speaker is recognized, control is performed to display the registered image 50 of the registered image information 15 on the display 110.
  • the display control means 186 performs control to display the emotion deformed image processed by the image processing means 185 in the display area of the display 110 as shown in FIG. 9A or FIG. 9C, for example.
  • the voice change determination means 184 determines that the voice state of the other party's voice information is laughing, for example, and the image processing means 185 is registered according to the emotion information 44.
  • the image 50 is a deformed image.
  • the voice change determination means 184 determines that the voice pattern of the voice information of the other party's call is “angry”, for example, and performs image processing according to the emotion information 44.
  • Means 185 is an image obtained by deforming the registered image 50.
  • FIG. 9A the voice change determination means 184 determines that the voice state of the other party's voice information is laughing, for example, and the image processing means 185 is registered according to the emotion information 44.
  • the image 50 is a deformed image.
  • the voice change determination means 184 determines that the voice pattern of the voice information of the other party'
  • the display control means 186 controls the video input from the camera unit 101 as image information and displays it on the display area of the display 110.
  • the image capturing unit 187 When the image capturing unit 187 recognizes request information for capturing an image by an input operation of the operation unit 140 by the user, the image capturing unit 187 controls the camera unit 101 to be in a state in which an image can be captured. In addition, the display 110 is controlled to display an image of the shooting range of the camera unit 101. Furthermore, when the image capturing unit 187 recognizes request information indicating that an image that can be processed by the image processing unit 185 is input by a user's setting input, the frame 51 and the frame as illustrated in FIG. Control the display of line 52 on display 110. When the request information indicating that the image is to be captured is recognized by the user's input operation, the image within the imaging range of the camera unit 101 is captured and stored in the storage unit 160 as image information.
  • FIG. 10 is a flowchart showing the photographing process of the mobile phone 100.
  • the processing unit 180 of the mobile phone 100 receives the shooting request information for requesting shooting of a predetermined video by the camera unit 101 by the operation of the operation unit 140 of the user. Recognizing the request information (step S101), the camera unit 101 for shooting is activated (step S102).
  • the processing unit 180 controls the display 110 with the display control means 186, and performs image processing according to the deformable image, that is, the voice state of the other speaker during a call with the other speaker.
  • a display screen is displayed asking whether or not the user has the power to capture an image (step S103).
  • the shooting unit 187 of the processing unit 180 causes the display control unit 186 to enter a video that falls within the shooting range of the camera unit 101. Is controlled to display in the display area of display 110.
  • step S103 when the processing unit 180 recognizes information indicating that a deformable image is to be captured by the user's setting input, the display unit 110 displays the display 110 on the photographing unit 187. Control is performed to display the frame 51 and the frame line 52 as shown in FIG. 4 in the area (step S104). Then, so that the user enters the face of the subject within the frame 51, the right eyebrow and the left eyebrow enter the right eyebrow region 53 and the left eyebrow region 54, respectively, and the mouth enters the mouth region 55.
  • the shooting unit 187 of the processing unit 180 captures the video displayed on the display 110 as image information (step S105). ).
  • the processing unit 180 recognizes the input indicating that the deformable image is not captured in step S103, and the imaging unit 187 also captures the image displayed on the display 110 as image information. .
  • the processing unit 180 stores the captured image in the storage unit 160 so that it can be appropriately read (step S106). Further, when the processing unit 180 recognizes the setting input for capturing the deformable image in step S103, the processing unit 180 selects whether or not it is the power to register the captured deformable image in the registration information 11. Then, when recognizing the information indicating that it is registered in the registered image information 15 of the predetermined registration information 11 by the user's setting input, the processing unit 180 sets the captured image information to the predetermined image specified by the user's setting input. Record as registered image information 15 of registered information 11.
  • FIG. 11 is a flowchart of the incoming call process when a mobile phone receives an incoming call.
  • the processing unit 180 of the mobile phone 100 recognizes that the mobile phone 100 receives an incoming call from another communication device such as a mobile phone or a general phone at the arrival / departure recognition means 181.
  • the partner speaker recognition means 182 recognizes the registration information 11 of the called partner speaker (step S 202).
  • the partner speaker recognition means 182 of the processing unit 180 recognizes the telephone number of the partner speaker recorded in the incoming call information, and stores the registration information 11 having the telephone number information 14 that matches the telephone number of the partner speaker. recognize.
  • the partner speaker recognition means 182 recognizes the registered image information 15 of the recognized registration information 11 (step S203).
  • the processing unit 180 controls the display control means 186 to display the destination telephone number in the display area of the display 110, for example. Then, when the processing unit 180 recognizes an operation signal indicating that the user responds to the incoming call by operating the operation unit 140 (step S204), the processing unit 180 controls the transmission / reception unit 130 to determine the destination partner speaker. For example, a communication connection is established with a communication device through a network such as a telephone line or the Internet so that they can talk to each other.
  • the processing unit 180 recognizes information indicating that the call with the partner speaker is to be terminated by operating the operation unit 140 of the user, or receives an incoming call indicating that the call is to be terminated from the communication device of the partner speaker.
  • the transmission / reception unit 130 is controlled to cancel the call ready state, and the communication is terminated (step S205).
  • the processing unit 180 displays the registered information 11 in the display control unit 186.
  • the display 110 is controlled to display the registered image 50 of the registered image information 15 in the display area of the display 110 (step S206).
  • step S207 when the processing unit 180 recognizes an operation signal indicating that the user responds to the incoming call by operating the operation unit 140 (step S207), the processing unit 180 controls the transmission / reception unit 130 to receive the destination partner speaker. For example, a communication connection is established with a communication device such as a telephone line or the Internet so as to be able to talk to each other.
  • step S207 if the operation signal indicating that the user responds to the incoming call cannot be recognized by operating the operation unit 140 or if the operation signal indicating that the incoming call is rejected is received, the reception of the incoming information is terminated. To terminate the incoming call processing.
  • step S207 when the processing unit 180 recognizes the operation signal for responding to the incoming call and is connected to the communication apparatus of the other party's speaker, the voice recognition means 183 causes the other party The voice information of the speaker is recognized (step S208).
  • the voice recognition means 183 of the processing unit 180 analyzes the received voice information and determines the voice status of the other speaker, that is, the voice level, strength, weakness, tempo of the other speaker's speech, etc. Recognized as information (step S209).
  • the voice recognition means 183 stores the recognized voice basic information in the memory 170 so that it can be read out as appropriate.
  • the processing unit 180 changes the voice status of the other speaker while the user is talking to the other speaker.
  • Step S210 that is, when a change in the voice state of the voice information transmitted from the communication device of the other speaker is detected, the voice change determination means 184 detects the voice of the received voice information.
  • the emotion information 44 of the other speaker is recognized from the state (step S211). Specifically, the voice change determination means 184 prays the voice state of the voice information of the other speaker and compares it with the voice basic information stored in the memory 170.
  • the voice change determination means 184 of the processing unit 180 has the changed voice pattern, for example, the amount of change in the voice, the amount of change in the voice, Recognize strength changes, tempo changes, etc. Also, the voice change determination means 184 searches the voice pattern information 43 matching the voice pattern of the other speaker based on the emotion recognition information table 40, and recognizes the emotion recognition information 41 having the voice pattern information 43. . The voice change determination means 184 also recognizes the phrase information recorded in the voice pattern information 43 in the voice information of the other speaker, for example, phrases representing emotions such as “Kora” and “Hahaha”, and these phrase information Emotion recognition information 41 corresponding to is recognized. Further, the voice change determining means 184 stores the recognized emotion recognition information 41 in the memory 170 so that it can be read out appropriately. .
  • the image processing means 185 of the processing unit 180 reads the registered image 50 recorded in the registered image information 15 and edits the image of the registered image 50 (step S212).
  • the image processing means 185 of the processing unit 180 reads the first image template table 20 and the second image template table 30.
  • the emotion information 44 of the emotion recognition information 41 recognized in step S211 and stored in the memory 170 is recognized, and the first image template information 21 and the first image template information 21 having corresponding emotion information 25 and 35 corresponding to the emotion information 44 are stored. 2 Recognize image template information 31.
  • the image processing means 185 of the processing unit 180 registers in accordance with the first image pattern information 24 of the first image template information 21 and the second image pattern information 34 of the second image template information 31.
  • the emotional change image is created by deforming the right eyebrow region 53, the left eyebrow region 54, and the mouth region 55 of the image 50 (step S213).
  • step S211 an example in which information indicating "smiling state" is recorded in emotion information 44 of emotion recognition information 41 recognized by voice change determination means 184 is shown.
  • the image processing means 185 includes the first image pattern information 24 of the first image template information 21 and the second pattern information of the second image template information 31 having corresponding emotion information 25, 35 corresponding to the emotion information 44.
  • the images of the right eyebrow region 53, the left eyebrow region 54, and the mouth region 55 of the registered image 50 are transformed into, for example, a smile image as shown in FIG. 9A.
  • the right eyebrow region 53 and the left eyebrow region of the registered image are recorded.
  • the images of 54 and mouth area 55 are transformed into an “angry” face image as shown in FIG. 9B, for example.
  • the right eyebrow region 53 and the left eyebrow region 54 of the registered image are recorded.
  • the image of the mouth region 55 are transformed into a face image in a “sadness state” as shown in FIG. 9C, for example.
  • the image processing means 185 of the processing unit 180 stores the emotion-changed image deformed as described above in the memory 170 so that it can be read out appropriately.
  • the processing unit 180 controls the display 110 with the display control means 186 to display the emotion change image created at step S212 in the display area of the display 110.
  • step S214 the processing unit 180 recognizes an operation signal for ending the call with the other party speaker by the operation of the operation unit 140 of the user.
  • the processing unit 180 releases the communication connection and ends the call. If the call is to be continued, the process returns to step S210 to perform processing for recognizing the voice state of the other speaker (step S214).
  • FIG. 12 is a flowchart of the calling process when a mobile phone is called.
  • the same reference numerals are given to the processing that is substantially the same as the incoming processing of the mobile phone 100 in FIG. 11, and the description thereof is omitted or simplified.
  • the mobile phone 100 first sends a call to the communication device of the other party by operating the operation unit 140 by the user at the arrival / recognition recognition unit 181 at the arrival / departure recognition unit 181 of the processing unit.
  • the call request information is recognized (step S301).
  • the processing unit 180 is required to If the registration information 11 of the other party's speaker is recorded in the request information, it is determined that the destination telephone number is set and entered by operating the operation unit 140 of the user.
  • the transmission information is transmitted to the other communication device (step S302).
  • the transmission / reception unit 130 is controlled to communicate with the other party's caller's communication device via a network such as a telephone line or the Internet.
  • the processing unit 180 recognizes information indicating that the call with the other speaker is to be terminated by operating the operation unit 140 of the user, or receives incoming call information indicating that the communication with the other speaker is ended. If it is recognized, the transmission / reception unit 130 is controlled to release the call ready state, and the communication is terminated (step S205).
  • step S 301 when recognizing that the registration information 11 of the transmission destination is recorded in the registration information table 10 in the transmission request information, the processing unit 180 stores the registered image information 15 in the registration information 11. Whether or not is recorded is determined (step S304). In step S304, when the registered image information 15 is not recorded, the processing unit 180 executes step S302 and transmits the call to the telephone number recorded in the telephone number information 14 of the registration information 11.
  • step S 304 when the registered image information 15 is recorded in the registered information 11, the processing unit 180 causes the display control unit 186 to control the display 110 to register the registered image information 50 in the registered image information 15. Is displayed in the display area of the display 110 (step S30 5).
  • processing unit 180 controls transmission / reception unit 130 to transmit the transmission information to the telephone number recorded in telephone number information 14 of registration information 11 (step S306).
  • processing unit 180 controls transmission / reception unit 130 to determine the destination partner speaker. For example, a communication connection is established with a communication device through a network such as a telephone line or the Internet so that they can talk to each other. On the other hand, if the other speaker does not respond to the call information in step S307, the call processing of the mobile phone 100 is terminated.
  • step S307 when the other speaker responds to the transmission information, the processing unit 180 performs the processing from step S208 to step S214 of the incoming call processing described above. That is, The processing unit 180 executes step S208, and the voice recognition means 183 recognizes the voice information of the called party's partner. After that, the voice recognition means 183 of the processing unit 180 analyzes the received voice information and uses the voice status of the other speaker, that is, the voice level, strength and weakness, and the other speaker's speaking tempo as the basic voice information. recognize. The voice recognition means 183 stores the recognized voice basic information in the memory 170 so that it can be read out as appropriate.
  • processing unit 180 performs the process of step S 209, detects that the voice state of the other speaker has changed while the user is talking to the other speaker, and sends it to voice change determination means 184.
  • the other speaker's emotion information 44 is recognized from the voice state of the received voice information.
  • the voice change determination means 184 stores the recognized emotion recognition information 41 in the memory 170 so that it can be read out appropriately. .
  • the processing unit 180 performs the processing of step S210, causes the image processing unit 185 to read the registered image 50 in which the registered image information 15 is recorded, edits the image of the registered image 50, and Create a change image. Further, the image processing means 185 of the processing unit stores the emotion change image in the memory 170 so that it can be read out as appropriate.
  • the processing unit 180 performs the process of step S213, controls the display 110 with the display control means 186, and displays the emotion change image created in step S212 in the display area of the display 110. Display.
  • step S214 when the processing unit 180 performs the process of step S214 and recognizes an operation signal indicating that the call with the other party is terminated by the operation of the operation unit 140 of the user, the processing unit 180 releases the communication connection. And end the call. If the call is to be continued, the process returns to step S210 to perform processing for recognizing the voice state of the other speaker.
  • the mobile phone 100 causes the voice change determination unit 184 of the processing unit 180 to recognize the voice pattern of the other speaker's voice information and determines that the voice pattern has changed. Then, the image processing means 185 changes the registered image 50 of the registered image information 15 in accordance with the change of the sound pattern. For this reason, the registered image 50 of the registered image information 15 can be transformed as if a moving image is reproduced according to the change of the voice pattern. Also, the voice pattern of the other speaker changes depending on the emotion of the other speaker. Therefore, opponent The registered image 50 of the registered image information 15 can be transformed according to the speaker's emotion, and an appropriate emotion change image corresponding to the emotion of the other speaker can be displayed on the display 110.
  • the registered image 50 of the registered image information 15 is deformed and displayed as an emotion change image. For this reason, it is not necessary to prepare a plurality of images corresponding to each emotion. Therefore, it is possible to effectively utilize the free capacity that does not press the storage capacity of the storage unit 160.
  • the image processing means 185 deforms the registered image 50 of the registered image information 15 based on the first image template information 21 and the second image template information 31, and generates an emotion change image. For this reason, only the facial expression can be transformed without modifying the background of the face image displayed in the display area of the display 110, the position of the face image, or the like. Therefore, when the image displayed on the display 110 is switched from the registered image 50 to the emotional change image, or when switching from the emotional change image to another emotional change image, the image does not shift and becomes difficult to see. You can switch images.
  • the image processing means 185 determines the right eyebrow region 53, the left eyebrow region 54, the eyebrow region 54 based on the eyebrow and mouth change rates recorded in the first image pattern information 24 and the second image pattern information 34.
  • the left eyebrow, the right eyebrow, and the mouth of the mouth region 55 are deformed by moving the dots by a predetermined amount, for example. Therefore, the facial expression can be easily changed by subjecting only a part of the registered image 50 of the registered image information 15 to the image transformation process. Therefore, since only part of the registered image 50 is processed, the processing load on the processing unit associated with the image processing can be reduced.
  • the image processing means 185 changes the right eyebrow of the right eyebrow region 53, the left eyebrow of the left eyebrow region 54, and the mouth of the mouth region 55, respectively. For this reason, it is possible to deform the eyebrows and mouth of the face that can best display emotional changes. Therefore, the user can easily confirm the other speaker's emotion from the emotion change image.
  • the image processing means 185 determines the right eyebrow region 53, the left eyebrow region based on the eyebrow deformation rate and the mouth deformation rate recorded in the first image pattern information 24 and the second image pattern information 34.
  • the eyebrow in area 54 and the mouth in area 55 are deformed.
  • the voice of the other speaker's voice According to the pattern, a predetermined part of the image can be appropriately deformed by moving it in a predetermined direction by a predetermined amount.
  • the image processing means 185 also includes the first image basic information 23 that substantially matches the right eyebrow region 53 of the registered image 50, the eyebrow of the left eyebrow region 54, and the second image that substantially matches the mouth of the mouth region 55.
  • the voice change determination means 184 compares the voice basic information recognized by the voice recognition means 183 with the voice information of the other speaker, and the voice pattern of the voice information of the other speaker is the voice basic information. If the voice pattern is different, the voice pattern power also recognizes the emotion of the other speaker. Then, the image processing means 185 processes the image according to the emotion of the other speaker. For this reason, the voice change determination means 184 can easily recognize the change of the voice pattern. Therefore, the image processing means 185 can appropriately transform the registered image 50 into an emotion change image corresponding to the change in the other speaker's emotion.
  • the voice recognition means 183 receives the first voice information transmitted from the communication device of the other speaker after the mobile phone 100 and the communication device of the other speaker are connected for communication and ready to communicate. Recognize a voice pattern as basic voice information. For this reason, the voice change determination means 184 can recognize how the emotion of the other speaker has changed compared to the state at the beginning of the conversation, based on the emotion of the other speaker at the beginning of the conversation. It is possible to display on the display 110 how the other speaker's emotion has changed compared to the initial state of speaking.
  • the display control means 186 causes the display 110 to display the registered image 50 recorded in the registered image information 15 when an incoming call is received from the other speaker.
  • the image processing means 185 converts the registered image 50 based on the first image template table 20 and the second image template table 30. Let For this reason, the user can make a call while confirming the feeling of the other party's speaker in a device that makes a call with the other party's speaker, such as the mobile phone 100 as in the above embodiment. Shi Therefore, even if there is no function such as a videophone, for example, it is possible to talk with the other party while confirming the emotional change of the other party with a simple configuration and to provide good call support.
  • the display control means 186 causes the display 110 to display the registered image 50 recorded in the registered image information 15 in the same way even during outgoing calls for transmitting information to the other party. Then, when the sound change determination means 184 determines that the state of the sound information has changed, the image processing means 185 transforms the registered image 50 based on the first image template table 20 and the second image template table 30. Let For this reason, as with incoming calls, the user can make a call while checking the feelings of the other speaker, and can support good calls.
  • the photographing means 187 displays the frame 51 and the frame line 52 on the display 110 when photographing the registered image 50 recorded in the registered image information 15. For this reason, when the user captures the registered image 50, the user focuses on the subject so that his / her eye is placed in the frame 51 and the eyebrow and mouth are placed in line with the frame line 52. Is possible. Therefore, the registered image 50 recorded in the registered image information 15 can be easily taken.
  • image information captured using the frame 51 and the frame line 52 is recorded in the registered image information 15. Therefore, the image processing means 185 can easily recognize the right eyebrow region 53, the left eyebrow region 54, and the mouth region 55, and can transform the image of the eyebrow and mouth. Therefore, the registered image 50 can be appropriately edited according to the voice pattern of the voice information of the other speaker and displayed on the display 110.
  • the cellular phone 100 causes the image processing means 185 to transform the image recorded in the registered image information 15, and causes the display control means 186 to display the image on the display 110.
  • the registered image 50 of the registered image information 15 is transformed according to the received audio information, the emotional change image that has been stably transformed without the need to send and receive large files such as image information by the communication means. Can be displayed.
  • the power illustrated as the mobile phone 100 as the image processing device and the image display device is not limited to this.
  • the present invention may be applied to other electric devices such as a personal computer, a general telephone device, and a car navigation device.
  • a personal computer it recognizes the voice state such as voice information in which microphone power is input, voice information obtained from a network such as the Internet, voice information obtained from a storage medium such as an optical disk, and the like. It is also possible to adopt a configuration in which a change in the sound pattern is detected and the image is deformed by the image processing means.
  • the operation unit 140 includes an operation button and an operation knob.
  • the present invention is not limited to this.
  • a touch panel that allows a predetermined setting input operation by touching the display 110 may be used even if it is a keyboard or a mouse connected to the mobile phone 100.
  • V and shift configurations that can be set and input various setting items such as voice input operations and configurations that output signals via a wireless medium such as a remote controller can be applied.
  • the arrival / departure recognition means 181 has shown a configuration in which when the incoming call information is received, the voice output unit 120 outputs a ringtone such as a voice, a warning sound, and a notification sound to notify the user of the incoming call.
  • a ringtone such as a voice, a warning sound, and a notification sound to notify the user of the incoming call.
  • the incoming call sound may be output after the called party recognition means 182 recognizes the called party.
  • the ring tone can be changed depending on the destination, and the user can be informed of who the other party is by using only the ring tone.
  • the mobile phone 100 is configured to include a vibration unit, and when the incoming / outgoing recognition unit 181 recognizes an incoming call, the mobile unit 100 vibrates the vibration unit and reports incoming information. Further, as described above, the vibration means may be vibrated after the partner speaker recognition means 182 recognizes the called party. In this case, the vibration pattern can tell the user who the call is.
  • the voice recognition means 183 recognizes the gender of the other speaker, basic voice voice information, etc., based on the contents described in the registration details information 16 of the registration information 11. It is good also as a structure. In such a configuration, the voice recognition means 183 records the registration details information 16. Based on the recorded voice basic information, it is possible to compare the voice state of the other speaker at the time of incoming or outgoing call and the information recorded in the voice basic information. Therefore, for example, even when the other speaker is angry at the time of an incoming call or outgoing call, the “anger state” of the other speaker can be displayed on the display 110. Therefore, an image corresponding to the emotion of the other speaker can be displayed more appropriately.
  • the basic voice information may be transmitted to the mobile phone 100 such as the communication device of the other speaker. Even in such a configuration, the cellular phone 100 can eliminate the process of recognizing the basic voice information, which simplifies the configuration and reduces the processing load.
  • the mobile phone 100 records a plurality of registered images associated with the emotion information 44 in the registered image information 15. Then, when the voice change determination means 184 of the processing unit 180 recognizes the change in the voice pattern of the voice information of the other speaker, it recognizes the emotion recognition information 41 having the voice pattern information 43 corresponding to the voice pattern. Then, the display control means 186 performs control to read a registered image corresponding to the emotion information 44 of the emotion recognition information 41 from a plurality of registered images recorded in the registered image information 15 and display it in the display area of the display 110.
  • the image displayed on the display 110 may be switched to an image corresponding to the emotion of the other speaker according to a change in the voice state of the other speaker. Even in such a configuration, it is possible to display an appropriate emotion change image on the display 110 according to the change in the emotion of the other speaker, as in the above-described effects. Furthermore, since the image recorded in advance in the storage unit 160 is switched, it is not necessary to process the image. Therefore, the processing load on the processing unit 180 can be reduced, and the processing speed can be increased.
  • the image processing means 185 places a part of the registered image 50, for example, the eyebrows and the mouth, on the other eyebrows and mouth element images in accordance with the emotion recognition information 41 recognized by the sound change judgment means 184. It is good also as a structure to replace.
  • an eyebrow image as an element image as shown in FIG. 13 is recorded in the first image pattern information 24 of the first image template information 21.
  • a mouth image as an element image as shown in FIG. 14 is recorded in the second image pattern information 34 of the second image template information 31.
  • Figure 13A shows the person in the eyebrow image of the first basic image information 23 It is an example of the eyebrow image image
  • FIG. 13B is an example of an eyebrow image captured when the person in the eyebrow image of the first image basic information 23 is “angry”.
  • FIG. 13C is an example of an eyebrow image captured when the person of the eyebrow image of the first image basic information 23 is in a sad state.
  • FIG. 14A is an example of a mouth image taken when a person in the mouth image of the second image basic information 33 is smiling.
  • FIG. 14B is an example of a mouth image taken when the person of the mouth image of the second image basic information 33 is “angry”.
  • FIG. 14C is an example of a mouth image taken when the person in the mouth image of the second image basic information 33 is in a sad state.
  • the voice change determination means 184 determines that the voice pattern of the other speaker's voice information is different from the voice pattern of the voice basic information
  • the emotion recognition of the voice pattern information 43 corresponding to the voice pattern is performed. Recognize information 41.
  • the image processing means 185 obtains the first image basic information 23 and the second image basic information 33 of the eyebrow image whose shapes substantially coincide with the eyebrows of the right and left eyebrow regions 53 and 54 and the mouth region 55 of the registered image 50.
  • the first image template information 21 and the second image template information 31 are searched.
  • the image processing means 185 uses the searched first image template information 21 and second template information 31 to detect the emotion information 44 of the emotion recognition information 41 recognized by the voice change determination means 184.
  • First image template information 21 and second template information 31 having 25 and 35 are recognized.
  • the image processing means 185 converts the images of the left and right eyebrow regions 53 and 54 and the mouth region 55 of the registered image 50 into the recognized first image pattern information 21 of the first image template information 21 and second template information 31.
  • the second image pattern information 34 is replaced with the image. Even in such a configuration, the image information can be switched and displayed according to the voice information of the other speaker, and the emotion of the other speaker can be easily known.
  • the image processing means 185 may be configured to similarly deform other parts of the face image, such as the nose, forehead, left and right eyelids, left and right jaws, ears, and the face contour. Further, the color of each of these parts may be transformed. For example, if there is information indicating “angry state” in the corresponding emotion information, the image processing means 185 changes the contrast and color by, for example, increasing the red intensity of the ears and performing image transformation processing. May be.
  • the registered image 50 recorded in the registered image information 15 has a face in the frame 51, Although the front image of the face in a state where the eyebrow image and the mouth image are in the left and right eyebrow areas 53 and 54 and the mouth area 55, respectively, is taken, the present invention is not limited to this.
  • a configuration in which a face image taken from a side face is used, or a face image taken from an oblique direction may be used.
  • a profile is used as a registered image, it is possible to use a frame dedicated to the landscape image and focus on the profile so that the profile appears within this frame. Furthermore, it is good also as a structure which recognizes the outline of a face automatically.
  • the boundary force between the face color of the registered image in the registered image information 15 and the color of the landscape image also recognizes the outline of the face, and also recognizes the power of the eyes, mouth, eyebrows, etc. inside the face such as different colors. .
  • the boundary force between the face color of the registered image in the registered image information 15 and the color of the landscape image also recognizes the outline of the face, and also recognizes the power of the eyes, mouth, eyebrows, etc. inside the face such as different colors. .
  • not only images taken from the front but also images taken from various directions can be processed and deformed.
  • an animal image, a doll image, or the like may be recorded. Even with such individuality, it is possible to create an emotion change image in which the expression of an animal or doll is changed by the same processing as in the above-described embodiment, and display it in the display area of the display 110. It is also possible to switch between a plurality of images, for example, a plurality of animal images, a plurality of doll images, etc., according to the voice state of the other speaker's voice information.
  • the mobile phone 100 is not limited to the power shown in the configuration for recognizing the voice information of the other party speaker transmitted from the communication device and recognizing the voice state pattern of the voice information. Absent.
  • the voice state of voice information may be transmitted from the communication device of the other speaker, and the mobile phone 100 may be configured to receive the voice state transmitted from the communication device of the other speaker. This configuration eliminates the need for detecting the voice state of the voice information with the mobile phone 100, thereby reducing the processing load and simplifying the configuration.
  • the image information modified according to the sound pattern of the sound information by the image processing apparatus of the present invention may be transmitted together with the sound information to a receiving device such as a mobile phone.
  • both the user and the other party's speaker are configured as a communication system using the mobile phone 100 of the above embodiment.
  • both the user and the other speaker use the mobile phone 100 of the above embodiment as a transmission / reception terminal. You can.
  • each user can easily feel the emotion of the caller of the mobile phone 100 and the other party of the callee by viewing the image displayed on the display 110. Can be recognized.
  • the display 110 may be displayed only when a call is made to a communication device such as a general telephone of the other speaker, and the image may be changed according to the voice state of the other speaker.
  • a configuration may be adopted in which an image is displayed on the display 110 only when the communication device power of the speaker is received and the image is changed according to the voice state of the other speaker.
  • Each function described above has been constructed as a program, but it can be used in any form, for example, configured by hardware such as a circuit board or an element such as a single IC (Integrated Circuit). .
  • hardware such as a circuit board or an element such as a single IC (Integrated Circuit).
  • the mobile phone 100 causes the voice change determination unit 184 of the processing unit 180 to recognize the voice pattern of the voice information of the other speaker and determines that the voice pattern has changed. Then, the image processing means 185 changes the registered image 50 of the registered image information 15 according to the change of the sound pattern. Therefore, the registered image 50 of the registered image information 15 can be deformed according to the change of the voice pattern, and the registered image 50 of the registered image information 15 can be deformed according to the emotion of the other speaker. It is possible to display on the display 110 an appropriate emotion change image corresponding to the emotion of the person.
  • the present invention can be used for an image processing device that displays an image, an image display device, a receiving device, a transmitting device, a communication system, an image processing method, an image processing program, and a recording medium that records the image processing program.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Quality & Reliability (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Telephone Function (AREA)

Abstract

A mobile telephone (100) causes voice change judgment means of a processing unit (180) to recognize a voice pattern of voice information on a partner-speaker. When it is judged that the voice pattern has been changed, the mobile telephone (100) causes image processing means to change a registered image of the registered image information according to the change of the voice pattern. Thus, it is possible to change the registered image of the registered image information according to the voice pattern change and display an appropriate feeling change image on a display (110) according to the partner-speaker.

Description

明 細 書  Specification
画像処理装置、画像表示装置、受信装置、送信装置、通信システム、画 像処理方法、画像処理プログラム、画像処理プログラムを記録した記録媒体 技術分野  Image processing apparatus, image display apparatus, receiving apparatus, transmitting apparatus, communication system, image processing method, image processing program, and recording medium recording image processing program
[0001] 本発明は、画像を表示する画像処理装置、画像表示装置、受信装置、送信装置、 通信システム、画像処理方法、画像処理プログラム、画像処理プログラムを記録した 記録媒体に関する。  The present invention relates to an image processing device that displays an image, an image display device, a receiving device, a transmitting device, a communication system, an image processing method, an image processing program, and a recording medium that records the image processing program.
背景技術  Background art
[0002] 従来、画像を表示させるディスプレイを備えた携帯用電話機にお!ヽて、受信者側の 携帯電話機のディスプレイに表示させる画像を発信者が選択可能な構成が知られて いる。(例えば、特許文献 1参照)。  Conventionally, there has been known a configuration in which a caller can select an image to be displayed on a display of a mobile phone on the receiver side in a mobile phone having a display for displaying an image. (For example, see Patent Document 1).
[0003] 特許文献 1に記載のものは、発信者側の携帯電話機から受信者側の携帯電話機 に発信する際に、発信者側の携帯電話機は、発信者により複数の画像から発信者の 感情に応じて選択された画像を受信者側の携帯電話機に送信する。そして、受信者 側の携帯電話機は、着信と同時に送信された画像をディスプレイに表示させる構成 が採られている。  [0003] In the device described in Patent Document 1, when a call is made from a mobile phone on the sender side to a mobile phone on the receiver side, the mobile phone on the caller side sends the emotion of the caller from a plurality of images by the caller. The selected image is transmitted to the mobile phone on the receiver side. The mobile phone on the receiver side is configured to display on the display an image transmitted simultaneously with the incoming call.
[0004] 特許文献 1 :特開 2004— 15158号公報 (第 4頁ないし第 5頁、図 3ないし図 4参照) 発明の開示  Patent Document 1: Japanese Patent Application Laid-Open No. 2004-15158 (refer to pages 4 to 5 and FIGS. 3 to 4) Disclosure of the Invention
発明が解決しょうとする課題  Problems to be solved by the invention
[0005] ところで、特許文献 1のような構成では、発信者側が受信者側の携帯電話機に発信 する際に、受信者側に表示させる画像を選択するため、発信時の画像しか伝達する ことができない。このため、通話中にディスプレイに表示される画像を変更することが できない。したがって、例えば通話中に感情が変化したとしても、その感情に応じた 画像を受信者側の携帯電話機のディスプレイに表示させることができな 、と 、う問題 がー例として挙げられる。  [0005] By the way, in the configuration as in Patent Document 1, when the caller side makes a call to the mobile phone on the receiver side, an image to be displayed on the receiver side is selected, so only the image at the time of transmission can be transmitted. Can not. For this reason, the image displayed on the display during a call cannot be changed. Therefore, for example, even if the emotion changes during a call, an image corresponding to the emotion cannot be displayed on the receiver's mobile phone display.
[0006] 本発明は、適切な画像を表示できる画像処理装置、画像表示装置、受信装置、送 信装置、通信システム、画像処理方法、画像処理プログラム、画像処理プログラムを 記録した記録媒体を提供することを 1つの目的とする。 [0006] The present invention relates to an image processing device, an image display device, a receiving device, a transmitting device, a communication system, an image processing method, an image processing program, and an image processing program capable of displaying an appropriate image. One objective is to provide a recorded recording medium.
課題を解決するための手段  Means for solving the problem
[0007] 本発明の画像処理装置は、表示手段に表示される画像を、受信して得られる音声 情報に応じて処理する画像処理装置であって、前記音声情報における音声状態の 変化を認識する音声状態認識手段と、前記画像を前記表示手段に表示させるととも に、前記表示された画像を前記音声状態の変化に応じて変更させる表示制御手段と 、を具備したことを特徴とする。  [0007] The image processing apparatus of the present invention is an image processing apparatus that processes an image displayed on a display unit according to audio information obtained by reception, and recognizes a change in an audio state in the audio information. And a display control unit configured to display the image on the display unit and to change the displayed image in accordance with the change in the audio state.
[0008] 本発明の画像表示装置は、画像を記憶する記憶手段と、前記音声情報を受信する 際に前記画像を表示する表示手段と、前述した本発明の画像処理装置と、を具備し たことを特徴としたことを特徴とする。  [0008] The image display device of the present invention comprises storage means for storing an image, display means for displaying the image when the audio information is received, and the image processing device of the present invention described above. It is characterized by that.
[0009] 本発明の受信装置は、前述した本発明の画像表示装置と、前記音声情報を受信 可能な受信手段と、を備え、前記表示制御手段は、前記受信手段により音声情報が 受信される際に、その音声情報の送信元に対応する画像を表示することを特徴とす る。  [0009] A receiving device of the present invention includes the above-described image display device of the present invention and a receiving unit capable of receiving the audio information, and the display control unit receives the audio information by the receiving unit. In this case, an image corresponding to the transmission source of the audio information is displayed.
[0010] 本発明の送信装置は、前述した本発明の画像表示装置と、前記音声信号を送受 信可能な送信手段と、を備え、前記送信手段は、前記音声情報を送信する送信先に 対して発呼する発呼手段を有し、前記表示手段は、前記発呼手段による発呼または 前記発呼に対する前記送信先の応答に応じて前記送信先に対応する画像を表示し 、前記画像処理装置は、前記発呼に応答して受信される送信先からの音声情報に おける音声状態の変化に応じて、表示された画像を変更することを特徴とする。  [0010] A transmission device of the present invention includes the above-described image display device of the present invention and a transmission unit capable of transmitting and receiving the audio signal, wherein the transmission unit is for a transmission destination that transmits the audio information. The display unit displays an image corresponding to the transmission destination in response to a call made by the calling unit or a response of the transmission destination to the call, and the image processing The apparatus is characterized in that the displayed image is changed in accordance with a change in a voice state in voice information from a transmission destination received in response to the call.
[0011] 本発明の通信システムは、相互に音声情報を送受信可能な送受信端末を備えた 通信システムであって、前記送受信端末の各々は、画像を記憶する記憶手段と、音 声情報を送信する送信手段と、音声情報を送信するために送信元に対して発呼する 発呼手段と、音声情報を受信する受信手段と、前記音声情報を受信する際に、また は前記発呼手段により発呼する際に、または前記発呼に対し前記送信先が応答する 際に前記画像を表示する表示手段と、前述した本発明の画像処理装置と、を具備し たことを特徴とする。  [0011] The communication system of the present invention is a communication system including a transmission / reception terminal capable of transmitting and receiving audio information to and from each other, and each of the transmission / reception terminals transmits storage means for storing images and audio information. A transmitting means, a calling means for making a call to a transmission source to transmit the voice information, a receiving means for receiving the voice information, and when receiving the voice information, or by the calling means. It is characterized by comprising display means for displaying the image when making a call or when the transmission destination responds to the call, and the image processing apparatus of the present invention described above.
[0012] 本発明の画像処理方法は、表示手段に表示される画像を、受信して得られる音声 情報に応じて処理する画像処理方法であって、前記音声情報における音声状態の 変化を認識し、前記画像を前記表示手段に表示させるとともに、前記表示された画 像を前記音声状態の変化に応じて変更させることを特徴とする。 [0012] The image processing method of the present invention provides a sound obtained by receiving an image displayed on the display means. An image processing method for processing according to information, comprising: recognizing a change in sound state in the sound information, displaying the image on the display means, and displaying the displayed image according to the change in sound state. It is characterized by being changed.
[0013] 本発明の画像処理プログラムは、演算手段を前述した本発明の画像処理装置とし て機能させることを特徴とする。  [0013] An image processing program of the present invention is characterized in that a computing means functions as the above-described image processing apparatus of the present invention.
[0014] 本発明の画像処理プログラムは、前述した本発明の画像処理方法を演算手段に実 施させることを特徴とする。 [0014] An image processing program of the present invention is characterized by causing a calculation means to perform the above-described image processing method of the present invention.
[0015] 本発明の画像処理プログラムを記録した記録媒体は、前述した本発明の画像処理 プログラムが演算手段にて読取可能に記録されたことを特徴とする。 図面の簡単な説明 A recording medium on which the image processing program of the present invention is recorded is characterized in that the above-described image processing program of the present invention is recorded so as to be readable by an arithmetic means. Brief Description of Drawings
[0016] [図 1]本発明の一実施の形態に係る携帯用電話機器の全体斜視図である。 FIG. 1 is an overall perspective view of a portable telephone device according to an embodiment of the present invention.
[図 2]携帯用電話機器の内部構成を模式的に示すブロック図である。  FIG. 2 is a block diagram schematically showing the internal configuration of the portable telephone device.
[図 3]記憶手段に記録される登録情報テーブルの概略を示す模式図である。  FIG. 3 is a schematic diagram showing an outline of a registration information table recorded in storage means.
[図 4]記憶手段に記録される登録画像情報の画像を示す模式図である。  FIG. 4 is a schematic diagram showing an image of registered image information recorded in a storage means.
[図 5]記憶手段に記録される第 1画像テンプレートテーブルの概略構成を示す模式 図である。  FIG. 5 is a schematic diagram showing a schematic configuration of a first image template table recorded in a storage means.
[図 6]記憶手段に記録される第 2画像テンプレートテーブルの概略構成を示す模式 図である。  FIG. 6 is a schematic diagram showing a schematic configuration of a second image template table recorded in a storage means.
[図 7]記憶手段に記録される感情認識情報テーブルの概略構成を示す模式図である  FIG. 7 is a schematic diagram showing a schematic configuration of an emotion recognition information table recorded in a storage means.
[図 8]処理部の構成を模式的に示すブロック図である。 FIG. 8 is a block diagram schematically showing a configuration of a processing unit.
[図 9A]画像処理手段に変形処理された感情変形画像を示す模式図である。  FIG. 9A is a schematic diagram showing an emotion deformation image transformed by the image processing means.
[図 9B]画像処理手段に変形処理された他の感情変形画像を示す模式図である。  FIG. 9B is a schematic diagram showing another emotion deformation image transformed by the image processing means.
[図 9C]画像処理手段に変形処理された他の感情変形画像を示す模式図である。  FIG. 9C is a schematic diagram showing another emotion deformation image transformed by the image processing means.
[図 10]携帯電話の撮影処理を示すフローチャートである。  FIG. 10 is a flowchart showing a photographing process of a mobile phone.
[図 11]携帯電話の着信時における着信処理のフローチャートである。  FIG. 11 is a flowchart of incoming call processing when a mobile phone is incoming.
[図 12]携帯電話の発信時の発信処理におけるフローチャートである。  FIG. 12 is a flowchart of a calling process when a mobile phone is called.
[図 13A]本実施の形態の携帯電話の変形例における第 1画像基本情報の眉画像の 人物が笑顔の状態の時に撮影される眉画像の一例を示す模式図である。 [FIG. 13A] Eyebrow image of the first image basic information in a modification of the mobile phone of the present embodiment It is a schematic diagram which shows an example of the eyebrow image image | photographed when a person is in the state of a smile.
[図 13B]本実施の形態の携帯電話の変形例における第 1画像基本情報の眉画像の 人物が「怒り状態」の時に撮影される眉画像の一例を示す模式図である。  FIG. 13B is a schematic diagram showing an example of an eyebrow image taken when a person in the eyebrow image of the first image basic information is “angry” in a modification of the mobile phone of the present embodiment.
[図 13C]本実施の形態の携帯電話の変形例における第 1画像基本情報の眉画像の 人物が悲しみ状態の時に撮影される眉画像の一例を示す模式図である。  FIG. 13C is a schematic diagram showing an example of an eyebrow image captured when a person in the eyebrow image of the first image basic information is in a sad state in a modification of the mobile phone according to the present embodiment.
[図 14A]本実施の形態の携帯電話の変形例における第 2画像基本情報の口画像の 人物が笑顔の状態の時に撮影される口画像の一例を示す模式図である。  FIG. 14A is a schematic diagram showing an example of a mouth image taken when a person in the mouth image of the second image basic information is in a smiling state in a modification of the mobile phone according to the present embodiment.
[図 14B]第 2画像基本情報の口画像の人物が「怒り状態」の時に撮影される口画像の 一例を示す模式図である。  FIG. 14B is a schematic diagram showing an example of a mouth image taken when a person in the mouth image of the second image basic information is “angry”.
[図 14C]第 2画像基本情報の口画像の人物が悲しみ状態の時に撮影される口画像 の一例を示す模式図である。  FIG. 14C is a schematic diagram showing an example of a mouth image taken when a person in the mouth image of the second image basic information is in a sad state.
符号の説明 Explanation of symbols
24変形量情報および変形方向情報としての第 1画像パターン情報 24 First image pattern information as deformation amount information and deformation direction information
34変形量情報および変形方向情報としての第 2画像パターン情報 34 Second image pattern information as deformation amount information and deformation direction information
50画像としての登録画像 Registered images as 50 images
100画像表示装置、受信装置、送信装置、および送受信端末としても  Also as 100 image display device, receiving device, transmitting device, and transmitting / receiving terminal
機能する携帯電話 Functioning mobile phone
110表示手段としてのディスプレイ  110 Display as display means
130受信手段、および送信手段としての送受信部  130 Receiving means and transmitting / receiving section as transmitting means
160記憶手段  160 storage means
180画像処理装置としての処理部  180 processing unit as image processing device
181発呼手段としての発着認識手段 181 Call arrival / reception recognition means
183音声情報認識手段、および標準音声状態認識手段としても機能す  183 Also functions as voice information recognition means and standard voice state recognition means
る音声認識手段 Voice recognition means
184音声状態認識手段としての音声変化判定手段  184 Voice change determination means as voice state recognition means
185画像情報認識手段、および画像変形処理手段としても機能する画  Image that also functions as 185 image information recognition means and image deformation processing means
像処理手段 Image processing means
186表示制御手段 発明を実施するための最良の形態 186 Display control means BEST MODE FOR CARRYING OUT THE INVENTION
[0018] 以下、本発明における一実施の形態を図面に基づいて説明する。  Hereinafter, an embodiment of the present invention will be described with reference to the drawings.
[0019] 〔携帯電話機の構成〕  [Configuration of mobile phone]
図 1は、本発明の一実施の形態に係る携帯用電話機器の全体斜視図である。図 2 は、携帯用電話機器の内部構成を模式的に示すブロック図である。図 3は、記憶手 段に記録される登録情報テーブルの概略を示す模式図である。図 4は、記憶手段に 記録される登録画像情報の画像を示す模式図である。図 5は、記憶手段に記録され る第 1画像テンプレートテーブルの概略構成を示す模式図である。図 6は、記憶手段 に記録される第 2画像テンプレートテーブルの概略構成を示す模式図である。図 7は 、記憶手段に記録される感情認識情報テーブルの概略構成を示す模式図である。 図 8は、処理部の構成を模式的に示すブロック図である。図 9Aは、画像処理手段に 変形処理された感情変形画像を示す模式図である。図 9Bは、画像処理手段に変形 処理された他の感情変形画像を示す模式図である。図 9Cは、画像処理手段に変形 処理された他の感情変形画像を示す模式図である  FIG. 1 is an overall perspective view of a portable telephone device according to an embodiment of the present invention. FIG. 2 is a block diagram schematically showing the internal configuration of the portable telephone device. FIG. 3 is a schematic diagram showing an outline of a registration information table recorded in the storage device. FIG. 4 is a schematic diagram showing an image of registered image information recorded in the storage means. FIG. 5 is a schematic diagram showing a schematic configuration of the first image template table recorded in the storage means. FIG. 6 is a schematic diagram showing a schematic configuration of the second image template table recorded in the storage means. FIG. 7 is a schematic diagram showing a schematic configuration of an emotion recognition information table recorded in the storage means. FIG. 8 is a block diagram schematically showing the configuration of the processing unit. FIG. 9A is a schematic diagram showing an emotion-transformed image transformed by the image processing means. FIG. 9B is a schematic diagram showing another emotion deformation image transformed by the image processing means. FIG. 9C is a schematic diagram showing another emotion deformation image transformed by the image processing means.
[0020] 図 1および図 2において、 100は、画像表示装置、受信装置、送信装置、および送 受信端末としても機能する携帯用電話機器 (以後、携帯電話と称する)であり、この携 帯電話 100は、他の携帯電話や一般電話機器などの送信装置としての通信機器 ( 図示せず)と電話回線やインターネット回線などのネットワークを介して通信し、この 通信機器と通話可能な状態に通信接続する。そして、通信先の相手話者に応じて所 定の画像をディスプレイ 110に表示させる。なお、本実施の形態では、画像表示装置 として、携帯電話 100を用いて説明するが、これに限定されず、例えば画像表示装 置として、パーソナルコンピュータで画像を処理してモニターなどの画像表示部に表 示させる装置などであってもよい。そして、この携帯電話 100は、内部に回路基板な どを収納可能な上部筐体 100Aおよび下部筐体 100Bを備えて構成されて 、る。ま た、上部筐体 100Aは、回動部 100Cにより下部筐体 100Bに回動可能に設けられて いる。なお、図 1には、上部筐体 100Aが下部筐体 100Bに回動可能に取り付けられ る例を示した力 これに限定されず、例えば上部筐体 100Aと下部筐体 100Bとが一 体に形成されて 、るものであってもよ 、。 [0021] そして、上部筐体 100Aには、ディスプレイ 110、音声出力部 120、アンテナ 130A (図 1参照)を有する受信手段および送信手段としての送受信部 130 (図 2参照)が設 けられ、下部筐体 100Bには、操作部 140、音声入力部 150、記憶手段 160、メモリ 1 70、および画像処理装置としても機能する処理部 180などが設けられている。また、 上部筐体 100Aの内部に設けられ、ディスプレイ 110、音声出力部 120、および送受 信部 130が接続される回路基板は、下部筐体 100Bの内部の処理部 180が接続さ れる回路基板に、回動部 100Cの内部を挿通されるフレキシブル基板などにより電気 的に接続されている。また、上部筐体 100Aには、一部に内部と外部とを連通する窓 部が形成されており、この窓部には、写真撮影用のカメラ部 101が外部に臨む状態 で取り付けられている。なお、カメラ部 101は、上部筐体 100Aのディスプレイ 110が 臨む側と反対側の面、またはその他の位置に設けられてもよ 、。 In FIG. 1 and FIG. 2, reference numeral 100 denotes a portable telephone device (hereinafter referred to as a mobile phone) that also functions as an image display device, a receiving device, a transmitting device, and a transmitting / receiving terminal. The 100 communicates with a communication device (not shown) as a transmitting device such as another mobile phone or a general telephone device via a network such as a telephone line or an Internet line, and is connected so as to be able to make a call with this communication device. To do. Then, a predetermined image is displayed on the display 110 according to the communication partner. In this embodiment, the mobile phone 100 is used as the image display device. However, the present invention is not limited to this. For example, as the image display device, an image display unit such as a monitor that processes an image with a personal computer. It may be a device that is displayed on the screen. The cellular phone 100 is configured to include an upper casing 100A and a lower casing 100B that can accommodate a circuit board and the like. Further, the upper casing 100A is rotatably provided to the lower casing 100B by a rotating portion 100C. Note that FIG. 1 illustrates an example in which the upper casing 100A is pivotably attached to the lower casing 100B. For example, the upper casing 100A and the lower casing 100B are combined. It may be something that is formed. [0021] The upper casing 100A is provided with a display 110, an audio output unit 120, a receiving unit having an antenna 130A (see Fig. 1) and a transmitting / receiving unit 130 (see Fig. 2) as a transmitting unit. The housing 100B is provided with an operation unit 140, an audio input unit 150, a storage unit 160, a memory 170, a processing unit 180 that also functions as an image processing device, and the like. In addition, the circuit board provided in the upper casing 100A and connected to the display 110, the audio output unit 120, and the transmission / reception unit 130 is connected to the circuit board to which the processing unit 180 in the lower casing 100B is connected. They are electrically connected by a flexible board inserted through the inside of the rotating part 100C. In addition, the upper casing 100A is partially formed with a window portion that communicates the inside and the outside, and the camera portion 101 for taking a picture is attached to the window portion so as to face the outside. . The camera unit 101 may be provided on the surface of the upper housing 100A opposite to the side on which the display 110 faces, or on another position.
[0022] ディスプレイ 110は、処理部 180の制御により、所定の画像情報やテキスト情報など の各種情報が表示される表示領域を備えている。この画像情報としては、例えば記 憶手段 160に記録された画像情報などの他、図示しな ヽ TV受信機で受信した TV 画像データ、外部装置など光ディスクや磁気ディスク、メモリカードなどの記録媒体に 記録されドライブやドライバなどにて読み取った画像データ、メモリ 170からの画像デ ータなどである。また、ディスプレイ 110としては、例えば液晶表示パネルや有機 EL ( Electro Luminescence)ノヽネノレ、 PDP (Plasma Display Panel)、 CRT (し atnode— Ray T ube)、 FED (Field Emission Display)、電気泳動ディスプレイパネルなどが例示できる  The display 110 includes a display area in which various information such as predetermined image information and text information is displayed under the control of the processing unit 180. This image information includes, for example, image information recorded in the storage means 160, TV image data received by a TV receiver (not shown), recording media such as an external device such as an optical disk, a magnetic disk, and a memory card. Image data recorded and read by a drive or driver, or image data from the memory 170. Examples of the display 110 include a liquid crystal display panel, an organic EL (Electro Luminescence), a PDP (Plasma Display Panel), a CRT (and atnode-Ray Tube), an FED (Field Emission Display), and an electrophoretic display panel. Can be illustrated
[0023] 音声出力部 120は、処理部 180の制御により、所定の音声情報を音声に変換して 出力する。この音声情報としては、例えば、通信相手話者の音声に関する音声情報 や、記憶手段 160などに記録された音楽情報、メモリ 170などに記憶されている警告 音などが例示できる。また、図 1において示される音声出力部 120は、利用者の操作 により指向性および無指向性を切り替え可能なスピーカを有しており、例えば利用者 が携帯電話 100を手放した状態で利用する際にも、音声出力部 120から出力される 音声を聴認可能となるように構成されている。なお、例えば下部筐体 100Bの操作部 140が設けられる面とは反対側面に、別途無指向性のスピーカを設ける構成としても よい。 The sound output unit 120 converts predetermined sound information into sound and outputs the sound under the control of the processing unit 180. As this voice information, for example, voice information related to the voice of the communication partner speaker, music information recorded in the storage means 160 or the like, a warning sound stored in the memory 170 or the like can be exemplified. In addition, the audio output unit 120 shown in FIG. 1 has a speaker that can be switched between directivity and omnidirectionality by the user's operation. For example, when the user uses the mobile phone 100 without using it, In addition, the sound output from the sound output unit 120 is configured to be audible. For example, a separate omnidirectional speaker may be provided on the side surface of the lower housing 100B opposite to the surface on which the operation unit 140 is provided. Good.
[0024] 送受信部 130は、アンテナ 130Aを備えている。また、送受信部 130は、処理部 18 0の制御により、音声入力部 150から入力された利用者の音声情報を、アンテナ 130 Aを介して、例えば通信先となる相手話者の携帯電話や一般電話機器などの通信機 器に送信する。また、相手話者の通信機器などカゝら送信される音声情報をアンテナ 1 30Aを介して受信し、処理部 180に出力する。さらに、送受信部 130は、処理部 180 の制御により、例えばネットワーク上の他のサーノから情報を受信して処理部 180に 出力する。また送受信部 130は、処理部 180の制御により、処理部 180から入力され た情報を所定のサーバや端末機器に送信する。  [0024] The transceiver 130 includes an antenna 130A. Further, the transmission / reception unit 130 controls the voice information of the user input from the voice input unit 150 through the antenna 130A, for example, by the control of the processing unit 180. Send to a communication device such as a telephone device. Also, audio information transmitted from the other party's communication device or the like is received via the antenna 130A and output to the processing unit 180. Further, the transmission / reception unit 130 receives information from, for example, another sano on the network and outputs the information to the processing unit 180 under the control of the processing unit 180. Further, the transmission / reception unit 130 transmits information input from the processing unit 180 to a predetermined server or terminal device under the control of the processing unit 180.
[0025] 操作部 140は、各種操作ボタンや操作つまみなどを有している。これら操作ボタン や操作つまみの入力操作の内容としては、例えば相手話者の携帯電話に電話をか ける際の相手番号の入力、または相手話者の登録情報の閲覧操作、通話の開始や 終了操作、他サーバから情報を取得する設定などが例示できる。そして、操作部 140 は、利用者の入力操作により、所定の信号を処理部 180へ適宜出力する。  The operation unit 140 has various operation buttons and operation knobs. The contents of the input operation of these operation buttons and control knobs include, for example, input of the other party's number when making a call to the other party's mobile phone, operation for browsing the registered information of the other party, and start / end operation of the call A setting for acquiring information from another server can be exemplified. Then, the operation unit 140 appropriately outputs a predetermined signal to the processing unit 180 by a user input operation.
[0026] 音声入力部 150は、下部筐体 100B側に設けられ、音声を入力可能なマイクを備 えている。この音声入力部 150は、例えば、操作部 140の操作により指向性および無 指向性を切り替え可能で、例えば利用者が携帯電話 100を手放した状態で利用す る際にも、音声入力部 150から音声を入力可能となるように構成されている。  [0026] The audio input unit 150 is provided on the lower housing 100B side and includes a microphone capable of inputting audio. For example, the voice input unit 150 can switch between directivity and non-directionality by operating the operation unit 140. For example, the voice input unit 150 can be switched from the voice input unit 150 when the user uses the cellular phone 100 without using the mobile phone 100. It is configured to be able to input voice.
[0027] 記憶手段 160は、例えば、図 3に示すような登録情報テーブル 10や、図 5に示すよ うな第 1画像テンプレートテーブル 20や、図 6に示すような第 2画像テンプレートテー ブル 30や、図 7に示すような感情認識情報テーブル 40などを格納、すなわち読み出 し可能に記憶する。そして、記憶手段 160には、登録情報テーブル 10が記録される 登録情報記憶領域と、第 1画像テンプレートテーブル 20が記録される第 1画像記憶 領域と、第 2画像テンプレートテーブル 30が記憶される第 2画像記憶領域と、感情認 識情報テーブル 40が記憶される感情認識情報記憶領域と、などを備えている。なお 、記憶手段 160は、これらの 4つの記憶領域以外にも他の情報を記録する記録領域 を備えた構成としてもよい。また、記憶手段に登録情報記憶領域、第 1画像記憶領域 、第 2画像記憶領域、および感情認識情報記憶領域などは、メモリ 170に記憶される 構成としてもよい。そして、記憶手段 160としては、 HD (Hard Disk)、メモリカードなど の記録媒体に読み出し可能に記憶するドライブやドライバなどを備えた構成などが例 示できる。なお、 DVD (Digital Versatile Disc)、光ディスクなどを備えた構成としても よい。 [0027] The storage means 160 includes, for example, the registration information table 10 as shown in FIG. 3, the first image template table 20 as shown in FIG. 5, the second image template table 30 as shown in FIG. The emotion recognition information table 40 shown in FIG. 7 is stored, that is, stored so as to be readable. The storage means 160 stores a registration information storage area in which the registration information table 10 is recorded, a first image storage area in which the first image template table 20 is recorded, and a second image template table 30 in which the second image template table 30 is stored. 2 image storage areas and an emotion recognition information storage area in which the emotion recognition information table 40 is stored. Note that the storage means 160 may be configured to include a recording area for recording other information in addition to these four storage areas. In addition, a registration information storage area, a first image storage area, a second image storage area, an emotion recognition information storage area, and the like are stored in the memory 170 in the storage means. It is good also as a structure. As the storage unit 160, a configuration including a drive or a driver that can be read and stored in a recording medium such as an HD (Hard Disk) or a memory card can be illustrated. A configuration including a DVD (Digital Versatile Disc), an optical disc, or the like may be used.
[0028] 登録情報テーブル 10は、例えば利用者が相手話者と通話する際に、相手話者の 情報を確認する際に利用される情報である。この登録情報テーブル 10は、登録 ID 情報 12と、登録者名情報 13と、電話番号情報 14と、画像情報としての登録画像情 報 15と、登録詳細情報 16と、などを関連付けて 1つのデータとして構成される登録情 報 11を複数記録したテーブル構造に構成されて 、る。  [0028] The registration information table 10 is information used when, for example, the user confirms the information of the other speaker when the user makes a call with the other speaker. This registration information table 10 includes registration ID information 12, registrant name information 13, telephone number information 14, registration image information 15 as image information, registration detailed information 16, and the like, in association with one piece of data. It is configured in a table structure in which a plurality of registration information 11 configured as follows is recorded.
[0029] 登録 HD情報 12は、登録情報 11を特定するための固有情報であり、登録情報 11毎 に設定される情報である。この登録 Iひ f青報 12は、例えば利用者が登録情報テープ ル 10に登録情報 11を登録する際に自動的に連番で付与される。  The registered HD information 12 is unique information for specifying the registered information 11 and is set for each registered information 11. For example, when the user registers the registration information 11 in the registration information table 10, the registration I blueprint 12 is automatically given as a serial number.
[0030] 登録者名情報 13は、登録 HD情報 12にて特定される登録情報 11として登録される 登録者の名前に関する情報である。この登録者名情報 13は、例えばテキスト形式に て記録される。  The registrant name information 13 is information relating to the name of the registrant registered as the registration information 11 specified by the registration HD information 12. The registrant name information 13 is recorded in a text format, for example.
[0031] 電話番号情報 14は、登録 HD情報 12にて特定される登録情報 11として登録される 登録者の携帯電話の電話番号、または一般電話の電話番号に関する情報である。 この電話番号情報 14は、例えば登録者から利用者の携帯電話 100に着信があった 場合や、登録者に利用者が発信する場合に参照される情報である。  [0031] The telephone number information 14 is information related to the telephone number of the registrant's mobile phone registered as the registration information 11 specified by the registration HD information 12, or the telephone number of a general telephone. The telephone number information 14 is information that is referred to when, for example, an incoming call is received from the registrant to the user's mobile phone 100 or when the user makes a call to the registrant.
[0032] 登録画像情報 15は、登録 HD情報 12にて特定される登録情報 11として登録される 登録者から利用者の携帯電話 100に着信があった場合や、利用者が登録者に発信 する際にディスプレイ 110に表示させる画像に関する画像情報である。この登録画像 情報 15の画像は、図 4に示すように、例えば登録者の正面顔画像を所定のフレーム 内に収納されるように撮影した画像である。図 4において、この登録画像情報 15の登 録画像 50は、登録者の顔がフレーム 51内に収納される状態で撮影された画像情報 である。このフレーム 51は、 5本の略平行のフレーム線 52により分割され、分割され た領域にそれぞれ眉毛、目、鼻、口が配置されるように登録者の顔画像が配置される 。また、眉毛が配置されるフレーム領域 Aの、フレーム 51の中央線力も左右方向の所 定寸法だけ離れた位置までは、それぞれ右眉毛が配置される右眉領域 53、左眉毛 が配置される左眉領域 54となっている。さらに、口が配置されるフレーム領域 Dの、フ レーム 51の中央線力も左右方向に所定寸法移動した位置で囲まれる領域は、口が 配置される口領域 55となっている。なお、図 4において、フレーム 51およびフレーム 線 52を図示したが、これは、例えば携帯電話 100のカメラ部 101で登録者を撮影す る際に補助的に表示されるものである。したがって、実際に登録された登録画像情報 15の登録画像 50には、フレーム 51およびフレーム線 52は表示されない。このような 登録画像 50は、例えば携帯電話 100に設けられた撮影用のカメラ部 101にて撮影さ れて登録されたものであってもよぐ別途市販のデジタルカメラ機器などで撮影したも のを登録したものであってもよ!/、。 [0032] The registered image information 15 is registered as the registered information 11 specified by the registered HD information 12. When the registrant receives an incoming call to the mobile phone 100 of the user, the registered image information 15 is transmitted to the registrant. This is image information related to an image to be displayed on the display 110 at this time. As shown in FIG. 4, the image of the registered image information 15 is an image obtained by capturing a front face image of a registrant, for example, so as to be stored in a predetermined frame. In FIG. 4, a registered image 50 of the registered image information 15 is image information taken in a state where the registrant's face is housed in a frame 51. The frame 51 is divided by five substantially parallel frame lines 52, and the registrant's face image is arranged so that the eyebrows, eyes, nose, and mouth are arranged in the divided areas. In addition, the center line force of frame 51 in frame area A where the eyebrows are placed is also in the horizontal direction. Up to a position separated by a fixed dimension, there are a right eyebrow region 53 where the right eyebrows are arranged and a left eyebrow region 54 where the left eyebrows are arranged. Further, a region surrounded by a position where the central line force of the frame 51 in the frame region D where the mouth is disposed is also moved by a predetermined dimension in the left-right direction is a mouth region 55 where the mouth is disposed. In FIG. 4, the frame 51 and the frame line 52 are illustrated, but these are supplementarily displayed when the registrant is photographed by the camera unit 101 of the mobile phone 100, for example. Therefore, the frame 51 and the frame line 52 are not displayed in the registered image 50 of the registered image information 15 that is actually registered. Such a registered image 50 may be, for example, a photograph taken with a camera unit 101 for photographing provided in the mobile phone 100 and registered with a commercially available digital camera device. Even if you have registered! /.
[0033] 登録詳細情報 16は、登録 HD情報 12にて特定される登録情報 11として登録される 登録者の詳細事項に関する情報である。この登録詳細情報 16の記録される詳細事 項としては、例えば登録者の住所や仕事先などの情報、登録者の性別、登録者から 着信があった場合のアクションに関する情報、登録者のメールアドレス、登録者の音 声に関する音声基礎情報などが挙げられる。  The registration detail information 16 is information relating to the detailed items of the registrant registered as the registration information 11 specified by the registration HD information 12. The detailed information recorded in the registration details information 16 includes, for example, information such as the registrant's address and work place, gender of the registrant, information on the action when there is an incoming call from the registrant, and the email address of the registrant. And basic voice information on the registrant's voice.
[0034] 第 1画像テンプレートテーブル 20は、処理部 180にて登録画像情報 15に記録され る登録画像 50を変形編集する際の右眉領域 53および左眉領域 54の眉画像の変形 パターンを記録した情報である。この第 1画像テンプレートテーブル 20は、図 5に示 すように第 1画像 HD情報 22と、第 1画像基本情報 23と、変形量情報および変形方向 情報としての第 1画像パターン情報 24と、対応感情情報 25と、などを関連付けた第 1 画像テンプレート情報 21を複数記録したテーブル構造に構築されている。  [0034] The first image template table 20 records the deformation patterns of the eyebrow images in the right eyebrow region 53 and the left eyebrow region 54 when the processing unit 180 deforms and edits the registered image 50 recorded in the registered image information 15. Information. This first image template table 20 corresponds to the first image HD information 22, the first image basic information 23, the first image pattern information 24 as deformation amount information and deformation direction information, as shown in FIG. It is constructed in a table structure in which a plurality of first image template information 21 in which emotion information 25 and the like are associated are recorded.
[0035] 第 1画像 HD情報 22は、第 1画像テンプレート情報 21を特定する固有情報である。  The first image HD information 22 is unique information that identifies the first image template information 21.
この第 1画像 HD情報 22は、各第 1画像テンプレート情報 21にそれぞれ異なった固有 の情報であり、例えば、数字の連番が記録されている。  The first image HD information 22 is unique information different from each first image template information 21, and for example, a serial number is recorded.
[0036] 第 1画像基本情報 23は、第 1画像 HD情報 22にて特定される第 1画像テンプレート 情報 21の基本となる画像情報である。この第 1画像基本情報 23には、例えば平常時 の眉毛の基本画像が記録されている。また、この第 1画像基本情報の基本画像は、 図 4に示すような登録画像 50の右眉領域 53および左眉領域 54の領域形状である矩 形と略同一の大きさの矩形状、または相似関係の矩形状に形成され、この矩形の内 部に例えば平常時の眉毛の画像が略中央に配置されている。 The first image basic information 23 is image information that is the basis of the first image template information 21 specified by the first image HD information 22. In the first image basic information 23, for example, a basic image of normal eyebrows is recorded. Further, the basic image of the first image basic information is a rectangular shape that is the shape of the right eyebrow region 53 and the left eyebrow region 54 of the registered image 50 as shown in FIG. A rectangular shape having the same size as the shape or a rectangular shape having a similar relationship is formed, and an image of a normal eyebrow, for example, is arranged in the center of the rectangle.
[0037] 第 1画像パターン情報 24は、第 1画像 情報 22にて特定される第 1画像テンプレ ート情報 21の第 1画像基本情報 23の基本画像を基本として、所定の感情の状態に おいて第 1画像基本情報 23の眉画像の各部位をどの方向にどれだけ変形させるか を示す変形方向および変形量としての変形率が記録されている。具体的には、この 第 1画像パターン情報 24は、例えば笑顔の状態のパターンを例にすると、眉の略中 心位置を上方側に例えば 10ドット移動させ、眉の両端を下方側に例えば 20ドット移 動させるといった情報が記録されている。なお、この第 1画像パターン情報 24は、例 えば右眉領域 53および左眉領域 54の矩形の上下寸法に対して、眉の各部位を上 方向に移動させる割合および下方向移動させる割合が記録されているものであって もよぐ変形量および変形方向がベクトル形式で記録されて 、るものであってもよ 、。  [0037] The first image pattern information 24 is based on the basic image of the first image basic information 23 of the first image template information 21 specified by the first image information 22, and is in a predetermined emotional state. In addition, a deformation direction and a deformation rate as a deformation amount indicating how much and how much each part of the eyebrow image of the first image basic information 23 is deformed are recorded. Specifically, for example, in the case of a smile pattern, the first image pattern information 24 moves the approximate center position of the eyebrows upward, for example, 10 dots, and both ends of the eyebrows downward, for example, 20 Information such as dot movement is recorded. In the first image pattern information 24, for example, the ratio of moving each part of the eyebrow upward and the ratio moving downward with respect to the rectangular vertical dimension of the right eyebrow region 53 and the left eyebrow region 54 is recorded. The amount of deformation and the direction of deformation may be recorded in vector format, and may be.
[0038] 対応感情情報 25は、第 1画像 HD情報 22にて特定される第 1画像テンプレート情報 21の第 1画像パターン情報 24が、どのような感情を示す変化率であるかを示す情報 である。この対応感情情報 25には、例えば「平常時」や「無表情」の状態を示す「0」、 「笑顔」の状態を示す「1」、怒り状態を示す「2」などといった、各感情を数値として記 録されている。なお、これらの感情に関する情報が例えば「平常時」「笑顔」「怒り」な どと!/、つたテキスト形式で記録されて 、るものであってもよ!/、。  [0038] Corresponding emotion information 25 is information indicating a change rate indicating what kind of emotion the first image pattern information 24 of the first image template information 21 specified by the first image HD information 22 is. is there. The corresponding emotion information 25 includes, for example, “0” indicating “normal” or “no expression”, “1” indicating “smile”, “2” indicating anger, etc. It is recorded as a numerical value. Information about these emotions may be recorded in a text format such as “normal”, “smile”, “anger”, etc.! /, Etc.
[0039] 第 2画像テンプレートテーブル 30は処理部 180にて登録画像情報 15に記録される 登録画像 50を変形編集する際の口領域 55の口画像の変形パターンを記録した情 報である。この第 2画像テンプレートテーブル 30は、図 6に示すように第 2画像 情 報 32と、第 2画像基本情報 33と、変形量情報および変形方向情報としての第 2画像 ノターン情報 34と、対応感情情報 35と、などを関連付けた第 2画像テンプレート情 報 31を複数記録したテーブル構造に構築されて 、る。  The second image template table 30 is information in which a deformation pattern of the mouth image of the mouth region 55 when the registered image 50 recorded in the registered image information 15 is modified and edited by the processing unit 180 is recorded. As shown in FIG. 6, the second image template table 30 includes second image information 32, second image basic information 33, second image notation information 34 as deformation amount information and deformation direction information, and corresponding emotions. It is constructed in a table structure that records a plurality of second image template information 31 that associates information 35 and the like.
[0040] 第 2画像 HD情報 32は、第 2画像テンプレート情報 31を特定する固有情報である。  The second image HD information 32 is unique information that identifies the second image template information 31.
この第 2画像 HD情報 32は、各第 2画像テンプレート情報 31にそれぞれ異なった固有 の情報であり、例えば、数字の連番が記録されている。  The second image HD information 32 is unique information different from each of the second image template information 31, and for example, a serial number is recorded.
[0041] 第 2画像基本情報 33は、第 2画像 HD情報 32にて特定される第 2画像テンプレート 情報 31の基本となる画像情報である。この第 2画像基本情報 33には、例えば平常時 の口の基本画像が記録されている。この基本画像は、図 4に示す登録画像の口領域 55の領域形状である矩形と略同一の大きさの矩形状、または相似関係の矩形状に 形成され、この矩形の内部に例えば平常時の口の画像が略中央に配置されている。 [0041] The second image basic information 33 is a second image template specified by the second image HD information 32. This is image information that is the basis of information 31. In the second image basic information 33, for example, a basic image of the mouth at normal times is recorded. This basic image is formed in a rectangular shape that is substantially the same size as the rectangular shape of the mouth area 55 of the registered image shown in FIG. 4 or a similar rectangular shape. The mouth image is arranged in the approximate center.
[0042] 第 2画像パターン情報 34は、第 2画像 情報 32にて特定される第 2画像テンプレ ート情報 31の第 2画像基本情報 33を基本として、所定の感情の状態において第 2画 像基本情報 33の口画像の各部位をどの方向にどれだけ移動させるかを示す変形方 向および変形量としての変化率が記録されている。例えば、笑顔の状態を例示する と、口画像の両端を上方側に例えば 20ドット移動させると 、つた情報が記録されて!ヽ る。なお、この変化率としては、例えば口領域 55の矩形の上下寸法または左右方向 に対して、口の各部位を上方向に移動させる割合、下方向移動させる割合、左右方 向に移動させる割合などが記録されて!、るものであってもよぐ例えばベクトル形式で 記録されて 、るものであってもよ!/、。  The second image pattern information 34 is based on the second image basic information 33 of the second image template information 31 specified by the second image information 32, and the second image pattern information 34 in a predetermined emotional state. The deformation direction and the rate of change as the amount of deformation indicating how much each part of the mouth image of the basic information 33 is moved in which direction are recorded. For example, in the case of a smiling face, if both ends of the mouth image are moved upward, for example, 20 dots, the information is recorded. The rate of change may be, for example, the rate of moving each part of the mouth upward, the rate of moving downward, the rate of moving left or right, etc. May be recorded! For example, it may be recorded in vector format! /.
[0043] 対応感情情報 35は、第 1画像テンプレート情報 21と同様に、第 2画像 HD情報 32に て特定される第 2画像テンプレート情報 31の第 2画像パターン情報 34が、どのような 感情を示す変化率であるかを示す情報である。この対応感情情報 35には、例えば「 平常時」や「無表情」の状態を示す「0」、「笑顔」の状態を示す「1」、怒り状態を示す「 2」などといった、各感情を数値として記録してもよぐテキスト形式で記録されている ものであってもよい。  [0043] Corresponding emotion information 35 is similar to the first image template information 21, and the second image pattern information 34 of the second image template information 31 specified by the second image HD information 32 indicates what kind of emotion. It is the information which shows whether it is the change rate to show. The corresponding emotion information 35 includes, for example, “0” indicating “normal” or “no expression”, “1” indicating “smile”, “2” indicating anger, etc. It may be recorded in a text format which may be recorded as a numerical value.
[0044] 感情認識情報テーブル 40は、処理部 180にて、相手話者の音声情報の状態、例 えば声の強弱やテンポなどから、音声を発生した話者の感情を認識させる際に利用 されるデータ群である。この感情認識情報テーブル 40は、感情 情報 42と、音声状 態変化情報としての音声パターン情報 43と、感情情報 44と、などを関連付けて 1つ のデータとして構築される感情認識情報 41を複数記録したテーブル構造に構成され ている。  [0044] The emotion recognition information table 40 is used when the processing unit 180 recognizes the emotion of the speaker who generated the voice from the state of the voice information of the other speaker, for example, the strength or tempo of the voice. Data group. This emotion recognition information table 40 records a plurality of emotion recognition information 41 constructed as one piece of data by associating emotion information 42, voice pattern information 43 as voice state change information, emotion information 44, etc. The table structure is configured.
[0045] 感情 情報 42は、感情認識情報 41を特定する固有情報である。この感情 情報 The emotion information 42 is unique information that identifies the emotion recognition information 41. This emotion information
42は、感情認識情報 41毎に設けられ、それぞれ異なる情報を有している。 42 is provided for each emotion recognition information 41 and has different information.
[0046] 音声パターン情報 43は、音声の音声状態に関する情報である。具体的には、この 音声パターン情報 43には、例えば「こら」や「ははは」などといった感情を表す語句に 関する語句情報、音声の強弱状態に関する強弱情報、音声の高低状態に関する高 低情報、音声のテンポに関するテンポ情報などが記録されて 、る。 [0046] The voice pattern information 43 is information relating to the voice state of the voice. Specifically, this The voice pattern information 43 includes phrase information related to emotional phrases such as “Kora” and “Hahaha”, strength information related to the strength of the voice, height information related to the voice high / low state, and tempo related to the voice tempo. Information is recorded.
[0047] 感情情報 44は、感情 Iひ f青報 42にて特定される感情認識情報 41の音声パターン 情報 43に対応する感情に関する情報である。この感情情報 44には、例えば音声パ ターン情報 43に音声の強弱状態が強ぐ語句情報として「ははは」という語句が記録 されている場合、例えば大げさに笑っている状態を示す旨の HD情報である例えば「1 」が記録される。なお、感情情報 44は、例えば「笑顔」「怒り」などといったテキスト形 式にて記録されていてもよい。また、この感情情報 44は、前述した第 1画像テンプレ ート情報 21の対応感情情報 25や、第 2画像テンプレート情報 31の対応感情情報 35 に関連付けられている。  [0047] The emotion information 44 is information related to the emotion corresponding to the voice pattern information 43 of the emotion recognition information 41 specified by the emotion I and the blueprint 42. In this emotion information 44, for example, if the phrase “hahaha” is recorded as the phrase information in which the voice strength is strong in the voice pattern information 43, for example, HD indicating that the voice is laughing loudly. Information such as “1” is recorded. The emotion information 44 may be recorded in a text format such as “smile” or “anger”. The emotion information 44 is associated with the corresponding emotion information 25 of the first image template information 21 and the corresponding emotion information 35 of the second image template information 31 described above.
[0048] メモリ 170は、操作部 140で入力操作される設定事項などを適宜読み出し可能に 記憶する。また、メモリ 170は、携帯電話 100全体を動作制御する OS (Operating Sys tem)上に展開される各種プログラムなどを記憶している。なお、メモリ 170としては、 H D (Hard Disk)や光磁気ディスクなどの記録媒体に読み取り可能に記憶するドライブ やドライバなどを備えた構成としてもょ 、。  [0048] The memory 170 stores setting items input and operated by the operation unit 140 so that they can be read as appropriate. The memory 170 stores various programs developed on an OS (Operating System) that controls the operation of the entire mobile phone 100. The memory 170 may be configured to include a drive and a driver that are readable and stored in a recording medium such as an HD (Hard Disk) or a magneto-optical disk.
[0049] 処理部 180は、図示しない各種入出力ポート、例えばディスプレイ 110が接続され る表示制御ポート、音声出力部 120が接続される音声出力制御ポート、送受信部 13 0が接続される送受信ポート、操作部 140が接続される入力ポート、音声入力部 150 が接続される音声入力制御ポート、記憶手段 160が接続される記憶ポート、メモリ 17 0が接続されるメモリポートなどを有する。そして、処理部 180は、各種プログラムとし て、図 8に示すように、発呼手段としても機能する発着認識手段 181と、相手話者認 識手段 182と、音声情報認識手段としての音声認識手段 183と、音声状態認識手段 としての音声変化判定手段 184と、画像情報認識手段、および画像変形処理手段と しても機能する画像処理手段 185と、表示制御手段 186と、撮影手段 187と、などを 備えている。  [0049] The processing unit 180 includes various input / output ports (not shown), such as a display control port to which the display 110 is connected, an audio output control port to which the audio output unit 120 is connected, a transmission / reception port to which the transmission / reception unit 130 is connected, An input port to which the operation unit 140 is connected, a voice input control port to which the voice input unit 150 is connected, a storage port to which the storage unit 160 is connected, a memory port to which the memory 170 is connected, and the like. Then, the processing unit 180, as various programs, as shown in FIG. 8, an incoming / outgoing recognition means 181 that also functions as a calling means, a partner speaker recognition means 182 and a voice recognition means as voice information recognition means. 183, sound change determination means 184 as sound state recognition means, image information recognition means, image processing means 185 that also functions as image deformation processing means, display control means 186, photographing means 187, etc. It is equipped with.
[0050] 発着認識手段 181は、他の通信機器から携帯電話 100への着信、および携帯電 話 100から他の通信機器への発着を認識する。具体的には、発着認識手段 181は、 送受信部 130を制御して、他の通信機器から通話を要求する旨の着信情報を受信 する。また、発着認識手段 181は、着信情報を認識すると、音声出力部 120を制御し て、着信情報を受信した旨を音声、警告音、報知音などの着信音を出力させて利用 者に報知する。また、発着認識手段 181は、利用者の操作部 140の入力操作により 、所定の発信先通信機器に発信する旨の要求情報を認識すると、この発信先通信 機器に通話を要求する旨の発信情報を送信する。 [0050] The arrival / departure recognition means 181 recognizes incoming calls from other communication devices to the mobile phone 100 and arrivals / departures from the mobile phone 100 to other communication devices. Specifically, the arrival and departure recognition means 181 Controls the transmission / reception unit 130 to receive incoming call information requesting a call from another communication device. In addition, when the arrival / departure recognition means 181 recognizes the incoming call information, it controls the audio output unit 120 to output the incoming sound such as voice, warning sound, notification sound, etc. to notify the user that the incoming call information has been received. . Further, when the arrival / departure recognition means 181 recognizes request information indicating that a call is to be made to a predetermined destination communication device by an input operation of the operation unit 140 of the user, the transmission information indicating that a call is requested to the destination communication device. Send.
[0051] 相手話者認識手段 182は、発着認識手段 181にて着信情報を受信したり、発信先 の通信機器に発信情報を送信したりした際に、着信相手および発信相手を認識する 。具体的には、相手話者認識手段 182は、発着認識手段 181にて受信した着信情 報から着信情報を送信した着信相手の電話番号を認識する。そして、登録情報テー ブル 10から、認識した着信相手の電話番号と一致する電話番号情報 14を有する登 録情報 11を認識する。さら〖こ、相手話者認識手段 182は、発着認識手段 181にて送 信された発信情報の発信先電話番号を認識する。そして、登録情報テーブル 10から 発信先電話番号を一致する電話番号情報 14を有する登録情報 11を認識する。そし て、相手話者認識手段 182は、認識した登録情報 11を適宜読み出し可能にメモリ 1 70に記憶する。 [0051] When the incoming / outgoing recognition means 181 receives the incoming call information or transmits the outgoing call information to the destination communication device, the called party recognition means 182 recognizes the incoming call partner and the outgoing call partner. Specifically, the partner speaker recognition means 182 recognizes the telephone number of the called party who has sent the incoming information from the incoming information received by the arrival / reception recognition means 181. Then, the registration information 11 having the telephone number information 14 that matches the recognized telephone number of the called party is recognized from the registration information table 10. Furthermore, the other speaker recognition means 182 recognizes the destination telephone number of the transmission information transmitted by the arrival / reception recognition means 181. Then, the registration information 11 having the telephone number information 14 that matches the destination telephone number is recognized from the registration information table 10. Then, the partner speaker recognition means 182 stores the recognized registration information 11 in the memory 170 so that it can be read out appropriately.
[0052] 音声認識手段 183は、送受信部 130にて受信した相手話者の音声情報力も相手 話者の音声の状態を認識し、相手話者の標準音声状態情報としての音声基礎情報 として認識する。具体的には、音声認識手段 183は、送受信部 130を介して入力さ れる相手話者の音声情報を認識する。そして、音声情報の音声の状態から相手話者 の音声パターン、すなわち音声の高低、音声の強弱、音声を発するテンポなどを判 別し、音声基礎情報として認識する。なお、音声認識手段 183は、登録情報 11の登 録詳細情報 16に記載された内容に基づいて、相手話者の性別や基本音声情報な どを認識し、音声基礎情報とする構成としてもよい。  [0052] The voice recognition means 183 also recognizes the voice status of the other speaker as received by the transmitting / receiving unit 130, and recognizes the voice status of the other speaker as voice basic information as standard voice status information of the other speaker. . Specifically, the voice recognition means 183 recognizes the voice information of the partner speaker input via the transmission / reception unit 130. Then, the voice pattern of the other speaker is determined from the voice state of the voice information, that is, the voice level, the voice strength, the tempo at which the voice is emitted, and the like are recognized as the voice basic information. The voice recognition means 183 may be configured to recognize the other speaker's gender, basic voice information, etc. as voice basic information based on the contents described in the registration details information 16 of the registration information 11. .
[0053] 音声変化判定手段 184は、送受信部 130にて受信した音声情報の音声が、音声 認識手段 183にて認識した相手話者の音声基礎情報カゝら変化した場合に、変化し た音声情報力 相手話者の感情の変化を認識する。具体的には、音声変化判定手 段 184は、相手話者の音声情報の音声状態が変化すると、その音声状態の音声パ ターンを検出する。そして、例えば音声基礎情報の音声パターンに比べて検出した 音声情報の音声パターンの高低や強弱、テンポなどが異なる場合、感情認識情報テ 一ブル 40に基づ 、て、受信した音声の音声パターンに略一致する音声パターン情 報 42を認識し、この音声パターン情報 42に対応する感情認識情報 41を認識する。 また、音声変化判定手段 184は、受信した音声情報の音声に、例えば「こら」や「は はは」などと 、つた感情を表す語句情報が含まれることを認識すると、これらの語句情 報を音声パターン情報 43に含む感情認識情報 41を認識する。また、音声変化判定 手段 184は、認識した感情情報 44を含む感情認識情報 41を適宜読み出し可能にメ モリ 170に記憶する。 [0053] The voice change determination means 184 changes the voice when the voice of the voice information received by the transmission / reception unit 130 changes from the voice basic information of the other speaker recognized by the voice recognition means 183. Information skills Recognize changes in emotions of the other speaker. Specifically, when the voice state of the voice information of the other speaker changes, the voice change determination unit 184 determines the voice parameter of the voice state. Detect turn. For example, if the detected voice pattern is different in height, strength, tempo, etc. compared to the voice pattern of the basic voice information, the voice pattern of the received voice is changed based on the emotion recognition information table 40. The voice pattern information 42 that substantially matches is recognized, and the emotion recognition information 41 corresponding to the voice pattern information 42 is recognized. When the voice change determination means 184 recognizes that the voice of the received voice information includes phrase information representing the emotion, such as “Kora”, “Hahaha”, etc., these phrases information is displayed. The emotion recognition information 41 included in the voice pattern information 43 is recognized. The voice change determination means 184 stores the emotion recognition information 41 including the recognized emotion information 44 in the memory 170 so that it can be read out as appropriate.
[0054] 画像処理手段 185は、相手話者に対応する登録情報 11の登録画像情報 15を認 識する。そして、音声変化判定手段 184にて認識した感情認識情報 41の感情情報 4 4に基づいて、登録画像情報 15の画像を変化させる。具体的には、画像処理手段 1 85は、登録画像情報 15に記録される登録画像 50の右眉領域 53、左眉領域 54を認 識する。そして、画像処理手段 185は、第 1画像テンプレートテーブル 20に基づいて 、これらの右眉領域 53、左眉領域 54の眉画像に略一致する画像の第 1画像基本情 報 23を有する第 1画像テンプレート情報 21を検索する。これには、例えば、第 1画像 基本情報 23に記録された眉画像を登録画像情報 15の右眉領域 53、左眉領域 54 に重ね合わせ、例えば眉部分の重なり度が最も大きい第 1画像基本情報 23を検索 する。次に、画像処理手段 185は、検索して絞り込まれた第 1画像テンプレート情報 21のうちから、音声変化判定手段 184にて認識した感情情報 44に対応する対応感 情情報 25を有する第 1画像テンプレート情報 21を認識する。そして、この第 1画像テ ンプレート情報 21の第 1画像パターン情報 24に記録される変化率に応じて登録画 像情報 15の登録画像 50の右眉領域 53および左眉領域 54の眉の形状を変形させる  [0054] The image processing means 185 recognizes the registered image information 15 of the registered information 11 corresponding to the partner speaker. Then, based on the emotion information 44 of the emotion recognition information 41 recognized by the voice change determination means 184, the image of the registered image information 15 is changed. Specifically, the image processing means 185 recognizes the right eyebrow region 53 and the left eyebrow region 54 of the registered image 50 recorded in the registered image information 15. Then, based on the first image template table 20, the image processing means 185 has a first image having first image basic information 23 of an image that substantially matches the eyebrow images of the right eyebrow region 53 and the left eyebrow region 54. Search for template information 21. For example, the eyebrow image recorded in the first image basic information 23 is overlaid on the right eyebrow region 53 and the left eyebrow region 54 of the registered image information 15, for example, the first image basics with the highest degree of overlap of the eyebrow portions. Search for information 23. Next, the image processing means 185 includes a first image having corresponding emotion information 25 corresponding to the emotion information 44 recognized by the voice change determination means 184 from the first image template information 21 narrowed down by searching. Recognizes template information 21. The shape of the eyebrows of the right eyebrow region 53 and the left eyebrow region 54 of the registered image 50 of the registered image information 15 according to the change rate recorded in the first image pattern information 24 of the first image template information 21 Transform
[0055] また、画像処理手段 185は、右眉領域 53および左眉領域 54の眉画像の変形処理 と同様に、口画像も変形させる。すなわち、画像処理手段 185は、登録画像 50の口 領域 55を認識する。そして、第 2画像テンプレートテーブル 30に基づいて、このロ領 域 55の口画像に略一致する画像の第 2画像基本情報 33を有する第 2画像テンプレ ート情報 31を検索する。これには、例えば、第 2画像基本情報 33に記録された口画 像を登録画像情報 15の口領域 55に重ね合わせ、例えば口部分の重なり度が最も大 きい第 2画像基本情報 33を検索する。次に、画像処理手段 185は、検索して絞り込 まれた第 2画像テンプレート情報 31のうちから、音声変化判定手段 184にて認識した 感情情報 44に対応する対応感情情報 35を有する第 2画像テンプレート情報 31を認 識する。そして、この第 2画像テンプレート情報 31の第 2画像パターン情報 34に記録 される変化率に応じて登録画像情報 15の登録画像 50の口領域 55の口の形状を変 形させる。また、画像処理手段 185は、眉および口を変形させた登録画像 50を感情 変形画像として適宜読み出し可能にメモリ 170に記憶する。 In addition, the image processing means 185 also deforms the mouth image in the same manner as the deformation process of the eyebrow image in the right eyebrow region 53 and the left eyebrow region 54. That is, the image processing means 185 recognizes the mouth area 55 of the registered image 50. Then, based on the second image template table 30, the second image template having the second image basic information 33 of the image that substantially matches the mouth image in the region 55. Search for port information 31. To do this, for example, the mouth image recorded in the second image basic information 33 is overlaid on the mouth area 55 of the registered image information 15, and for example, the second image basic information 33 with the largest degree of overlap of the mouth portion is searched. To do. Next, the image processing means 185 includes a second image having corresponding emotion information 35 corresponding to the emotion information 44 recognized by the voice change determination means 184 from the second image template information 31 searched and narrowed down. Recognize template information 31. Then, the mouth shape of the mouth region 55 of the registered image 50 of the registered image information 15 is changed according to the change rate recorded in the second image pattern information 34 of the second image template information 31. In addition, the image processing means 185 stores the registered image 50 in which the eyebrows and the mouth are deformed in the memory 170 so as to be appropriately readable as an emotion deformed image.
[0056] 表示制御手段 186は、ディスプレイ 110を制御して登録情報 11の登録画像情報 1 5の画像を表示領域に表示させる制御をする。具体的には、表示制御手段 186は、 発着認識手段 181にて利用者による操作部 140の操作により相手話者の通信機器 に発信する旨の発信要求情報を認識されると、この相手話者の登録情報 11に登録 画像情報 15に記録される登録画像 50をディスプレイ 110に表示させる制御をする。 また、表示制御手段 186は、発着認識手段 181にて携帯電話 100に他の携帯電話 や一般電話などの通信機器力も着信があることが認識され、相手話者認識手段 182 にて送信先である相手話者の登録情報 11の登録画像情報 15が認識されると、この 登録画像情報 15の登録画像 50をディスプレイ 110に表示させる制御をする。  [0056] The display control means 186 controls the display 110 to display the image of the registered image information 15 of the registration information 11 in the display area. Specifically, the display control means 186, when the arrival / reception recognition means 181 recognizes the call request information indicating that the call is sent to the communication apparatus of the other speaker by the operation of the operation unit 140 by the user, Control is performed to display the registered image 50 recorded in the registered image information 15 on the display 110. In addition, the display control means 186 recognizes that the mobile phone 100 has an incoming communication device such as another mobile phone or a general telephone at the arrival / departure recognition means 181 and the destination speaker recognition means 182 is the transmission destination. When the registered image information 15 of the registered information 11 of the other speaker is recognized, control is performed to display the registered image 50 of the registered image information 15 on the display 110.
[0057] また、表示制御手段 186は、画像処理手段 185にて画像処理された感情変形画像 を、例えば図 9 Aな 、し図 9Cに示すようにディスプレイ 110の表示領域に表示させる 制御をする。ここで図 9Aは、音声変化判定手段 184にて通話の相手話者の音声情 報の音声状態が例えば笑っている状態であると判断され、その感情情報 44に応じて 画像処理手段 185が登録画像 50を変形させた画像である。また、図 9Bは、音声変 化判定手段 184にて通話の相手話者の音声情報の音声状態の音声パターンが例 えば「怒り状態」であると判断され、その感情情報 44に応じて画像処理手段 185が登 録画像 50を変形させた画像である。図 9Cは、音声変化判定手段 184にて通話の相 手話者の音声情報の音声パターンが例えば「悲しみ状態」と判断され、その感情情 報 44に応じて画像処理手段 185が登録画像 50を変形させた画像である。さらに、表 示制御手段 186は、カメラ部 101から入力された映像を画像情報として処理し、ディ スプレイ 110の表示領域に表示させる制御をする。 [0057] Further, the display control means 186 performs control to display the emotion deformed image processed by the image processing means 185 in the display area of the display 110 as shown in FIG. 9A or FIG. 9C, for example. . Here, in FIG. 9A, the voice change determination means 184 determines that the voice state of the other party's voice information is laughing, for example, and the image processing means 185 is registered according to the emotion information 44. The image 50 is a deformed image. In FIG. 9B, the voice change determination means 184 determines that the voice pattern of the voice information of the other party's call is “angry”, for example, and performs image processing according to the emotion information 44. Means 185 is an image obtained by deforming the registered image 50. FIG. 9C shows that the voice pattern of the caller's voice information is determined to be, for example, “sad state” by the voice change judging means 184, and the image processing means 185 transforms the registered image 50 according to the emotion information 44. It is the image made to do. In addition, the table The display control means 186 controls the video input from the camera unit 101 as image information and displays it on the display area of the display 110.
[0058] 撮影手段 187は、利用者の操作部 140の入力操作により、画像を撮影する旨の要 求情報を認識すると、カメラ部 101を制御して画像撮影可能な状態とする。また、ディ スプレイ 110にカメラ部 101の撮影範囲の映像を表示させる制御をする。さらに、撮 影手段 187は、利用者の設定入力により画像処理手段 185にて画像処理可能な画 像を撮影する旨の要求情報を認識すると、図 3に示すような、フレーム 51およびフレ ーム線 52をディスプレイ 110に表示させる制御をする。そして、利用者の入力操作に より撮影する旨の要求情報を認識すると、カメラ部 101の撮影範囲に入っている映像 を撮影し、画像情報として記憶手段 160に記憶する。  When the image capturing unit 187 recognizes request information for capturing an image by an input operation of the operation unit 140 by the user, the image capturing unit 187 controls the camera unit 101 to be in a state in which an image can be captured. In addition, the display 110 is controlled to display an image of the shooting range of the camera unit 101. Furthermore, when the image capturing unit 187 recognizes request information indicating that an image that can be processed by the image processing unit 185 is input by a user's setting input, the frame 51 and the frame as illustrated in FIG. Control the display of line 52 on display 110. When the request information indicating that the image is to be captured is recognized by the user's input operation, the image within the imaging range of the camera unit 101 is captured and stored in the storage unit 160 as image information.
[0059] 〔携帯電話の動作〕  [0059] [Operation of mobile phone]
(撮影処理)  (Shooting process)
次に、上記携帯電話 100の動作として、変形可能画像撮影処理について、図 10に 基づいて説明する。図 10は、携帯電話 100の撮影処理を示すフローチャートである  Next, as an operation of the mobile phone 100, a deformable image photographing process will be described with reference to FIG. FIG. 10 is a flowchart showing the photographing process of the mobile phone 100.
[0060] まず、携帯電話 100の処理部 180は、利用者の操作部 140の操作によりカメラ部 1 01にて所定の映像の撮影を要求する撮影要求情報が設定入力されると、この撮影 情報要求情報を認識して (ステップ S101)、撮影用のカメラ部 101を起動させる (ステ ップ S 102)。 [0060] First, the processing unit 180 of the mobile phone 100 receives the shooting request information for requesting shooting of a predetermined video by the camera unit 101 by the operation of the operation unit 140 of the user. Recognizing the request information (step S101), the camera unit 101 for shooting is activated (step S102).
[0061] 次に、処理部 180は、表示制御手段 186にてディスプレイ 110を制御させ、変形可 能画像、すなわち相手話者と通話中に相手話者の音声状態に応じて画像処理を実 施するための画像の撮影である力否かを問う表示画面を表示させる (ステップ S 103) 。そして、このステップ S 103において、利用者が変形可能画像の撮影をしない旨の 情報を設定入力すると、処理部 180の撮影手段 187は、表示制御手段 186にカメラ 部 101の撮影範囲内に入る映像をディスプレイ 110の表示領域に表示させる制御を させる。  [0061] Next, the processing unit 180 controls the display 110 with the display control means 186, and performs image processing according to the deformable image, that is, the voice state of the other speaker during a call with the other speaker. A display screen is displayed asking whether or not the user has the power to capture an image (step S103). In step S 103, when the user inputs and inputs information indicating that the deformable image is not shot, the shooting unit 187 of the processing unit 180 causes the display control unit 186 to enter a video that falls within the shooting range of the camera unit 101. Is controlled to display in the display area of display 110.
[0062] 一方、ステップ S103において、処理部 180は、利用者の設定入力により変形可能 画像の撮影をする旨の情報を認識すると、撮影手段 187にディスプレイ 110の表示 領域内に図 4に示すようなフレーム 51およびフレーム線 52を表示させる制御をさせ る (ステップ S 104)。そして、利用者が被撮影者の顔をフレーム 51内に入るように、か つ、右眉および左眉がそれぞれ右眉領域 53および左眉領域 54に入り、口が口領域 55に入るようにカメラ部 101の焦点を合わせ、操作部 140にて撮影を実施する旨の 情報を入力すると、処理部 180の撮影手段 187は、ディスプレイ 110に表示されてい る映像を画像情報として取り込む (ステップ S 105)。なお、処理部 180がステップ S1 03にて変形可能画像の撮影ではな ヽ旨の入力を認識して 、る場合も同様に、撮影 手段 187は、ディスプレイ 110に表示される映像を画像情報として取り込む。 On the other hand, in step S103, when the processing unit 180 recognizes information indicating that a deformable image is to be captured by the user's setting input, the display unit 110 displays the display 110 on the photographing unit 187. Control is performed to display the frame 51 and the frame line 52 as shown in FIG. 4 in the area (step S104). Then, so that the user enters the face of the subject within the frame 51, the right eyebrow and the left eyebrow enter the right eyebrow region 53 and the left eyebrow region 54, respectively, and the mouth enters the mouth region 55. When the camera unit 101 is focused and information indicating that shooting is to be performed is input by the operation unit 140, the shooting unit 187 of the processing unit 180 captures the video displayed on the display 110 as image information (step S105). ). Similarly, in the case where the processing unit 180 recognizes the input indicating that the deformable image is not captured in step S103, and the imaging unit 187 also captures the image displayed on the display 110 as image information. .
[0063] そして、処理部 180は、取り込まれた画像を適宜読み出し可能に記憶手段 160に 記憶する(ステップ S106)。また、処理部 180は、ステップ S 103にて変形可能画像を 撮影する旨の設定入力を認識して ヽる場合、取り込んだ変形可能画像を登録情報 1 1に登録する力否かを選択する。そして、利用者の設定入力により所定の登録情報 1 1の登録画像情報 15に登録する旨の情報を認識すると、処理部 180は、取り込んだ 画像情報を利用者の設定入力により指定される所定の登録情報 11の登録画像情報 15として記録する。  [0063] Then, the processing unit 180 stores the captured image in the storage unit 160 so that it can be appropriately read (step S106). Further, when the processing unit 180 recognizes the setting input for capturing the deformable image in step S103, the processing unit 180 selects whether or not it is the power to register the captured deformable image in the registration information 11. Then, when recognizing the information indicating that it is registered in the registered image information 15 of the predetermined registration information 11 by the user's setting input, the processing unit 180 sets the captured image information to the predetermined image specified by the user's setting input. Record as registered image information 15 of registered information 11.
[0064] (携帯電話の着信処理)  [0064] (Incoming call processing of mobile phone)
次に、携帯電話 100に他の携帯電話や一般電話などから着信があった場合の着 信処理を図 10に基づいて説明する。図 11は、携帯電話の着信時における着信処理 のフローチャートである。  Next, an incoming call process when there is an incoming call to the mobile phone 100 from another mobile phone or a general phone will be described with reference to FIG. FIG. 11 is a flowchart of the incoming call process when a mobile phone receives an incoming call.
[0065] 携帯電話 100の処理部 180は、発着認識手段 181にて携帯電話 100に他の携帯 電話や一般電話などの通信機器から着信があることを認識、すなわち送受信部 130 力も通話を求める着信情報が入力されたことを認識されると (ステップ S 201)、相手 話者認識手段 182に、着信先の相手話者の登録情報 11を認識する (ステップ S202 )。すなわち、処理部 180の相手話者認識手段 182は、着信情報に記録された相手 話者の電話番号を認識し、この相手話者の電話番号と一致する電話番号情報 14を 有する登録情報 11を認識する。また、ステップ S202の後、相手話者認識手段 182 は、認識した登録情報 11の登録画像情報 15を認識する (ステップ S203)。  [0065] The processing unit 180 of the mobile phone 100 recognizes that the mobile phone 100 receives an incoming call from another communication device such as a mobile phone or a general phone at the arrival / departure recognition means 181. When it is recognized that the information has been input (step S 201), the partner speaker recognition means 182 recognizes the registration information 11 of the called partner speaker (step S 202). In other words, the partner speaker recognition means 182 of the processing unit 180 recognizes the telephone number of the partner speaker recorded in the incoming call information, and stores the registration information 11 having the telephone number information 14 that matches the telephone number of the partner speaker. recognize. Further, after step S202, the partner speaker recognition means 182 recognizes the registered image information 15 of the recognized registration information 11 (step S203).
[0066] このステップ S202で相手話者の登録情報 11がない場合、またはステップ S203に て相手話者の登録画像情報 15がない場合、処理部 180は、例えば表示制御手段 1 86に着信先の電話番号をディスプレイ 110の表示領域に表示させる制御をする。そ して、処理部 180は、利用者による操作部 140の操作により着信に応答する旨の操 作信号を認識すると (ステップ S 204)、送受信部 130を制御して着信先の相手話者 の通信機器に例えば電話回線やインターネットなどのネットワークを介して、互いに 通話可能な状態に通信接続する。そして、処理部 180は、利用者の操作部 140の操 作により、相手話者との通話を終了する旨の情報を認識したり、相手話者の通信機 器から通話を終了する旨の着信情報を認識したりすると、送受信部 130を制御して 通話可能状態を解除し、通信を終了する (ステップ S205)。 [0066] If there is no registration information 11 of the other speaker in this step S202, or in step S203 If there is no registered speaker information 15 of the other party, the processing unit 180 controls the display control means 186 to display the destination telephone number in the display area of the display 110, for example. Then, when the processing unit 180 recognizes an operation signal indicating that the user responds to the incoming call by operating the operation unit 140 (step S204), the processing unit 180 controls the transmission / reception unit 130 to determine the destination partner speaker. For example, a communication connection is established with a communication device through a network such as a telephone line or the Internet so that they can talk to each other. Then, the processing unit 180 recognizes information indicating that the call with the partner speaker is to be terminated by operating the operation unit 140 of the user, or receives an incoming call indicating that the call is to be terminated from the communication device of the partner speaker. When the information is recognized, the transmission / reception unit 130 is controlled to cancel the call ready state, and the communication is terminated (step S205).
[0067] 一方、処理部 180は、ステップ S202において、相手話者認識手段 182にて登録情 報 11が認識され、さらにステップ S203において、登録画像情報 15を認識されると、 表示制御手段 186にてディスプレイ 110を制御させてディスプレイ 110の表示領域に 登録画像情報 15の登録画像 50を表示させる (ステップ S206)。  On the other hand, when the registered information 11 is recognized by the partner speaker recognizing unit 182 in step S202 and the registered image information 15 is further recognized in step S203, the processing unit 180 displays the registered information 11 in the display control unit 186. The display 110 is controlled to display the registered image 50 of the registered image information 15 in the display area of the display 110 (step S206).
[0068] この後、処理部 180は、利用者による操作部 140の操作により着信に応答する旨の 操作信号を認識する (ステップ S207)と、送受信部 130を制御して着信先の相手話 者の通信機器に例えば電話回線やインターネットなどのネットワークを介して、互 ヽ に通話可能な状態に通信接続する。一方、ステップ S207において、利用者による操 作部 140の操作により着信に応答する旨の操作信号が認識できない場合、または着 信を拒否する旨の操作信号を認識した場合、着信情報の受信を終了させ、着信処 理を終了させる。  [0068] After that, when the processing unit 180 recognizes an operation signal indicating that the user responds to the incoming call by operating the operation unit 140 (step S207), the processing unit 180 controls the transmission / reception unit 130 to receive the destination partner speaker. For example, a communication connection is established with a communication device such as a telephone line or the Internet so as to be able to talk to each other. On the other hand, in step S207, if the operation signal indicating that the user responds to the incoming call cannot be recognized by operating the operation unit 140 or if the operation signal indicating that the incoming call is rejected is received, the reception of the incoming information is terminated. To terminate the incoming call processing.
[0069] そして、ステップ S207において、処理部 180は、着信に応答する旨の操作信号を 認識して、相手話者の通信機器と通信接続されると、音声認識手段 183にて着信先 の相手話者の音声情報が認識する (ステップ S208)。この後、処理部 180の音声認 識手段 183は、受信した音声情報を解析して、相手話者の音声の状態、すなわち音 声の高低、強弱、相手話者の話し方のテンポなどを音声基礎情報として認識する (ス テツプ S209)。また、音声認識手段 183は、認識した音声基礎情報を適宜読み出し 可能にメモリ 170に記憶する。  [0069] Then, in step S207, when the processing unit 180 recognizes the operation signal for responding to the incoming call and is connected to the communication apparatus of the other party's speaker, the voice recognition means 183 causes the other party The voice information of the speaker is recognized (step S208). After that, the voice recognition means 183 of the processing unit 180 analyzes the received voice information and determines the voice status of the other speaker, that is, the voice level, strength, weakness, tempo of the other speaker's speech, etc. Recognized as information (step S209). The voice recognition means 183 stores the recognized voice basic information in the memory 170 so that it can be read out as appropriate.
[0070] 次に、処理部 180は、利用者が相手話者と通話中に、相手話者の音声状態が変化 したことを検知する (ステップ S210)、すなわち、相手話者の通信機器から送信され る音声情報の音声の状態の変化を検出すると、音声変化判定手段 184にて、受信し た音声情報の音声の状態から相手話者の感情情報 44を認識する (ステップ S211) 。具体的には、音声変化判定手段 184は、相手話者の音声情報の音声の状態を解 祈し、メモリ 170に記憶された音声基礎情報と比較する。そして、処理部 180の音声 変化判定手段 184は、音声基礎情報の音声の状態および受信した音声情報の音声 状態が異なる場合、その変化した音声のパターン、例えば音声の高低の変化量、音 声の強弱の変化量、テンポの変化量などを認識する。また、音声変化判定手段 184 は、感情認識情報テーブル 40に基づいて、相手話者の音声のパターンと一致する 音声パターン情報 43を検索し、この音声パターン情報 43を有する感情認識情報 41 を認識する。また、音声変化判定手段 184は、相手話者の音声情報に音声パターン 情報 43に記録される語句情報、例えば「こら」や「ははは」といった感情を表す語句を 認識し、これらの語句情報に対応する感情認識情報 41を認識する。さらに、音声変 化判定手段 184は、認識した感情認識情報 41を、適宜読み出し可能にメモリ 170に feす。。 [0070] Next, the processing unit 180 changes the voice status of the other speaker while the user is talking to the other speaker. (Step S210), that is, when a change in the voice state of the voice information transmitted from the communication device of the other speaker is detected, the voice change determination means 184 detects the voice of the received voice information. The emotion information 44 of the other speaker is recognized from the state (step S211). Specifically, the voice change determination means 184 prays the voice state of the voice information of the other speaker and compares it with the voice basic information stored in the memory 170. Then, when the voice state of the voice basic information and the voice state of the received voice information are different, the voice change determination means 184 of the processing unit 180 has the changed voice pattern, for example, the amount of change in the voice, the amount of change in the voice, Recognize strength changes, tempo changes, etc. Also, the voice change determination means 184 searches the voice pattern information 43 matching the voice pattern of the other speaker based on the emotion recognition information table 40, and recognizes the emotion recognition information 41 having the voice pattern information 43. . The voice change determination means 184 also recognizes the phrase information recorded in the voice pattern information 43 in the voice information of the other speaker, for example, phrases representing emotions such as “Kora” and “Hahaha”, and these phrase information Emotion recognition information 41 corresponding to is recognized. Further, the voice change determining means 184 stores the recognized emotion recognition information 41 in the memory 170 so that it can be read out appropriately. .
[0071] このステップ S211の後、処理部 180の画像処理手段 185は、登録画像情報 15に 記録された登録画像 50を読み込み、登録画像 50の画像を編集する (ステップ S212 ) o具体的には、処理部 180の画像処理手段 185は、第 1画像テンプレートテーブル 20および第 2画像テンプレートテーブル 30を読み込む。そして、ステップ S211にて 認識してメモリ 170に記憶された感情認識情報 41の感情情報 44を認識し、この感情 情報 44に対応する対応感情情報 25, 35を有する第 1画像テンプレート情報 21、第 2画像テンプレート情報 31を認識する。  [0071] After step S211, the image processing means 185 of the processing unit 180 reads the registered image 50 recorded in the registered image information 15 and edits the image of the registered image 50 (step S212). The image processing means 185 of the processing unit 180 reads the first image template table 20 and the second image template table 30. Then, the emotion information 44 of the emotion recognition information 41 recognized in step S211 and stored in the memory 170 is recognized, and the first image template information 21 and the first image template information 21 having corresponding emotion information 25 and 35 corresponding to the emotion information 44 are stored. 2 Recognize image template information 31.
[0072] この後、処理部 180の画像処理手段 185は、第 1画像テンプレート情報 21の第 1画 像パターン情報 24、および第 2画像テンプレート情報 31の第 2画像パターン情報 34 に応じて、登録画像 50の右眉領域 53、左眉領域 54、および口領域 55を変形させて 感情変化画像を作成する (ステップ S213)。  Thereafter, the image processing means 185 of the processing unit 180 registers in accordance with the first image pattern information 24 of the first image template information 21 and the second image pattern information 34 of the second image template information 31. The emotional change image is created by deforming the right eyebrow region 53, the left eyebrow region 54, and the mouth region 55 of the image 50 (step S213).
[0073] 例えば、ステップ S211にお 、て音声変化判定手段 184にて認識された感情認識 情報 41の感情情報 44に「笑顔状態」である旨の情報が記録されて ヽる場合を例示 する。この場合、画像処理手段 185は、この感情情報 44に対応する対応感情情報 2 5, 35を有する第 1画像テンプレート情報 21の第 1画像パターン情報 24および第 2 画像テンプレート情報 31の第 2パターン情報 34に応じて、登録画像 50の右眉領域 5 3、左眉領域 54、および口領域 55の画像を、例えば図 9Aに示すような笑顔の画像 に変形させる。同様に、音声変化判定手段 184にて認識された感情認識情報 41の 感情情報 44に「怒り状態」である旨の情報が記録されて 、る場合、登録画像の右眉 領域 53、左眉領域 54、および口領域 55の画像を、例えば図 9Bに示すような、「怒り 状態」の顔画像に変形させる。さらに、音声変化判定手段 184にて認識された感情 認識情報 41の感情情報 44に「悲しみ状態」である旨の情報が記録されて ヽる場合、 登録画像の右眉領域 53、左眉領域 54、および口領域 55の画像を、例えば図 9Cに 示すような、「悲しみ状態」の顔画像に変形させる。 [0073] For example, in step S211, an example in which information indicating "smiling state" is recorded in emotion information 44 of emotion recognition information 41 recognized by voice change determination means 184 is shown. To do. In this case, the image processing means 185 includes the first image pattern information 24 of the first image template information 21 and the second pattern information of the second image template information 31 having corresponding emotion information 25, 35 corresponding to the emotion information 44. In response to 34, the images of the right eyebrow region 53, the left eyebrow region 54, and the mouth region 55 of the registered image 50 are transformed into, for example, a smile image as shown in FIG. 9A. Similarly, when information indicating “anger state” is recorded in the emotion information 44 of the emotion recognition information 41 recognized by the voice change determination means 184, the right eyebrow region 53 and the left eyebrow region of the registered image are recorded. The images of 54 and mouth area 55 are transformed into an “angry” face image as shown in FIG. 9B, for example. In addition, when information indicating that “sadness” is recorded in the emotion information 44 of the emotion recognition information 41 recognized by the voice change determination means 184, the right eyebrow region 53 and the left eyebrow region 54 of the registered image are recorded. And the image of the mouth region 55 are transformed into a face image in a “sadness state” as shown in FIG. 9C, for example.
[0074] そして、処理部 180の画像処理手段 185は、上記のように変形した感情変化画像 を適宜読み出し可能にメモリ 170に記憶する。  [0074] Then, the image processing means 185 of the processing unit 180 stores the emotion-changed image deformed as described above in the memory 170 so that it can be read out appropriately.
[0075] この後、処理部 180は、表示制御手段 186にてディスプレイ 110を制御させて、ス テツプ S212にて作成した感情変化画像をディスプレイ 110の表示領域に表示させる  [0075] Thereafter, the processing unit 180 controls the display 110 with the display control means 186 to display the emotion change image created at step S212 in the display area of the display 110.
[0076] そして、処理部 180は、利用者の操作部 140の操作により相手話者との通話を終 了する旨の操作信号を認識すると、通信接続を解除し、通話を終了させる。また、通 話を続行する場合は、ステップ S210に戻って、相手話者の音声状態を認識する処 理をする(ステップ S214)。 [0076] Then, when the processing unit 180 recognizes an operation signal for ending the call with the other party speaker by the operation of the operation unit 140 of the user, the processing unit 180 releases the communication connection and ends the call. If the call is to be continued, the process returns to step S210 to perform processing for recognizing the voice state of the other speaker (step S214).
[0077] (携帯電話の発信処理)  [0077] (Call processing of mobile phone)
次に、携帯電話 100から相手話者に発信して通話をする発信処理について図 12 に基づいて説明する。図 12は、携帯電話の発信時の発信処理におけるフローチヤ ートである。なお、図 12において図 11の携帯電話 100の着信処理と略同様の処理 には同符号を付与し、その説明を省略、または簡略化する。  Next, call processing for making a call by calling the mobile phone 100 to the other party speaker will be described with reference to FIG. Fig. 12 is a flowchart of the calling process when a mobile phone is called. In FIG. 12, the same reference numerals are given to the processing that is substantially the same as the incoming processing of the mobile phone 100 in FIG. 11, and the description thereof is omitted or simplified.
[0078] 図 12において、携帯電話 100は、先ず、処理部の発着認識手段 181にて、発着認 識手段 181にて利用者による操作部 140の操作により他者の通信機器に発信する 旨の発信要求情報を認識する (ステップ S301)。そして、処理部 180は、この発進要 求情報に発信先の相手話者の登録情報 11が記録されて 、な 、と判断した場合、利 用者の操作部 140の操作により設定入力される発信先の電話番号を認識し、発信先 の通信機器に発信情報を送信する (ステップ S302)。そして、発信先の相手話者が 発信に応答すると (ステップ S303)、送受信部 130を制御して着信先の相手話者の 通信機器に例えば電話回線やインターネットなどのネットワークを介して、互いに通 話可能な状態に通信接続する。そして、処理部 180は、利用者の操作部 140の操作 により、相手話者との通話を終了する旨の情報を認識したり、相手話者の通信機器 力 通話を終了する旨の着信情報を認識したりすると、送受信部 130を制御して通 話可能状態を解除し、通信を終了する (ステップ S205)。 In FIG. 12, the mobile phone 100 first sends a call to the communication device of the other party by operating the operation unit 140 by the user at the arrival / recognition recognition unit 181 at the arrival / departure recognition unit 181 of the processing unit. The call request information is recognized (step S301). Then, the processing unit 180 is required to If the registration information 11 of the other party's speaker is recorded in the request information, it is determined that the destination telephone number is set and entered by operating the operation unit 140 of the user. The transmission information is transmitted to the other communication device (step S302). Then, when the other party's caller answers the call (step S303), the transmission / reception unit 130 is controlled to communicate with the other party's caller's communication device via a network such as a telephone line or the Internet. Make a communication connection where possible. Then, the processing unit 180 recognizes information indicating that the call with the other speaker is to be terminated by operating the operation unit 140 of the user, or receives incoming call information indicating that the communication with the other speaker is ended. If it is recognized, the transmission / reception unit 130 is controlled to release the call ready state, and the communication is terminated (step S205).
[0079] 一方、ステップ S301において、処理部 180は、発信要求情報に発信先の登録情 報 11が登録情報テーブル 10に記録されて 、ることを認識すると、この登録情報 11に 登録画像情報 15が記録されて 、るか否かを判断する (ステップ S304)。このステップ S304において、登録画像情報 15が記録されていない場合は、処理部 180は、ステ ップ S302を実施し、登録情報 11の電話番号情報 14に記録された電話番号に発信 する。 On the other hand, in step S 301, when recognizing that the registration information 11 of the transmission destination is recorded in the registration information table 10 in the transmission request information, the processing unit 180 stores the registered image information 15 in the registration information 11. Whether or not is recorded is determined (step S304). In step S304, when the registered image information 15 is not recorded, the processing unit 180 executes step S302 and transmits the call to the telephone number recorded in the telephone number information 14 of the registration information 11.
[0080] また、ステップ S304において、登録情報 11に登録画像情報 15が記録されている 場合、処理部 180は、表示制御手段 186にディスプレイ 110を制御させて、登録画 像情報 15の登録画像 50をディスプレイ 110の表示領域に表示させる (ステップ S30 5)。  In step S 304, when the registered image information 15 is recorded in the registered information 11, the processing unit 180 causes the display control unit 186 to control the display 110 to register the registered image information 50 in the registered image information 15. Is displayed in the display area of the display 110 (step S30 5).
[0081] さらに、処理部 180は、送受信部 130を制御して、登録情報 11の電話番号情報 14 に記録されて 、る電話番号に発信情報を送信させる (ステップ S306)。  Furthermore, processing unit 180 controls transmission / reception unit 130 to transmit the transmission information to the telephone number recorded in telephone number information 14 of registration information 11 (step S306).
[0082] そして、ステップ S306にて送信した発信情報に対して、発信先の相手話者が応答 すると (ステップ S307)、処理部 180は、送受信部 130を制御して着信先の相手話者 の通信機器に例えば電話回線やインターネットなどのネットワークを介して、互いに 通話可能な状態に通信接続する。一方、ステップ S307において、相手話者が発信 情報に応答しない場合は、携帯電話 100の発信処理を終了させる。  [0082] Then, when the destination partner speaker responds to the transmission information transmitted in step S306 (step S307), processing unit 180 controls transmission / reception unit 130 to determine the destination partner speaker. For example, a communication connection is established with a communication device through a network such as a telephone line or the Internet so that they can talk to each other. On the other hand, if the other speaker does not respond to the call information in step S307, the call processing of the mobile phone 100 is terminated.
[0083] ステップ S307において、相手話者が発信情報に応答すると、処理部 180は、前述 した着信処理のステップ S 208ないしステップ S 214の処理を実施する。すなわち、処 理部 180は、ステップ S208を実施して、音声認識手段 183にて着信先の相手話者 の音声情報が認識する。この後、処理部 180の音声認識手段 183は、受信した音声 情報を解析して、相手話者の音声の状態、すなわち音声の高低、強弱、相手話者の 話し方のテンポなどを音声基礎情報として認識する。また、音声認識手段 183は、認 識した音声基礎情報を適宜読み出し可能にメモリ 170に記憶する。 [0083] In step S307, when the other speaker responds to the transmission information, the processing unit 180 performs the processing from step S208 to step S214 of the incoming call processing described above. That is, The processing unit 180 executes step S208, and the voice recognition means 183 recognizes the voice information of the called party's partner. After that, the voice recognition means 183 of the processing unit 180 analyzes the received voice information and uses the voice status of the other speaker, that is, the voice level, strength and weakness, and the other speaker's speaking tempo as the basic voice information. recognize. The voice recognition means 183 stores the recognized voice basic information in the memory 170 so that it can be read out as appropriate.
[0084] 次に、処理部 180は、ステップ S209の処理を実施し、利用者が相手話者と通話中 に、相手話者の音声状態が変化したことを検知し、音声変化判定手段 184にて受信 した音声情報の音声の状態から相手話者の感情情報 44を認識させる。また、音声変 化判定手段 184は、認識した感情認識情報 41を、適宜読み出し可能にメモリ 170に feす。。 Next, processing unit 180 performs the process of step S 209, detects that the voice state of the other speaker has changed while the user is talking to the other speaker, and sends it to voice change determination means 184. The other speaker's emotion information 44 is recognized from the voice state of the received voice information. Also, the voice change determination means 184 stores the recognized emotion recognition information 41 in the memory 170 so that it can be read out appropriately. .
[0085] そして、処理部 180は、ステップ S210の処置を実施し、画像処理手段 185にて登 録画像情報 15が記録された登録画像 50を読み込ませて登録画像 50の画像を編集 させ、感情変化画像を作成する。また、処理部の画像処理手段 185は、感情変化画 像を適宜読み出し可能にメモリ 170に記憶する。  [0085] Then, the processing unit 180 performs the processing of step S210, causes the image processing unit 185 to read the registered image 50 in which the registered image information 15 is recorded, edits the image of the registered image 50, and Create a change image. Further, the image processing means 185 of the processing unit stores the emotion change image in the memory 170 so that it can be read out as appropriate.
[0086] この後、処理部 180は、ステップ S213の処理を実施し、表示制御手段 186にてデ イスプレイ 110を制御させて、ステップ S212にて作成した感情変化画像をディスプレ ィ 110の表示領域に表示させる。  Thereafter, the processing unit 180 performs the process of step S213, controls the display 110 with the display control means 186, and displays the emotion change image created in step S212 in the display area of the display 110. Display.
[0087] そして、処理部 180は、ステップ S214の処理を実施して、利用者の操作部 140の 操作により相手話者との通話を終了する旨の操作信号を認識すると、通信接続を解 除し、通話を終了させる。また、通話を続行する場合は、ステップ S210に戻って、相 手話者の音声状態を認識する処理をする。  [0087] Then, when the processing unit 180 performs the process of step S214 and recognizes an operation signal indicating that the call with the other party is terminated by the operation of the operation unit 140 of the user, the processing unit 180 releases the communication connection. And end the call. If the call is to be continued, the process returns to step S210 to perform processing for recognizing the voice state of the other speaker.
[0088] 〔携帯電話の作用効果〕  [0088] [Function and effect of mobile phone]
上述したように、本実施の形態の携帯電話 100は、処理部 180の音声変化判定手 段 184にて相手話者の音声情報の音声パターンを認識させ、音声パターンが変化し たと判断されると、画像処理手段 185にて音声パターンの変化に応じて登録画像情 報 15の登録画像 50を変化させる。このため、登録画像情報 15の登録画像 50を音 声パターンの変化に応じて動画を再生させて 、るかのように変形させることができる。 また、相手話者の音声パターンは相手話者の感情により変化する。したがって、相手 話者の感情に応じて登録画像情報 15の登録画像 50を変形することができ、相手話 者の感情に応じた適切な感情変化画像をディスプレイ 110に表示させることができる As described above, the mobile phone 100 according to the present embodiment causes the voice change determination unit 184 of the processing unit 180 to recognize the voice pattern of the other speaker's voice information and determines that the voice pattern has changed. Then, the image processing means 185 changes the registered image 50 of the registered image information 15 in accordance with the change of the sound pattern. For this reason, the registered image 50 of the registered image information 15 can be transformed as if a moving image is reproduced according to the change of the voice pattern. Also, the voice pattern of the other speaker changes depending on the emotion of the other speaker. Therefore, opponent The registered image 50 of the registered image information 15 can be transformed according to the speaker's emotion, and an appropriate emotion change image corresponding to the emotion of the other speaker can be displayed on the display 110.
[0089] また、登録画像情報 15の登録画像 50を変形させて感情変化画像として表示させ ている。このため、各感情に対応する複数の画像を用意する必要がない。したがって 、記憶手段 160の記憶容量を圧迫することがなぐ空き容量を有効に活用することが できる。 Further, the registered image 50 of the registered image information 15 is deformed and displayed as an emotion change image. For this reason, it is not necessary to prepare a plurality of images corresponding to each emotion. Therefore, it is possible to effectively utilize the free capacity that does not press the storage capacity of the storage unit 160.
[0090] また、画像処理手段 185は、第 1画像テンプレート情報 21および第 2画像テンプレ ート情報 31に基づいて、登録画像情報 15の登録画像 50を変形させて感情変化画 像を生成する。このため、ディスプレイ 110の表示領域に表示される顔画像の背景や 、顔画像の配置位置などを変形させず、顔の表情だけを変形させることができる。し たがって、ディスプレイ 110に表示される画像を、登録画像 50から感情変化画像に 切り替える時、また感情変化画像から他の感情変化画像に切り替える時に、画像が ずれて見辛くなることがなぐスムーズに画像を切り替えることができる。  [0090] Further, the image processing means 185 deforms the registered image 50 of the registered image information 15 based on the first image template information 21 and the second image template information 31, and generates an emotion change image. For this reason, only the facial expression can be transformed without modifying the background of the face image displayed in the display area of the display 110, the position of the face image, or the like. Therefore, when the image displayed on the display 110 is switched from the registered image 50 to the emotional change image, or when switching from the emotional change image to another emotional change image, the image does not shift and becomes difficult to see. You can switch images.
[0091] さらに、画像処理手段 185は、第 1画像パターン情報 24および第 2画像パターン情 報 34に記録された眉および口の変化率に基づいて、右眉領域 53、左眉領域 54、お よび口領域 55の左眉、右眉、口を例えば所定量だけドット移動させて変形させてい る。このため、登録画像情報 15の登録画像 50の一部のみを画像変形処理すること で容易に顔の表情を変えることができる。したがって、登録画像 50の一部の画像処 理を実施するだけでょ 、ので、画像処理に伴う処理部の処理負荷を軽減させること ができる。  [0091] Furthermore, the image processing means 185 determines the right eyebrow region 53, the left eyebrow region 54, the eyebrow region 54 based on the eyebrow and mouth change rates recorded in the first image pattern information 24 and the second image pattern information 34. The left eyebrow, the right eyebrow, and the mouth of the mouth region 55 are deformed by moving the dots by a predetermined amount, for example. Therefore, the facial expression can be easily changed by subjecting only a part of the registered image 50 of the registered image information 15 to the image transformation process. Therefore, since only part of the registered image 50 is processed, the processing load on the processing unit associated with the image processing can be reduced.
[0092] さらに、画像処理手段 185は、右眉領域 53の右眉、左眉領域 54の左眉、口領域 5 5の口をそれぞれ変化させている。このため、感情の変化を最も表示できる顔の眉お よび口を変形させることができる。したがって、利用者は感情変化画像から相手話者 の感情を容易に確認することができる。  Further, the image processing means 185 changes the right eyebrow of the right eyebrow region 53, the left eyebrow of the left eyebrow region 54, and the mouth of the mouth region 55, respectively. For this reason, it is possible to deform the eyebrows and mouth of the face that can best display emotional changes. Therefore, the user can easily confirm the other speaker's emotion from the emotion change image.
[0093] また、画像処理手段 185は、第 1画像パターン情報 24および第 2画像パターン情 報 34に記録される眉の変形率および口の変形率に基づいて、右眉領域 53、左眉領 域 54の眉および口領域 55の口を変形している。このため、相手話者の音声の音声 パターンに応じて適切に画像の所定の一部を所定の方向に所定量がけ移動させて 変形させることができる。また、この時、画像処理手段 185は、登録画像 50の右眉領 域 53、左眉領域 54の眉と略一致する第 1画像基本情報 23、および口領域 55の口と 略一致する第 2画像基本情報 33を認識し、これらの第 1画像基本情報 23、第 2画像 基本情報 33に対応する第 1画像パターン情報 24および第 2画像パターン情報 34の 変形率に基づいて画像を変形している。このため、感情変化画像を相手話者の実際 の感情変化による表情に近付けることができる。したがって、利用者は相手話者の感 情の確認がより確実に実施できる。 [0093] Further, the image processing means 185 determines the right eyebrow region 53, the left eyebrow region based on the eyebrow deformation rate and the mouth deformation rate recorded in the first image pattern information 24 and the second image pattern information 34. The eyebrow in area 54 and the mouth in area 55 are deformed. For this reason, the voice of the other speaker's voice According to the pattern, a predetermined part of the image can be appropriately deformed by moving it in a predetermined direction by a predetermined amount. At this time, the image processing means 185 also includes the first image basic information 23 that substantially matches the right eyebrow region 53 of the registered image 50, the eyebrow of the left eyebrow region 54, and the second image that substantially matches the mouth of the mouth region 55. Recognize the basic image information 33 and transform the image based on the deformation rate of the first image pattern information 24 and the second image pattern information 34 corresponding to the first image basic information 23 and the second image basic information 33. Yes. For this reason, it is possible to bring the emotional change image closer to the facial expression of the other speaker's actual emotional change. Therefore, the user can confirm the feeling of the other speaker more reliably.
[0094] そして、音声変化判定手段 184は、音声認識手段 183にて認識する音声基礎情報 と、相手話者の音声情報とを比較して、相手話者の音声情報の音声パターンが音声 基礎情報の音声パターンと異なると、その音声パターン力も相手話者の感情を認識 する。そして、画像処理手段 185は、相手話者の感情に応じて画像を処理する。この ため、音声変化判定手段 184は、音声パターンの変化を容易に認識することができ る。したがって、画像処理手段 185は、登録画像 50を適切に相手話者の感情の変化 に応じた感情変化画像に変形処理することができる。  Then, the voice change determination means 184 compares the voice basic information recognized by the voice recognition means 183 with the voice information of the other speaker, and the voice pattern of the voice information of the other speaker is the voice basic information. If the voice pattern is different, the voice pattern power also recognizes the emotion of the other speaker. Then, the image processing means 185 processes the image according to the emotion of the other speaker. For this reason, the voice change determination means 184 can easily recognize the change of the voice pattern. Therefore, the image processing means 185 can appropriately transform the registered image 50 into an emotion change image corresponding to the change in the other speaker's emotion.
[0095] また、音声認識手段 183は、携帯電話 100と相手話者の通信機器とが通信接続さ れ通話可能状態となった後、相手話者の通信機器から送信される最初の音声情報 の音声パターンを音声基礎情報として認識する。このため、音声変化判定手段 184 は、話し始めの相手話者の感情を基本として、相手話者の感情が話し始めの時の状 態に比べてどのように変化したかを認識することができ、相手話者の感情の変化を話 し始めの時の状態に比べてどのように変化したかをディスプレイ 110に表示させること ができる。  [0095] Further, the voice recognition means 183 receives the first voice information transmitted from the communication device of the other speaker after the mobile phone 100 and the communication device of the other speaker are connected for communication and ready to communicate. Recognize a voice pattern as basic voice information. For this reason, the voice change determination means 184 can recognize how the emotion of the other speaker has changed compared to the state at the beginning of the conversation, based on the emotion of the other speaker at the beginning of the conversation. It is possible to display on the display 110 how the other speaker's emotion has changed compared to the initial state of speaking.
[0096] さらに、表示制御手段 186は、相手話者から着信時に登録画像情報 15に記録され た登録画像 50をディスプレイ 110に表示させる。そして、画像処理手段 185は、音声 変化判定手段 184にて音声情報の状態が変化したと判断されると、登録画像 50を 第 1画像テンプレートテーブル 20、第 2画像テンプレートテーブル 30に基づいて変 形させる。このため、上記実施の形態のような携帯電話 100など、相手話者と通話す る機器において、利用者は相手話者の感情を確認しながら通話することができる。し たがって、例えばテレビ電話などの機能がなくとも、簡単な構成で相手話者の感情変 化を確認しながら相手話者と通話ができ、良好な通話支援を提供できる。 [0096] Further, the display control means 186 causes the display 110 to display the registered image 50 recorded in the registered image information 15 when an incoming call is received from the other speaker. When the sound change determination means 184 determines that the state of the sound information has changed, the image processing means 185 converts the registered image 50 based on the first image template table 20 and the second image template table 30. Let For this reason, the user can make a call while confirming the feeling of the other party's speaker in a device that makes a call with the other party's speaker, such as the mobile phone 100 as in the above embodiment. Shi Therefore, even if there is no function such as a videophone, for example, it is possible to talk with the other party while confirming the emotional change of the other party with a simple configuration and to provide good call support.
[0097] また、表示制御手段 186は、相手話者へ通話する旨の情報を送信する発信時にお いても同様に、登録画像情報 15に記録された登録画像 50をディスプレイ 110に表 示させる。そして、画像処理手段 185は、音声変化判定手段 184にて音声情報の状 態が変化したと判断されると、登録画像 50を第 1画像テンプレートテーブル 20、第 2 画像テンプレートテーブル 30に基づいて変形させる。このため、着信時と同様に、利 用者は相手話者の感情を確認しながら通話することができ、良好な通話の支援がで きる。  [0097] Also, the display control means 186 causes the display 110 to display the registered image 50 recorded in the registered image information 15 in the same way even during outgoing calls for transmitting information to the other party. Then, when the sound change determination means 184 determines that the state of the sound information has changed, the image processing means 185 transforms the registered image 50 based on the first image template table 20 and the second image template table 30. Let For this reason, as with incoming calls, the user can make a call while checking the feelings of the other speaker, and can support good calls.
[0098] また、撮影手段 187は、登録画像情報 15に記録する登録画像 50を撮影する際に 、ディスプレイ 110上にフレーム 51およびフレーム線 52を表示させている。このため 、利用者は登録画像 50を撮影する際に、撮影対象者をこのフレーム 51に顔が入るよ うに、またフレーム線 52に合わせて、眉、口が配置されるように焦点を合わせることが できる。したがって、容易に登録画像情報 15に記録される登録画像 50を撮影するこ とがでさる。  Further, the photographing means 187 displays the frame 51 and the frame line 52 on the display 110 when photographing the registered image 50 recorded in the registered image information 15. For this reason, when the user captures the registered image 50, the user focuses on the subject so that his / her eye is placed in the frame 51 and the eyebrow and mouth are placed in line with the frame line 52. Is possible. Therefore, the registered image 50 recorded in the registered image information 15 can be easily taken.
[0099] さらに、フレーム 51およびフレーム線 52を用いて撮影された画像情報が登録画像 情報 15に記録されている。このため、画像処理手段 185は容易に右眉領域 53、左 眉領域 54、および口領域 55を認識して、眉および口の画像を変形処理させることが できる。したがって、登録画像 50を相手話者の音声情報の音声パターンに応じて適 切に画像編集してディスプレイ 110の表示させることができる。  Further, image information captured using the frame 51 and the frame line 52 is recorded in the registered image information 15. Therefore, the image processing means 185 can easily recognize the right eyebrow region 53, the left eyebrow region 54, and the mouth region 55, and can transform the image of the eyebrow and mouth. Therefore, the registered image 50 can be appropriately edited according to the voice pattern of the voice information of the other speaker and displayed on the display 110.
[0100] そして、携帯電話 100は、登録画像情報 15に記録された画像を画像処理手段 18 5にて変形処理させ、表示制御手段 186にてディスプレイ 110に表示させている。こ のため、動画のような大きなファイルを保存する記憶領域が不要にできる。さらに、受 信した音声情報に応じて登録画像情報 15の登録画像 50を変形処理するので、通信 手段で画像情報などの大きなファイルを送受信する必要がなぐ安定して変形処理さ れた感情変化画像を表示させることができる。  [0100] Then, the cellular phone 100 causes the image processing means 185 to transform the image recorded in the registered image information 15, and causes the display control means 186 to display the image on the display 110. This eliminates the need for a storage area for storing large files such as movies. Furthermore, since the registered image 50 of the registered image information 15 is transformed according to the received audio information, the emotional change image that has been stably transformed without the need to send and receive large files such as image information by the communication means. Can be displayed.
[0101] 〔実施形態の変形〕  [Modification of Embodiment]
なお、本発明は、上述した一実施の形態に限定されるものではなぐ本発明の目的を 達成できる範囲で以下に示される変形をも含むものである。 The present invention is not limited to the above-described embodiment. To the extent that can be achieved, the following modifications are also included.
[0102] すなわち、上記実施の形態では、画像処理装置および画像表示装置として携帯電 話 100を例示した力 これに限定されない。例えば、パーソナルコンピュータ、一般電 話機器、カーナビゲーシヨン装置など、他の電気機器に適用されてもよい。例えばパ 一ソナルコンピュータの場合では、マイク力も入力される音声情報や、インターネット などのネットワークから取得した音声情報、光ディスクなどの記憶媒体から取得した音 声情報などの音声状態を認識し、この音声状態の音声パターンの変化を検出して画 像処理手段にて画像を変形処理させる構成としてもよい。  That is, in the above-described embodiment, the power illustrated as the mobile phone 100 as the image processing device and the image display device is not limited to this. For example, the present invention may be applied to other electric devices such as a personal computer, a general telephone device, and a car navigation device. For example, in the case of a personal computer, it recognizes the voice state such as voice information in which microphone power is input, voice information obtained from a network such as the Internet, voice information obtained from a storage medium such as an optical disk, and the like. It is also possible to adopt a configuration in which a change in the sound pattern is detected and the image is deformed by the image processing means.
[0103] また、上記実施の形態では、操作部 140として操作ボタンや操作ツマミを備えた構 成としたが、これに限定されない。例えば携帯電話 100に接続されるキーボードやマ ウスであってもよぐディスプレイ 110に接触することで所定の設定入力操作が可能な タツチパネルであってもよい。さらに、音声による入力操作や、リモートコントローラな どの無線媒体を介して信号を出力する構成など、各種設定事項を設定入力可能な V、ずれの構成が適用できる。  [0103] In the above embodiment, the operation unit 140 includes an operation button and an operation knob. However, the present invention is not limited to this. For example, a touch panel that allows a predetermined setting input operation by touching the display 110 may be used even if it is a keyboard or a mouse connected to the mobile phone 100. In addition, V and shift configurations that can be set and input various setting items such as voice input operations and configurations that output signals via a wireless medium such as a remote controller can be applied.
[0104] さらに、発着認識手段 181は、着信情報を受信すると、音声出力部 120から音声、 警告音、報知音などの着信音を出力して利用者に着信を知らせる構成を示したが、 これに限らない。例えば、相手話者認識手段 182にて着信相手を認識した後に着信 音を出力させる構成としてもよい。この様な構成では、着信相手により、着信音を変え ることもでき、着信音のみで利用者に着信相手が誰であるかを伝えることができる。  [0104] Further, the arrival / departure recognition means 181 has shown a configuration in which when the incoming call information is received, the voice output unit 120 outputs a ringtone such as a voice, a warning sound, and a notification sound to notify the user of the incoming call. Not limited to. For example, the incoming call sound may be output after the called party recognition means 182 recognizes the called party. In such a configuration, the ring tone can be changed depending on the destination, and the user can be informed of who the other party is by using only the ring tone.
[0105] また、着信音により着信を報知する構成の他に、振動などにより利用者に着信を報 知する構成としてもよい。この構成では、携帯電話 100は、振動手段を備えた構成と し、発着認識手段 181にて着信を認識すると、振動手段を振動させて着信情報を報 知する。また、上記のように相手話者認識手段 182にて着信相手を認識した後に振 動手段を振動させる構成としてもよい。この場合、振動のパターンにより着信相手が 誰であるかを利用者に伝えることができる。  [0105] Further, in addition to the configuration in which an incoming call is notified by a ring tone, a configuration in which an incoming call is notified to a user by vibration or the like may be employed. In this configuration, the mobile phone 100 is configured to include a vibration unit, and when the incoming / outgoing recognition unit 181 recognizes an incoming call, the mobile unit 100 vibrates the vibration unit and reports incoming information. Further, as described above, the vibration means may be vibrated after the partner speaker recognition means 182 recognizes the called party. In this case, the vibration pattern can tell the user who the call is.
[0106] そして、音声認識手段 183は、前述したように、登録情報 11の登録詳細情報 16に 記載された内容に基づ ヽて、相手話者の性別や声の音声基礎情報などを認識する 構成としてもよい。このような構成では、音声認識手段 183は、登録詳細情報 16に記 録された音声基礎情報に基づいて、着信時、または発信時の相手話者の声の状態 、および音声基礎情報に記録された情報を比較することができる。したがって、例え ば着信時や発信時に相手話者が怒って ヽる場合などでも、相手話者の「怒り状態」を ディスプレイ 110に表示させることができる。よって、より適切に相手話者の感情に応 じた画像を表示させることができる。 [0106] Then, as described above, the voice recognition means 183 recognizes the gender of the other speaker, basic voice voice information, etc., based on the contents described in the registration details information 16 of the registration information 11. It is good also as a structure. In such a configuration, the voice recognition means 183 records the registration details information 16. Based on the recorded voice basic information, it is possible to compare the voice state of the other speaker at the time of incoming or outgoing call and the information recorded in the voice basic information. Therefore, for example, even when the other speaker is angry at the time of an incoming call or outgoing call, the “anger state” of the other speaker can be displayed on the display 110. Therefore, an image corresponding to the emotion of the other speaker can be displayed more appropriately.
[0107] さらに、相手話者の通信機器カゝら携帯電話 100に音声基礎情報が送信される構成 としてもよ ヽ。このような構成でも携帯電話 100は音声基礎情報を認識する処理を不 要にできるので、構成が簡単になり、処理負荷も軽減できる。  [0107] Further, the basic voice information may be transmitted to the mobile phone 100 such as the communication device of the other speaker. Even in such a configuration, the cellular phone 100 can eliminate the process of recognizing the basic voice information, which simplifies the configuration and reduces the processing load.
[0108] また、上記実施の形態では、画像処理手段 185にて登録画像を変形させて感情変 化画像を作成して表示させる構成を示したが、これに限定されない。例えば、携帯電 話 100は、登録画像情報 15に感情情報 44に関連付けられた複数の登録画像を記 録する。そして、処理部 180の音声変化判定手段 184は、相手話者の音声情報の音 声パターンの変化を認識すると、その音声パターンに対応する音声パターン情報 43 を有する感情認識情報 41を認識する。そして表示制御手段 186は、登録画像情報 1 5に記録された複数の登録画像から感情認識情報 41の感情情報 44に対応する登 録画像を読み込み、ディスプレイ 110の表示領域に表示させる制御をする。このよう に、相手話者の音声状態の変化に応じて、ディスプレイ 110に表示させる画像を相 手話者の感情に応じた画像に切り替える構成としてもよい。このような構成において も、前記した作用効果のように、相手話者の感情の変化に応じて適切な感情変化画 像をディスプレイ 110に表示させることができる。さらに、予め記憶手段 160に記録さ れた画像を切り替える構成としているため、画像を処理する必要がない。したがって、 処理部 180の処理負荷を軽減でき、処理を高速化することができる。  [0108] In the above-described embodiment, the configuration has been described in which the registered image is deformed by the image processing unit 185 to create and display an emotion-change image, but the present invention is not limited to this. For example, the mobile phone 100 records a plurality of registered images associated with the emotion information 44 in the registered image information 15. Then, when the voice change determination means 184 of the processing unit 180 recognizes the change in the voice pattern of the voice information of the other speaker, it recognizes the emotion recognition information 41 having the voice pattern information 43 corresponding to the voice pattern. Then, the display control means 186 performs control to read a registered image corresponding to the emotion information 44 of the emotion recognition information 41 from a plurality of registered images recorded in the registered image information 15 and display it in the display area of the display 110. In this manner, the image displayed on the display 110 may be switched to an image corresponding to the emotion of the other speaker according to a change in the voice state of the other speaker. Even in such a configuration, it is possible to display an appropriate emotion change image on the display 110 according to the change in the emotion of the other speaker, as in the above-described effects. Furthermore, since the image recorded in advance in the storage unit 160 is switched, it is not necessary to process the image. Therefore, the processing load on the processing unit 180 can be reduced, and the processing speed can be increased.
[0109] さらに、画像処理手段 185は、音声変化判定手段 184にて認識した感情認識情報 41に応じて、登録画像 50の一部、例えば眉および口を他の眉や口の要素画像に置 換する構成としてもよい。この場合、例えば第 1画像テンプレート情報 21の第 1画像 ノターン情報 24に、図 13に示すような要素画像としての眉画像を記録する。また第 2画像テンプレート情報 31の第 2画像パターン情報 34に図 14に示すような要素画像 としての口画像を記録する。図 13Aは、第 1画像基本情報 23の眉画像の人物が笑 顔の状態の時に撮影される眉画像の一例である。また、図 13Bは、第 1画像基本情 報 23の眉画像の人物が「怒り状態」の時に撮影される眉画像の一例である。また、図 13Cは、第 1画像基本情報 23の眉画像の人物が悲しみ状態の時に撮影される眉画 像の一例である。図 14Aは、第 2画像基本情報 33の口画像の人物が笑顔の状態の 時に撮影される口画像の一例である。また、図 14Bは、第 2画像基本情報 33の口画 像の人物が「怒り状態」の時に撮影される口画像の一例である。また、図 14Cは、第 2 画像基本情報 33の口画像の人物が悲しみ状態の時に撮影される口画像の一例で ある。この構成では、音声変化判定手段 184にて、相手話者の音声情報の音声バタ ーンが音声基礎情報の音声パターンと異なると判断すると、その音声パターンに対 応する音声パターン情報 43の感情認識情報 41を認識する。そして、画像処理手段 185は、登録画像 50の左右眉領域 53, 54の眉および口領域 55の口画像と形状が 略一致する眉画像の第 1画像基本情報 23および第 2画像基本情報 33を有する第 1 画像テンプレート情報 21および第 2画像テンプレート情報 31を検索する。さらに、画 像処理手段 185は、検索した第 1画像テンプレート情報 21および第 2テンプレート情 報 31から、音声変化判定手段 184にて認識した感情認識情報 41の感情情報 44〖こ 対応する対応感情情報 25, 35を有する第 1画像テンプレート情報 21および第 2テン プレート情報 31を認識する。そして、画像処理手段 185は、登録画像 50の左右眉領 域 53, 54および口領域 55の画像を、これらの認識した第 1画像テンプレート情報 21 および第 2テンプレート情報 31の第 1画像パターン情報 24および第 2画像パターン 情報 34の画像に置換する。このような構成でも、相手話者の音声情報に応じて画像 情報を切り替えて表示することができ、相手話者の感情を容易に知ることができる。 [0109] Furthermore, the image processing means 185 places a part of the registered image 50, for example, the eyebrows and the mouth, on the other eyebrows and mouth element images in accordance with the emotion recognition information 41 recognized by the sound change judgment means 184. It is good also as a structure to replace. In this case, for example, an eyebrow image as an element image as shown in FIG. 13 is recorded in the first image pattern information 24 of the first image template information 21. Further, a mouth image as an element image as shown in FIG. 14 is recorded in the second image pattern information 34 of the second image template information 31. Figure 13A shows the person in the eyebrow image of the first basic image information 23 It is an example of the eyebrow image image | photographed in the state of a face. FIG. 13B is an example of an eyebrow image captured when the person in the eyebrow image of the first image basic information 23 is “angry”. FIG. 13C is an example of an eyebrow image captured when the person of the eyebrow image of the first image basic information 23 is in a sad state. FIG. 14A is an example of a mouth image taken when a person in the mouth image of the second image basic information 33 is smiling. FIG. 14B is an example of a mouth image taken when the person of the mouth image of the second image basic information 33 is “angry”. FIG. 14C is an example of a mouth image taken when the person in the mouth image of the second image basic information 33 is in a sad state. In this configuration, when the voice change determination means 184 determines that the voice pattern of the other speaker's voice information is different from the voice pattern of the voice basic information, the emotion recognition of the voice pattern information 43 corresponding to the voice pattern is performed. Recognize information 41. Then, the image processing means 185 obtains the first image basic information 23 and the second image basic information 33 of the eyebrow image whose shapes substantially coincide with the eyebrows of the right and left eyebrow regions 53 and 54 and the mouth region 55 of the registered image 50. The first image template information 21 and the second image template information 31 are searched. Furthermore, the image processing means 185 uses the searched first image template information 21 and second template information 31 to detect the emotion information 44 of the emotion recognition information 41 recognized by the voice change determination means 184. First image template information 21 and second template information 31 having 25 and 35 are recognized. Then, the image processing means 185 converts the images of the left and right eyebrow regions 53 and 54 and the mouth region 55 of the registered image 50 into the recognized first image pattern information 21 of the first image template information 21 and second template information 31. Also, the second image pattern information 34 is replaced with the image. Even in such a configuration, the image information can be switched and displayed according to the voice information of the other speaker, and the emotion of the other speaker can be easily known.
[0110] また、登録画像の眉画像、および口画像を変形処理する例を示したが、これに限定 されない。例えば、画像処理手段 185は、顔画像の他の部位、例えば鼻、額、左右 頰、左右顎、耳など、さらには顔の輪郭なども同様にして変形処理する構成としても よい。さらに、これらの各部位の色を変形処理してもよい。例えば、対応感情情報に「 怒った状態」を示す情報があれば、画像処理手段 185は、例えば耳ゃ頰の赤色の強 さを強くして画像変形処理させるなど、コントラストや色彩を変更するなどしてもよい。  [0110] Further, although an example in which the eyebrow image and the mouth image of the registered image are transformed is shown, the present invention is not limited to this. For example, the image processing means 185 may be configured to similarly deform other parts of the face image, such as the nose, forehead, left and right eyelids, left and right jaws, ears, and the face contour. Further, the color of each of these parts may be transformed. For example, if there is information indicating “angry state” in the corresponding emotion information, the image processing means 185 changes the contrast and color by, for example, increasing the red intensity of the ears and performing image transformation processing. May be.
[0111] さらに、登録画像情報 15に記録される登録画像 50は、顔がフレーム 51内に入り、 眉画像、および口画像がそれぞれ左右眉領域 53, 54および口領域 55内に入った 状態の顔の正面画像が撮影されるとしたが、これに限定されない。例えば、横顔が撮 影される顔画像が用いられる構成としてもよぐ斜めから撮影された顔画像が用いら れる構成としてもよい。例えば横顔を登録画像として用いる場合には、横画像専用の フレームを用いてこのフレーム内に横顔が入るように焦点を合わせて撮影してもよ ヽ 。さらに、顔の輪郭を自動で認識する構成としてもよい。この場合、例えば登録画像 情報 15の登録画像の顔の色と風景画像の色との境界力も顔の輪郭を認識し、さらに 顔の内部の目、口、眉、などを色違いなど力も認識する。このような構成では、正面か ら撮影された画像に限らず、様々な方向から撮影された画像をも画像処理して変形 させることがでさる。 [0111] Furthermore, the registered image 50 recorded in the registered image information 15 has a face in the frame 51, Although the front image of the face in a state where the eyebrow image and the mouth image are in the left and right eyebrow areas 53 and 54 and the mouth area 55, respectively, is taken, the present invention is not limited to this. For example, a configuration in which a face image taken from a side face is used, or a face image taken from an oblique direction may be used. For example, when a profile is used as a registered image, it is possible to use a frame dedicated to the landscape image and focus on the profile so that the profile appears within this frame. Furthermore, it is good also as a structure which recognizes the outline of a face automatically. In this case, for example, the boundary force between the face color of the registered image in the registered image information 15 and the color of the landscape image also recognizes the outline of the face, and also recognizes the power of the eyes, mouth, eyebrows, etc. inside the face such as different colors. . In such a configuration, not only images taken from the front but also images taken from various directions can be processed and deformed.
[0112] さらには、登録画像情報 15に記録される登録画像 50には、例えば動物の画像や、 人形の画像などが記録されていてもよい。このような個性でも上記実施の形態と同様 の処理にて動物や人形の表情を変化させた感情変化画像を作成し、ディスプレイ 11 0の表示領域に表示させることができる。また、相手話者の音声情報の音声状態に応 じて、複数の画像、例えば複数の動物の画像、複数の人形の画像などを切り替えて 表示させる構成としてもょ ヽ。  Furthermore, in the registered image 50 recorded in the registered image information 15, for example, an animal image, a doll image, or the like may be recorded. Even with such individuality, it is possible to create an emotion change image in which the expression of an animal or doll is changed by the same processing as in the above-described embodiment, and display it in the display area of the display 110. It is also possible to switch between a plurality of images, for example, a plurality of animal images, a plurality of doll images, etc., according to the voice state of the other speaker's voice information.
[0113] そして、上記実施の形態の携帯電話 100では、通信機器から送信される相手話者 の音声情報を認識し、この音声情報の音声状態のパターンを認識する構成を示した 力 れに限らない。例えば、相手話者の通信機器から音声情報の音声状態を送信す る構成とし、携帯電話 100は、相手話者の通信機器から送信される音声状態を受信 する構成としてもよい。この構成では、携帯電話 100にて音声情報の音声状態を検 出する構成が不要となるので、処理負荷を軽減できるとともに、構成を簡単にできる。 さらに、本発明の画像処理装置にて音声情報の音声パターンに応じて変形処理され た画像情報が音声情報とともに携帯電話などの受信装置に送信される構成としても よい。  [0113] The mobile phone 100 according to the above embodiment is not limited to the power shown in the configuration for recognizing the voice information of the other party speaker transmitted from the communication device and recognizing the voice state pattern of the voice information. Absent. For example, the voice state of voice information may be transmitted from the communication device of the other speaker, and the mobile phone 100 may be configured to receive the voice state transmitted from the communication device of the other speaker. This configuration eliminates the need for detecting the voice state of the voice information with the mobile phone 100, thereby reducing the processing load and simplifying the configuration. Furthermore, the image information modified according to the sound pattern of the sound information by the image processing apparatus of the present invention may be transmitted together with the sound information to a receiving device such as a mobile phone.
[0114] さらに、利用者および相手話者の双方が上記実施の形態の携帯電話 100を利用し た通信システムとする構成としてもよい。このような構成では、利用者および相手話者 の双方が上記実施の形態の携帯電話 100をそれぞれ送受信端末として利用するこ とができる。そして、このような携帯電話 100を利用する通信システムでは、各利用者 は、ディスプレイ 110に表示される画像をみることで、携帯電話 100の発信元および 発信先の相手話者の感情を容易に認識することができる。 [0114] Furthermore, a configuration may be adopted in which both the user and the other party's speaker are configured as a communication system using the mobile phone 100 of the above embodiment. In such a configuration, both the user and the other speaker use the mobile phone 100 of the above embodiment as a transmission / reception terminal. You can. In such a communication system using the mobile phone 100, each user can easily feel the emotion of the caller of the mobile phone 100 and the other party of the callee by viewing the image displayed on the display 110. Can be recognized.
[0115] また、相手話者の一般電話などの通信機器に発信するときのみにディスプレイ 110 に画像を表示させ、相手話者の音声の状態に応じて画像を変化させる構成としても よぐまた相手話者の通信機器力も着信したときのみにディスプレイ 110に画像を表 示させ、相手話者の音声の状態に応じて画像を変化させる構成としてもよい。  [0115] The display 110 may be displayed only when a call is made to a communication device such as a general telephone of the other speaker, and the image may be changed according to the voice state of the other speaker. A configuration may be adopted in which an image is displayed on the display 110 only when the communication device power of the speaker is received and the image is changed according to the voice state of the other speaker.
[0116] 上述した各機能をプログラムとして構築したが、例えば回路基板などのハードウェア あるいは 1つの IC (Integrated Circuit)などの素子にて構成するなどしてもよぐいず れの形態としても利用できる。なお、プログラムや別途記録媒体から読み取らせる構 成とすることにより、取扱が容易で、利用の拡大が容易に図れる。  [0116] Each function described above has been constructed as a program, but it can be used in any form, for example, configured by hardware such as a circuit board or an element such as a single IC (Integrated Circuit). . By adopting a configuration that allows reading from a program or a separate recording medium, handling is easy and usage can be easily expanded.
[0117] その他、本発明の実施の際の具体的な構造および手順は、本発明の目的を達成 できる範囲で他の構造などに適宜変更できる。  [0117] In addition, the specific structure and procedure for carrying out the present invention can be appropriately changed to other structures and the like as long as the object of the present invention can be achieved.
[0118] 〔実施の形態の効果〕  [Effects of Embodiment]
上述したように、上記一実施の形態の携帯電話 100は、処理部 180の音声変化判定 手段 184にて相手話者の音声情報の音声パターンを認識させ、音声パターンが変 化したと判断されると、画像処理手段 185にて音声パターンの変化に応じて登録画 像情報 15の登録画像 50を変化させる。このため、登録画像情報 15の登録画像 50 を音声パターンの変化に応じて変形させることができ、相手話者の感情に応じて登 録画像情報 15の登録画像 50を変形することで相手話者の感情に応じた適切な感情 変化画像をディスプレイ 110に表示させることができる。  As described above, the mobile phone 100 according to the above-described embodiment causes the voice change determination unit 184 of the processing unit 180 to recognize the voice pattern of the voice information of the other speaker and determines that the voice pattern has changed. Then, the image processing means 185 changes the registered image 50 of the registered image information 15 according to the change of the sound pattern. Therefore, the registered image 50 of the registered image information 15 can be deformed according to the change of the voice pattern, and the registered image 50 of the registered image information 15 can be deformed according to the emotion of the other speaker. It is possible to display on the display 110 an appropriate emotion change image corresponding to the emotion of the person.
産業上の利用可能性  Industrial applicability
[0119] 本発明は、画像を表示する画像処理装置、画像表示装置、受信装置、送信装置、 通信システム、画像処理方法、画像処理プログラム、画像処理プログラムを記録した 記録媒体に利用できる。 The present invention can be used for an image processing device that displays an image, an image display device, a receiving device, a transmitting device, a communication system, an image processing method, an image processing program, and a recording medium that records the image processing program.

Claims

請求の範囲 The scope of the claims
[1] 表示手段に表示される画像を、受信して得られる音声情報に応じて処理する画像 処理装置であって、  [1] An image processing apparatus that processes an image displayed on a display unit according to audio information obtained by reception,
前記音声情報における音声状態の変化を認識する音声状態認識手段と、 前記画像を前記表示手段に表示させるとともに、前記表示された画像を前記音声 状態の変化に応じて変更させる表示制御手段と、  A voice state recognition unit for recognizing a change in a voice state in the voice information; a display control unit for causing the display unit to display the image and changing the displayed image according to the change in the voice state;
を具備したことを特徴とした画像処理装置。  An image processing apparatus comprising:
[2] 請求項 1に記載の画像処理装置であって、  [2] The image processing device according to claim 1,
前記音声状態の変化に応じて、前記画像を変形する画像変形処理手段を具備し、 前記表示制御手段は、前記音声状態の変化に応じて変形された前記画像を表示 させる  An image deformation processing unit that deforms the image according to the change in the sound state is provided, and the display control unit displays the image deformed according to the change in the sound state.
ことを特徴とする画像処理装置。  An image processing apparatus.
[3] 請求項 2に記載の画像処理装置であって、 [3] The image processing device according to claim 2,
前記画像変形処理手段は、前記音声状態の変化に応じて、前記画像の少なくとも 一部を画像全体に対して相対変形させる  The image deformation processing means relatively deforms at least a part of the image with respect to the entire image according to the change in the sound state.
ことを特徴とする画像処理装置。  An image processing apparatus.
[4] 請求項 3に記載の画像処理装置であって、 [4] The image processing device according to claim 3,
前記音声状態の変化に応じて前記画像の少なくとも一部を変形させる変形量に関 する変形量情報を有し、  Deformation amount information relating to a deformation amount that deforms at least a part of the image according to a change in the sound state;
前記画像変形処理手段は、前記画像の少なくとも一部を前記変形量情報の変形 量だけ相対変形させる  The image deformation processing means relatively deforms at least a part of the image by a deformation amount of the deformation amount information.
ことを特徴とする画像処理装置。  An image processing apparatus.
[5] 請求項 1に記載の画像処理装置であって、 [5] The image processing device according to claim 1,
前記音声状態の変化に応じて前記画像の少なくとも一部を変形させる変形方向に 関する変形方向情報を有し、  Deformation direction information about a deformation direction that deforms at least a part of the image in response to a change in the audio state,
前記画像変形処理手段は、前記画像の少なくとも一部を前記変形方向情報の変 形方向に相対変形させる  The image deformation processing means relatively deforms at least a part of the image in the deformation direction of the deformation direction information.
ことを特徴とする画像処理装置。 An image processing apparatus.
[6] 請求項 1に記載の画像処理装置であって、 [6] The image processing device according to claim 1,
前記画像は、複数の要素画像によって構成され、  The image is composed of a plurality of element images,
前記表示制御手段は、前記音声情報における音声状態が変化すると前記表示手 段に表示された前記画像の少なくとも一部の要素画像を前記音声状態の変化に応 じた他の要素画像に変えて表示させる  When the sound state in the sound information changes, the display control means changes and displays at least a part of the element image of the image displayed on the display means to another element image corresponding to the change of the sound state. Make
ことを特徴とする画像処理装置。  An image processing apparatus.
[7] 請求項 1な!、し請求項 6の 、ずれかに記載の画像処理装置であって、 [7] The image processing apparatus according to any one of claims 1 to 6 and claim 6,
前記音声情報の標準的な音声状態を認識する標準音声状態認識手段を具備し、 前記音声状態認識手段は、前記標準的な音声状態と異なる音声状態を認識する ことを特徴とする画像処理装置。  An image processing apparatus comprising: a standard voice state recognition unit that recognizes a standard voice state of the voice information, wherein the voice state recognition unit recognizes a voice state different from the standard voice state.
[8] 請求項 7に記載の画像処理装置であって、 [8] The image processing device according to claim 7,
前記音声情報を認識する音声情報認識手段を備え、  Voice information recognition means for recognizing the voice information;
前記標準音声状態認識手段は、前記音声情報認識手段にて最初に認識した音声 情報における音声状態を標準的な音声状態として認識する  The standard voice state recognition unit recognizes a voice state in voice information first recognized by the voice information recognition unit as a standard voice state.
ことを特徴とする画像処理装置。  An image processing apparatus.
[9] 画像を記憶する記憶手段と、 [9] storage means for storing images;
前記音声情報を受信する際に前記画像を表示する表示手段と、  Display means for displaying the image when receiving the audio information;
請求項 1な!ヽし請求項 8の ヽずれかに記載の画像処理装置と、  The image processing apparatus according to any one of claims 1 to 8, and
を具備したことを特徴とした画像表示装置。  An image display device comprising:
[10] 請求項 9に記載の画像表示装置と、 [10] The image display device according to claim 9,
前記音声情報を受信可能な受信手段と、を備え、  Receiving means capable of receiving the voice information,
前記表示制御手段は、前記受信手段により音声情報が受信される際に、その音声 情報の送信元に対応する画像を表示する  The display control unit displays an image corresponding to the transmission source of the audio information when the reception unit receives the audio information.
ことを特徴とする受信装置。  A receiving apparatus.
[11] 請求項 9に記載の画像表示装置と、 [11] The image display device according to claim 9,
前記音声信号を送受信可能な送信手段と、を備え、  Transmission means capable of transmitting and receiving the audio signal,
前記送信手段は、前記音声情報を送信する送信先に対して発呼する発呼手段を 有し、 前記表示手段は、前記発呼手段による発呼または前記発呼に対する前記送信先 の応答に応じて前記送信先に対応する画像を表示し、 The transmission means includes a calling means for calling a transmission destination for transmitting the voice information; The display means displays an image corresponding to the transmission destination according to a call made by the calling means or a response of the transmission destination to the call,
前記画像処理装置は、前記発呼に応答して受信される送信先からの音声情報に おける音声状態の変化に応じて、表示された画像を変更する  The image processing device changes a displayed image in response to a change in sound state in sound information from a transmission destination received in response to the call.
ことを特徴とする送信装置。  A transmission apparatus characterized by the above.
[12] 相互に音声情報を送受信可能な送受信端末を備えた通信システムであって、 前記送受信端末の各々は、  [12] A communication system including a transmission / reception terminal capable of transmitting / receiving voice information to / from each other,
画像を記憶する記憶手段と、  Storage means for storing images;
音声情報を送信する送信手段と、  A transmission means for transmitting voice information;
音声情報を送信するために送信元に対して発呼する発呼手段と、  A calling means for calling a transmission source to transmit voice information;
音声情報を受信する受信手段と、  Receiving means for receiving audio information;
前記音声情報を受信する際に、または前記発呼手段により発呼する際に、または 前記発呼に対し前記送信先が応答する際に前記画像を表示する表示手段と、 請求項 1な!ヽし請求項 8の ヽずれかに記載の画像処理装置と、  Display means for displaying the image when receiving the voice information, when making a call by the calling means, or when the destination responds to the call; and And the image processing device according to claim 8;
を具備したことを特徴とする通信システム。  A communication system comprising:
[13] 表示手段に表示される画像を、受信して得られる音声情報に応じて処理する画像 処理方法であって、 [13] An image processing method for processing an image displayed on a display means according to audio information obtained by reception,
前記音声情報における音声状態の変化を認識し、  Recognizing a change in voice state in the voice information;
前記画像を前記表示手段に表示させるとともに、前記表示された画像を前記音声 状態の変化に応じて変更させる  The image is displayed on the display means, and the displayed image is changed according to the change in the sound state.
ことを特徴とする画像処理方法。  An image processing method.
[14] 演算手段を請求項 1な!ヽし請求項 8の ヽずれかに記載の画像処理装置として機能 させる [14] Let the computing means function as the image processing device according to any one of claims 1 to 8
ことを特徴とする画像処理プログラム。  An image processing program characterized by that.
[15] 請求項 13に記載の画像処理方法を演算手段に実施させる [15] An image processing method according to claim 13 is executed by an arithmetic means.
ことを特徴とする画像処理プログラム。  An image processing program characterized by that.
[16] 請求項 14または請求項 15に記載の画像処理プログラムが演算手段にて読取可能 に記録された ことを特徴とする画像処理プログラムを記録した記録媒体。 [16] The image processing program according to claim 14 or claim 15 is recorded so as to be readable by an arithmetic means. The recording medium which recorded the image processing program characterized by the above-mentioned.
PCT/JP2006/306297 2005-03-31 2006-03-28 Image processing device, image display device, reception device, transmission device, communication system, image processing method, image processing program, and recording medium containing the image processing program WO2006106671A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2005-103092 2005-03-31
JP2005103092 2005-03-31

Publications (1)

Publication Number Publication Date
WO2006106671A1 true WO2006106671A1 (en) 2006-10-12

Family

ID=37073237

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2006/306297 WO2006106671A1 (en) 2005-03-31 2006-03-28 Image processing device, image display device, reception device, transmission device, communication system, image processing method, image processing program, and recording medium containing the image processing program

Country Status (1)

Country Link
WO (1) WO2006106671A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012138035A (en) * 2010-12-28 2012-07-19 Casio Comput Co Ltd Image display device and program
JP2016167815A (en) * 2011-09-09 2016-09-15 クゥアルコム・インコーポレイテッドQualcomm Incorporated Transmission of feeling as tactile sense feedback

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH05265482A (en) * 1992-03-17 1993-10-15 Matsushita Electric Ind Co Ltd Information processor
JPH09138767A (en) * 1995-11-14 1997-05-27 Fujitsu Ten Ltd Communication equipment for feeling expression
JPH10293860A (en) * 1997-02-24 1998-11-04 Nippon Telegr & Teleph Corp <Ntt> Person image display method and device using voice drive
JP2002215180A (en) * 2001-01-17 2002-07-31 Digital Media Lab Inc Communication device
JP2003037826A (en) * 2001-07-23 2003-02-07 Alpine Electronics Inc Substitute image display and tv phone apparatus

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH05265482A (en) * 1992-03-17 1993-10-15 Matsushita Electric Ind Co Ltd Information processor
JPH09138767A (en) * 1995-11-14 1997-05-27 Fujitsu Ten Ltd Communication equipment for feeling expression
JPH10293860A (en) * 1997-02-24 1998-11-04 Nippon Telegr & Teleph Corp <Ntt> Person image display method and device using voice drive
JP2002215180A (en) * 2001-01-17 2002-07-31 Digital Media Lab Inc Communication device
JP2003037826A (en) * 2001-07-23 2003-02-07 Alpine Electronics Inc Substitute image display and tv phone apparatus

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012138035A (en) * 2010-12-28 2012-07-19 Casio Comput Co Ltd Image display device and program
JP2016167815A (en) * 2011-09-09 2016-09-15 クゥアルコム・インコーポレイテッドQualcomm Incorporated Transmission of feeling as tactile sense feedback

Similar Documents

Publication Publication Date Title
US9792602B2 (en) Apparatus and method for providing emotion expression service in mobile communication terminal
EP2335400B1 (en) System and method for video telephony by converting facial motion to text
JPH0983630A (en) Telephone system
CN104539871B (en) Multimedia session method and device
JP2005525597A (en) Interactive control of electrical equipment
JP2019220848A (en) Data processing apparatus, data processing method and program
CN114845081A (en) Information processing apparatus, recording medium, and information processing method
CN105915521A (en) Multi-party communication management method, device and terminal
WO2018061173A1 (en) Tv conference system, tv conference method, and program
JP2010034695A (en) Voice response device and method
JPWO2019155735A1 (en) Information processing equipment, information processing methods and programs
CN106603381A (en) Chat information processing method and device
CN108509863A (en) Information cuing method, device and electronic equipment
JP2006253898A (en) Communication terminal
WO2006106671A1 (en) Image processing device, image display device, reception device, transmission device, communication system, image processing method, image processing program, and recording medium containing the image processing program
CN111108491A (en) Conference system
JP6846753B2 (en) Computer system, web conferencing audio assistance methods and programs
CN109587344A (en) Call control method, device, mobile terminal and medium based on mobile terminal
JP2009060220A (en) Communication system and communication program
KR20050079125A (en) Methods and a apparatus of setting normal display image and ringing signal for mobile phone
CN114222302A (en) Calling method and device for abnormal call, electronic equipment and storage medium
CN112700783A (en) Communication sound changing method, terminal equipment and storage medium
KR101677622B1 (en) Image display method and apparatus thereof
CN105376513A (en) Information transmission method and device
JP2015115926A (en) Portable terminal device, lip-reading communication method, and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application
DPE1 Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101)
NENP Non-entry into the national phase

Ref country code: DE

NENP Non-entry into the national phase

Ref country code: RU

122 Ep: pct application non-entry in european phase

Ref document number: 06730245

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP