WO2022052833A1 - Procédé de réglage de son de télévision, télévision et support d'enregistrement - Google Patents

Procédé de réglage de son de télévision, télévision et support d'enregistrement Download PDF

Info

Publication number
WO2022052833A1
WO2022052833A1 PCT/CN2021/115414 CN2021115414W WO2022052833A1 WO 2022052833 A1 WO2022052833 A1 WO 2022052833A1 CN 2021115414 W CN2021115414 W CN 2021115414W WO 2022052833 A1 WO2022052833 A1 WO 2022052833A1
Authority
WO
WIPO (PCT)
Prior art keywords
ear
preset
horizontal
left ear
area
Prior art date
Application number
PCT/CN2021/115414
Other languages
English (en)
Chinese (zh)
Inventor
徐遥令
李坚
洪文生
袁新艳
孙彦竹
Original Assignee
深圳创维-Rgb电子有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳创维-Rgb电子有限公司 filed Critical 深圳创维-Rgb电子有限公司
Publication of WO2022052833A1 publication Critical patent/WO2022052833A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • H04N21/44213Monitoring of end-user related data
    • H04N21/44218Detecting physical presence or behaviour of the user, e.g. using sensors to detect if the user is leaving the room or changes his face expression during a TV program
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/02Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/44Receiver circuitry for the reception of television signals according to analogue transmission standards
    • H04N5/60Receiver circuitry for the reception of television signals according to analogue transmission standards for the sound signals
    • H04N5/602Receiver circuitry for the reception of television signals according to analogue transmission standards for the sound signals for digital sound signals

Definitions

  • the present application relates to the field of smart TVs, and in particular, to a TV sound adjustment method, a TV, and a computer-readable storage medium.
  • the user When a user is watching TV, the user is in different viewing positions and experiences different sound effects.
  • the user can have the best sound effect only when the user is facing the center of the TV; when the user is in other areas, the pitch, timbre, loudness, etc. of the TV sound entering the left and right ears of the human will change, resulting in poor sound effect and affecting the sound experience.
  • the main purpose of the present application is to provide a TV sound adjustment method, a TV and a computer-readable storage medium, aiming to solve the problem of poor sound experience when existing users are not watching TV at the central point of the TV.
  • the present application provides a TV sound adjustment method, comprising the steps of:
  • left ear position information includes left ear vertical distance, left ear horizontal distance and left ear horizontal angle
  • right ear position information includes right ear vertical distance, Right ear horizontal distance and right ear horizontal angle
  • the preset camera angle of view, the preset standard definition, the preset standard vertical distance, the preset horizontal distance between the left and right speakers, the preset coefficient, the left ear area and the right ear area in the current target area image, and the steps of obtaining the left ear position information and the right ear position information include:
  • the position information of the right ear is obtained according to the right ear area, the preset camera field angle, the preset standard definition, the preset standard vertical distance, the preset horizontal distance between the left and right speakers, and the preset coefficient in the current target area image.
  • the step further includes:
  • the preset camera field angle When it is determined that only the left ear area exists in the current target area image, according to the left ear area in the current target area image, the preset camera field angle, the preset standard definition, the preset standard vertical distance, the preset left and right speaker horizontal spacing and Preset coefficients to obtain left ear position information;
  • the right ear position information is obtained.
  • the step further includes:
  • the preset camera field angle When it is determined that only the right ear area exists in the current target area image, according to the right ear area in the current target area image, the preset camera field angle, the preset standard definition, the preset standard vertical distance, the preset left and right speaker horizontal spacing and Preset coefficients to obtain right ear position information;
  • the left ear position information is obtained according to the right ear position information and the preset horizontal offset value.
  • the steps of left ear position information include:
  • the left ear horizontal distance according to the abscissa value of at least one pixel in the left ear area, the preset left and right speaker horizontal distance, the left ear vertical distance, the total number of horizontal pixels of the current target area image, and the preset camera field of view;
  • the steps of right ear position information include:
  • the horizontal distance of the right ear according to the abscissa value of at least one pixel in the right ear area, the preset horizontal distance between the left and right speakers, the vertical distance of the right ear, the total number of horizontal pixels of the current target area image, and the preset camera field angle;
  • the horizontal included angle of the right ear is obtained according to the horizontal distance of the right ear, the vertical distance of the right ear, and the preset horizontal distance between the left and right speakers.
  • the step of judging whether the left ear region and the right ear region both exist in the current target region image according to the current recognition result includes:
  • the current recognition result is the same as the recognition result corresponding to the previous sampling time, then the current left channel signal is adjusted according to the left channel difference parameter obtained at the previous sampling time, and the adjusted left channel signal is obtained. Adjust the current right channel signal with the right channel difference parameter obtained at the moment to obtain the adjusted right channel signal;
  • the step of obtaining the left channel difference parameter and the right channel difference parameter according to the left ear position information and the right ear position information includes:
  • the left ear horizontal distance, and the left ear vertical distance query the left channel difference parameters corresponding to the angle difference, the left ear horizontal distance, and the left ear vertical distance from the preset left channel difference parameters;
  • the right ear horizontal distance, and the right ear vertical distance query the right channel difference parameters corresponding to the angle difference, right ear horizontal distance and right ear vertical distance from a preset right channel difference parameter table .
  • the present application also provides a television comprising a memory, a processor and a computer program stored on the memory and executable on the processor, the computer program being executed by the processor When executed, it implements the steps of the TV sound adjustment method as described above.
  • the present application also provides a computer-readable storage medium, on which a computer program is stored, and when the computer program is executed by a processor, the steps of the above-mentioned TV sound adjustment method are realized. .
  • a TV sound adjustment method, a TV, and a computer-readable storage medium proposed in the present application collect user scenes through a smart TV camera, automatically identify the position of the user's head, locate the human ears, and calculate the regional positions where the ears are located, and the The angle difference between the ears, the pre-stored phase, frequency, and gain difference parameters are called according to the area position and the angle difference between the ears; then the difference parameters are used to adjust the decoded left channel signal and right channel signal
  • the phase, frequency and gain of the speaker are obtained, and the adjusted left channel signal and right channel signal are obtained; the adjusted left channel signal and right channel signal are driven by the left and right speakers of the speaker after driving processing, so that the user can produce sound.
  • the tone, timbre and loudness heard are always the same as when the user is at the center of the TV, so the sound the user hears is always the same as when the user is at the center of the TV.
  • the effect of the sound heard at the same time is the same, and the "sound moves with the ear" is realized, which provides the user with a good sound effect and improves the user experience.
  • FIG. 1 is a schematic structural diagram of a hardware operating environment involved in a solution according to an embodiment of the present application
  • FIG. 2 is a schematic flowchart of the first embodiment of the TV sound adjustment method of the application
  • FIG. 3 is a schematic flow chart of the refinement of step S40 in the first embodiment of the TV sound adjustment method of the present application;
  • FIG. 4 is a schematic flowchart of the second embodiment of the TV sound adjustment method of the present application.
  • FIG. 5 is a schematic flowchart of the third embodiment of the TV sound adjustment method of the present application.
  • FIG. 6 is a schematic flowchart of the fourth embodiment of the TV sound adjustment method of the present application.
  • FIG. 1 is a schematic diagram of a hardware structure of a TV provided in various embodiments of the present application.
  • the TV includes components such as a communication module 01 , a memory 02 and a processor 03 .
  • the processor 03 is respectively connected to the memory 02 and the communication module 01 , the memory 02 stores a computer program, and the computer program is simultaneously executed by the processor 03 .
  • the communication module 01 can be connected with external devices through the network.
  • the communication module 01 can receive data sent by an external device, and can also send data, instructions and information to the external device.
  • the external device can be an electronic device such as a mobile phone, a tablet computer, a notebook computer, and a desktop computer.
  • the memory 02 can be used to store software programs and various data.
  • the memory 02 may mainly include a stored program area and a stored data area, wherein the stored program area may store an operating system, an application program required for at least one function (according to the left ear position information and the right ear position information, the left channel difference parameters and Right channel difference parameter) etc.; the storage data area can store data or information etc. created according to the usage of the TV.
  • memory 02 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid state storage device.
  • the processor 03 is the control center of the TV, and uses various interfaces and lines to connect various parts of the entire TV, by running or executing the software programs and/or modules stored in the memory 02, and calling the data stored in the memory 02, Perform various functions of the TV and process data to monitor the TV as a whole.
  • the processor 03 may include one or more processing units; preferably, the processor 03 may integrate an application processor and a modem processor, wherein the application processor mainly processes the operating system, user interface and application programs, etc.
  • the processor mainly handles wireless communication. It can be understood that, the above-mentioned modulation and demodulation processor may also not be integrated into the processor 03 .
  • the above-mentioned television may further include a circuit control module, which is used for connecting with the commercial power to realize power control and ensure the normal operation of other components.
  • the TV structure shown in FIG. 1 does not constitute a limitation on the TV, and may include more or less components than the one shown, or combine some components, or arrange different components.
  • the TV sound adjustment method includes the steps:
  • Step S10 controlling the camera to collect the image of the target area with a preset sampling frequency
  • the TV has two left and right speakers, namely the left speaker and the right speaker, and the two speakers are in the same horizontal direction.
  • the camera can be built-in to the TV, or it can be an external device of the TV, and the camera is located on the vertical line of the connection between the left and right speakers.
  • the TV will control the camera to collect the image of the target area at the preset sampling frequency.
  • the sampling frequency can be preset by the manufacturer before leaving the factory, or can be preset by the user according to the actual needs, such as the sampling frequency is 5s, and the frequency is not used here. The specific value is limited.
  • the target area image is the image of the target area in front of the TV that is captured by the camera.
  • Step S20 performing image recognition on the currently collected target area image to obtain a current recognition result
  • Step S30 according to the current recognition result, determine whether the left ear area and the right ear area both exist in the current target area image;
  • the camera When the camera collects the image of the target area at the current sampling time, it will use face recognition technology to identify the current target area image, which can directly identify the area where the human ear is located, but because the human ear is located on both sides of the face, If the head is slightly deflected or the hair is covered, the ears may not be captured by the camera, which makes it difficult to capture the ears in most cases. Therefore, face recognition can also be performed to obtain the face in the image of the target area. It is located at a fixed position of the human face, and then can be indirectly and automatically located to the ear area of the human face according to the human face, and the ear area includes the left ear area and the right ear area. After image recognition, the current recognition result will be obtained.
  • the current recognition result may be the left ear area and the right ear area that exist in the image of the target area. It is also possible that in some cases, only the left ear area may exist in the image of the target area. area or the right ear area. After the recognition result is obtained, it is determined whether the left ear area and the right ear area exist in the current target area image.
  • Step S40 Obtain left ear position information and right ear position information according to the ear area in the current target area image, the preset camera angle of view, the preset standard definition, the preset standard vertical distance, the preset horizontal distance between the left and right speakers, and the preset coefficient.
  • Ear position information where the left ear position information includes the vertical distance of the left ear, the horizontal distance of the left ear and the horizontal angle of the left ear, and the position information of the right ear includes the vertical distance of the right ear, the horizontal distance of the right ear and the horizontal angle of the right ear;
  • the left ear position information includes the vertical distance of the left ear, the horizontal distance of the left ear and the horizontal angle of the left ear, and the position of the right ear.
  • the information includes the vertical distance of the right ear, the horizontal distance of the right ear, and the horizontal angle of the right ear.
  • the vertical distance of the left ear and the vertical distance of the right ear refer to the distance between the left ear and the plane where the TV is located and the distance between the right ear and the plane where the TV is located;
  • the horizontal distance refers to the distance from the left ear to the vertical line passing through the two speakers and is perpendicular to the plane of the TV;
  • the horizontal distance of the right ear refers to the distance from the right ear to the vertical line passing through the two speakers and is perpendicular to the TV.
  • the distance of the plane; the horizontal angle of the left ear refers to the projection point on the plane of the point determined by the horizontal distance of the left ear and the vertical distance of the left ear on the plane that is connected to the left and right speakers and perpendicular to the TV and the left speaker respectively.
  • the angle formed by the connection of the right speaker and the connection of the right speaker; the horizontal angle of the right ear refers to the point determined by the horizontal distance of the right ear and the vertical distance of the right ear on the plane passing through the connection of the left and right speakers and perpendicular to the TV.
  • the projection points on the plane are respectively the included angles formed by the connection line of the left speaker and the connection line of the right speaker.
  • the preset screening rules will be used first, such as the closest to the center of the image. For example, the face with the highest definition, select a face from the identified multiple faces, and then locate the ear area according to the selected face, and obtain the ear area as the recognition result.
  • FIG. 3 is a schematic flow chart of the refinement of step S40 in the first embodiment of the TV sound adjustment method according to the present application.
  • Step S40 includes:
  • Step S401 when it is determined that both the left ear area and the right ear area exist, according to the left ear area, the preset camera field of view, the preset standard definition, the preset standard vertical distance, and the preset left and right speaker levels in the current target area image. Spacing and preset coefficients to obtain left ear position information;
  • step S401 according to the left ear area in the current target area image, the preset camera field angle, the preset standard definition, the preset standard vertical distance, the preset horizontal distance between the left and right speakers, and the preset coefficient, obtain
  • the steps of left ear position information include:
  • Step S4011 performing spatial spectrum analysis on the left ear region in the current target region image to obtain the clarity of the left ear region;
  • Step S4012 obtaining the left ear vertical distance according to the left ear area definition, the preset standard definition, the preset standard vertical distance and the preset coefficient;
  • Step S4013 Obtain the left ear horizontal distance according to the abscissa value of at least one pixel in the left ear area, the preset left and right speaker horizontal distance, the left ear vertical distance, the total number of horizontal pixels of the current target area image, and the preset camera field of view. ;
  • Step S4014 Obtain the horizontal included angle of the left ear according to the horizontal distance of the left ear, the vertical distance of the left ear, and the preset horizontal distance between the left and right speakers.
  • the left ear vertical distance calculation formula is:
  • D left vertical ⁇ ⁇ X left ⁇ Ds/Xs, where D left vertical distance is the left ear, ⁇ is a preset coefficient, X left is the clarity of the left ear area, D s is the preset standard vertical distance, X s is Default standard definition.
  • the vertical distance of the left ear After obtaining the vertical distance of the left ear, select the pixel horizontal coordinate value of any pixel in the left ear area or obtain the average value of the pixel horizontal coordinate values of all pixels in the left ear area as the reference pixel horizontal coordinate value of the left ear area, Then input the horizontal coordinate value of the reference pixel of the left ear area, the vertical distance of the left ear, the preset horizontal distance between the left and right speakers, the total number of horizontal pixels of the current target area image and the preset camera field angle into the preset left ear horizontal distance calculation formula , obtain the horizontal distance of the left ear, where the calculation formula of the horizontal distance of the left ear is:
  • D left water is the horizontal distance of the left ear
  • is the preset coefficient
  • X left is the clarity of the left ear area
  • D s is the preset standard vertical distance
  • X s is the preset standard definition
  • ⁇ 0 is the preset camera.
  • Field of view N is the total number of horizontal pixels in the current target area image
  • Nleftwater is the horizontal coordinate value of the reference pixel in the left ear area.
  • the formula for the preset left ear horizontal angle is:
  • ⁇ left is the horizontal angle of the left ear
  • D left water is the horizontal distance of the left ear
  • D left drop is the vertical distance of the left ear
  • D x is the preset horizontal distance between the left and right speakers.
  • Step S402 Obtain right ear position information according to the right ear area, preset camera field angle, preset standard definition, preset standard vertical distance, preset horizontal distance between left and right speakers, and preset coefficients in the current target area image.
  • preset camera field angle preset standard definition
  • preset standard vertical distance preset left and right speaker horizontal distance and preset coefficient in the current target area image
  • step S402 includes:
  • Step S4021 performing spatial spectrum analysis on the right ear region in the current target region image to obtain the clarity of the right ear region;
  • Step S4022 obtain the vertical distance of the right ear according to the area definition of the right ear, the preset standard definition, the preset standard vertical distance and the preset coefficient;
  • Step S4023 Obtain the horizontal distance of the right ear according to the abscissa value of at least one pixel in the right ear area, the preset horizontal distance between the left and right speakers, the vertical distance of the right ear, the total number of horizontal pixels of the current target area image, and the preset camera field of view. ;
  • Step S4024 Obtain the horizontal included angle of the right ear according to the horizontal distance of the right ear, the vertical distance of the right ear, and the preset horizontal distance between the left and right speakers.
  • the right ear vertical distance calculation formula is:
  • D right vertical ⁇ ⁇ X right ⁇ Ds/Xs, where D right vertical distance is the right ear, ⁇ is a preset coefficient, X right is the clarity of the right ear area, D s is the preset standard vertical distance, X s is Default standard definition.
  • the vertical distance of the right ear After obtaining the vertical distance of the right ear, select the pixel horizontal coordinate value of any pixel in the right ear area or obtain the average value of the pixel horizontal coordinate values of all pixels in the right ear area as the reference pixel horizontal coordinate value of the right ear area, Then input the horizontal coordinate value of the reference pixel of the right ear area, the vertical distance of the right ear, the preset horizontal distance between the left and right speakers, the total number of horizontal pixels of the current target area image and the preset camera field angle into the preset right ear horizontal distance calculation formula , obtain the horizontal distance of the right ear, where the calculation formula of the horizontal distance of the right ear is:
  • D right water is the horizontal distance of the right ear
  • is the preset coefficient
  • X right is the definition of the right ear area
  • Ds is the preset standard vertical distance
  • Xs is the preset standard definition
  • ⁇ 0 is the preset camera field of view angle
  • N is the total number of horizontal pixels in the current target area image
  • N is the horizontal coordinate value of the reference pixel in the right ear area.
  • the formula for the preset horizontal angle of the right ear is:
  • ⁇ right is the horizontal angle of the right ear
  • D right water is the horizontal distance of the right ear
  • D right drop is the vertical distance of the right ear
  • D x is the preset horizontal distance between the left and right speakers.
  • Step S50 obtaining the left channel difference parameter and the right channel difference parameter according to the left ear position information and the right ear position information;
  • the left channel difference parameter includes the left channel difference phase ⁇ P_L , left channel difference frequency ⁇ F_L and left channel difference gain ⁇ A_L
  • right channel difference parameters include right channel difference phase ⁇ P_R, right channel difference frequency ⁇ F_R, right channel difference gain ⁇ A_R.
  • step S50 includes:
  • Step S51 according to the horizontal angle of the left ear and the horizontal angle of the right ear, obtain the angle difference between the horizontal angle of the left ear and the horizontal angle of the right ear;
  • Step S52 according to the angle difference, the left ear horizontal distance and the left ear vertical distance, from the preset left channel difference parameter table, look up the angle difference, the left ear horizontal distance and the left ear vertical distance corresponding to the left sound.
  • Road difference parameter
  • Step S53 according to the angle difference, the horizontal distance of the right ear and the vertical distance of the right ear, from the preset right channel difference parameter table, query the angle difference, the horizontal distance of the right ear and the vertical distance of the right ear corresponding to the right sound. track difference parameters.
  • a left channel difference parameter table and a right channel difference parameter table are preset in the TV. After obtaining the position information of the left ear and the right ear, obtain the angle difference between the horizontal angle of the left ear and the horizontal angle of the right ear according to the horizontal angle of the left ear and the horizontal angle of the right ear. The difference here is the absolute difference of the values. value.
  • a plurality of left channel difference parameters are stored in the left channel difference parameter table, and each left channel difference parameter corresponds to a different angle difference value, left ear horizontal distance and left ear vertical distance.
  • a plurality of right channel difference parameters are stored in the right channel difference parameter table, and each right channel difference parameter corresponds to a different angle difference value, right ear horizontal distance and right ear vertical distance.
  • the left channel difference parameters corresponding to the angle difference, the left ear horizontal distance and the left ear vertical distance are queried from the preset left channel difference parameter table.
  • the horizontal distance of the right ear and the vertical distance of the right ear, the right channel difference parameters corresponding to the angle difference, the horizontal distance of the right ear and the vertical distance of the right ear are inquired from the preset right channel difference parameter table.
  • step S52 and the execution of step S53 may be performed simultaneously, or step S52 may be performed first and then step S53 may be performed, or step S53 may be performed first and then step S52 may be performed, and the execution order of step S52 and step S53 is not performed here. limited.
  • Step S60 adjust the current left channel signal according to the left channel difference parameter, obtain the adjusted left channel signal, and adjust the current right channel signal according to the right channel difference parameter, and obtain the adjusted signal. the right channel signal;
  • Step S70 playing according to the adjusted left channel signal and the adjusted right channel signal.
  • the phase, frequency and gain of the current left channel signal L_S are P_L, F_L and A_L respectively
  • the current right channel The phase, frequency and gain of the signal R_S are P_R, F_R and A_R, respectively.
  • the signal function of the current left channel signal L_S is expressed as:
  • the signal function of the current right channel signal R_S is expressed as:
  • R_S ⁇ H(P_R, F_R, A_R).
  • left channel difference phase ⁇ P_L left channel difference frequency ⁇ F_L and left channel difference gain ⁇ A_L in the left channel difference parameters
  • the adjusted left channel signal, the signal function of the adjusted channel signal L_C is expressed as:
  • L_C ⁇ H(P_L+ ⁇ P_L, F_L+ ⁇ F_L, A_L ⁇ A_L),
  • right channel difference phase ⁇ P_R right channel difference frequency ⁇ F_R and right channel difference gain ⁇ A_R in the right channel difference parameter, respectively adjust the phase P_R, frequency F_R and gain A_R in the current right channel signal, and finally output
  • the adjusted right channel signal, the signal function of the adjusted channel signal R_C is expressed as:
  • R_C ⁇ H(P_R+ ⁇ P_R, F_R+ ⁇ F_R, A_R ⁇ A_R),
  • the left speaker driving signal L_D and the right speaker driving signal R_D are obtained, and finally the left speaker is driven according to the left speaker driving signal L_D for playback. , and at the same time drive the right speaker to play according to the right speaker driving signal R_D.
  • step S30 when the determination result in step S30 is that neither the left ear region nor the right ear region exists in the current region image, the decoded left and right channel signals will not be adjusted.
  • the smart TV camera captures the user scene, automatically identifies the user's head position and locates the human ear, calculates the regional position where the two ears are located, and the angle difference between the two ears, according to the regional position and the angle between the two ears. Then use the difference parameters to adjust the phase, frequency and gain of the decoded left channel signal and right channel signal, and obtain the adjusted left channel signal and right channel signal.
  • the left speaker and right speaker of the speaker are driven to emit sound, so that when the user is in different areas and the user's head is in different directions, the tones heard , timbre, and loudness are always the same as when the user is at the center of the TV, so the sound that the user hears is always the same as the sound when the user is at the center of the TV. Provide users with good sound effects and improve user experience.
  • FIG. 4 is a second embodiment of the TV sound adjustment method of the present application according to the first embodiment of the television sound adjustment method of the present application. After step S30, the method further includes:
  • Step S41 when it is determined that only the left ear area exists in the current target area image, according to the left ear area in the current target area image, the preset camera field of view, the preset standard definition, the preset standard vertical distance, and the preset left and right speakers. Horizontal spacing and preset coefficients to obtain left ear position information;
  • step S42 the position information of the right ear is obtained according to the position information of the left ear and the preset horizontal offset value.
  • the camera Since the user may be in the edge area of the scene that the camera can capture or the user's face is partially covered, the camera only captures part of the user's face, so when performing image recognition on the current area image, it may only be determined that the left ear area exists. .
  • the left ear area, preset camera field of view, preset standard definition, preset standard vertical distance, preset left and right speakers in the current target area image The horizontal distance and the preset coefficient are used to obtain left ear position information.
  • the specific algorithm and specific steps involved in the process of obtaining the left ear position information are the same as those in steps S4011 to S4013, and will not be described in detail here.
  • the right ear and the left ear are in the same horizontal direction and at the same height, but the distance between the left ear and the right ear in the horizontal direction is the width of the face.
  • the average width value of a face is obtained as the preset horizontal offset value.
  • the preset right ear horizontal distance second formula is:
  • D right water is the horizontal distance of the right ear
  • D left water is the horizontal distance of the left ear
  • D offset is the horizontal offset value.
  • the horizontal angle of the right ear is calculated by inputting the horizontal distance of the right ear, the vertical distance of the right ear and the preset horizontal distance of the left and right speakers into the formula of the preset horizontal angle of the right ear, and the preset right ear horizontal angle is calculated.
  • the formula for the included angle is:
  • ⁇ right is the horizontal angle of the right ear
  • D right water is the horizontal distance of the right ear
  • D right drop is the vertical distance of the right ear
  • D x is the preset horizontal distance between the left and right speakers.
  • This embodiment provides a strategy for obtaining left ear position information and right ear position information only according to the left ear region under the condition that only the left ear region can be obtained in the target region image.
  • FIG. 5 is a third embodiment of the TV sound adjustment method of the present application according to the first embodiment and the second embodiment of the television sound adjustment method of the present application. After step S30, the method further includes:
  • Step S43 when it is determined that only the right ear area exists in the current target area image, according to the right ear area, preset camera field of view, preset standard definition, preset standard vertical distance, preset left and right speakers in the current target area image Horizontal spacing and preset coefficients to obtain right ear position information;
  • Step S44 Obtain left ear position information according to the right ear position information and the preset horizontal offset value.
  • the camera Since the user may be in the edge area of the scene that the camera can capture or the user's face is partially covered, the camera only captures part of the user's face, so when performing image recognition on the current area image, it may only be determined that the left ear area exists. .
  • the current target area image When it is determined that there is only one right ear area in the current image of the current area, the current target area image will be based on the right ear area, the preset camera field of view, the preset standard definition, the preset standard vertical distance, and the preset left and right speakers.
  • the horizontal distance and the preset coefficient are used to obtain the position information of the right ear.
  • the specific algorithm and specific steps involved in the process of obtaining the position information of the right ear are the same as those in steps S4021 to S4023, and will not be described in detail here.
  • the right ear and the left ear are in the same horizontal direction and at the same height, but the distance between the left ear and the right ear in the horizontal direction is the width of the face.
  • the average width value of a face is obtained as the preset horizontal offset value.
  • the horizontal distance of the left ear first, according to the position of the image of the current target area of the right ear, it can be determined whether the point projected by the right ear on the TV is in the area where the camera and the left speaker are located in the TV, or in the TV where the camera and the right speaker are located. area. Then input the right ear horizontal distance and the horizontal offset value into the second preset left ear horizontal distance formula, and calculate the left ear horizontal distance.
  • the preset left ear horizontal distance second formula is:
  • D left water is the horizontal distance of the left ear
  • D right water is the horizontal distance of the right ear
  • D offset is the horizontal offset value.
  • the horizontal angle of the left ear is calculated by inputting the horizontal distance of the left ear, the vertical distance of the left ear and the preset horizontal distance of the left and right speakers into the formula of the preset horizontal angle of the left ear, and the preset left ear horizontal angle is calculated.
  • the formula for the included angle is:
  • ⁇ left is the horizontal angle of the left ear
  • D left water is the horizontal distance of the left ear
  • D left drop is the vertical distance of the left ear
  • D x is the preset horizontal distance between the left and right speakers.
  • This embodiment provides a strategy for obtaining left ear position information and right ear position information only according to the right ear region when only the right ear region can be obtained in the target region image.
  • FIG. 6 is a fourth embodiment of the TV sound adjustment method of the present application according to the aforementioned embodiment of the television sound adjustment method of the present application.
  • the steps before step S30 include:
  • Step S80 judging whether the current recognition result is the same as the recognition result corresponding to the previous sampling moment; the current recognition result is not the same as the recognition result corresponding to the previous sampling moment, then step S30 is performed; the current recognition result and the recognition corresponding to the previous sampling moment If the result is the same, step S81 is executed;
  • Step S81 adjust the current left channel signal according to the left channel difference parameter obtained at the previous sampling time, obtain the adjusted left channel signal, and adjust the current right channel signal according to the right channel difference parameter obtained at the previous sampling time. Adjust the channel signal to obtain the adjusted right channel signal;
  • Step S82 playing according to the adjusted left channel signal and the adjusted right channel signal.
  • step S30 is executed, that is, according to the current recognition result, it is judged whether both the left ear region and the right ear region exist in the current target region image.
  • the current recognition result is the same as the recognition result corresponding to the previous sampling time, then the currently received audio signal is decoded to obtain the current left channel signal and the current right channel signal, according to the difference of the left channel obtained at the previous sampling time
  • the parameter adjusts the current left channel signal to obtain the adjusted left channel signal, and adjusts the current right channel signal according to the right channel difference parameter obtained at the previous sampling moment to obtain the adjusted right channel signal.
  • the adjusted left channel signal and right channel signal are processed by signal amplification, amplitude limiting, etc., to obtain the left speaker drive signal and the right speaker drive signal, and finally drive the left speaker according to the left speaker drive signal for playback, and at the same time according to The right speaker drive signal drives the right speaker for playback.
  • the left ear position information and the right ear position information at each sampling moment it is first determined whether the position of the left ear area and the position of the right ear area in the target area image collected at the current sampling moment are the same as those collected at the previous sampling moment. Whether the position of the left ear area and the position of the right ear area in the target area image are the same, in the same case, you do not need to determine the ear position information, and directly use the channel difference parameter obtained last time as the current channel difference parameter, which can reduce Calculation process, improve processing speed.
  • the present application also proposes a computer-readable storage medium on which a computer program is stored.
  • the computer-readable storage medium may be the memory 02 in the television of FIG. 1, or may be, for example, ROM (Read-Only Memory, read-only memory)/RAM (Random Access Memory, random access memory), magnetic disk, optical disk
  • ROM Read-Only Memory, read-only memory
  • RAM Random Access Memory, random access memory
  • At least one of the computer-readable storage medium includes several pieces of information to enable the television to perform the methods described in the various embodiments of the present application.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Social Psychology (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Databases & Information Systems (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Computational Linguistics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • Stereophonic System (AREA)

Abstract

La présente demande concerne un procédé de réglage de son de télévision, une télévision et un support d'enregistrement. Le procédé comprend : l'acquisition et la reconnaissance d'une image de zone cible pour obtenir un résultat de reconnaissance ; en fonction du résultat de reconnaissance, la détermination de l'existence ou non d'une zone d'oreille gauche et d'une zone d'oreille droite ; si les zones existent, selon la zone d'oreille gauche, la zone d'oreille droite, un champ de vision d'une caméra, une définition standard, une distance verticale standard, une distance horizontale entre des haut-parleurs gauche et droit et des coefficients prédéfinis, l'acquisition d'informations de position d'oreille gauche et d'informations de position d'oreille droite ; l'acquisition de paramètres de différence de canal gauche et de paramètres de différence de canal droit en fonction des informations de position d'oreille gauche et des informations de position d'oreille droite ; l'ajustement d'un signal de canal gauche et d'un signal de canal droit selon les paramètres de différence de canal gauche et les paramètres de différence de canal droit pour obtenir un signal de canal gauche ajusté et un signal de canal droit ajusté ; et la réalisation d'une lecture selon le signal de canal gauche ajusté et le signal de canal droit ajusté.
PCT/CN2021/115414 2020-09-10 2021-08-30 Procédé de réglage de son de télévision, télévision et support d'enregistrement WO2022052833A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010951644.4A CN112073804B (zh) 2020-09-10 2020-09-10 电视声音调整方法、电视及存储介质
CN202010951644.4 2020-09-10

Publications (1)

Publication Number Publication Date
WO2022052833A1 true WO2022052833A1 (fr) 2022-03-17

Family

ID=73696165

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/115414 WO2022052833A1 (fr) 2020-09-10 2021-08-30 Procédé de réglage de son de télévision, télévision et support d'enregistrement

Country Status (2)

Country Link
CN (1) CN112073804B (fr)
WO (1) WO2022052833A1 (fr)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112073804B (zh) * 2020-09-10 2022-05-20 深圳创维-Rgb电子有限公司 电视声音调整方法、电视及存储介质
CN112380972B (zh) * 2020-11-12 2022-03-15 四川长虹电器股份有限公司 一种应用于电视场景的音量调节方法
CN113055810A (zh) * 2021-03-05 2021-06-29 广州小鹏汽车科技有限公司 音效控制方法、装置、系统、车辆以及存储介质
CN113301329B (zh) * 2021-05-21 2022-08-05 康佳集团股份有限公司 基于图像识别的电视声场校正方法、装置及显示设备
CN114157905B (zh) * 2021-11-22 2023-12-05 深圳康佳电子科技有限公司 基于图像识别的电视声音调节方法、装置、电视机
CN115022716A (zh) * 2022-05-09 2022-09-06 海信视像科技股份有限公司 显示设备、声场调整方法、存储介质和程序产品

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070154019A1 (en) * 2005-12-22 2007-07-05 Samsung Electronics Co., Ltd. Apparatus and method of reproducing virtual sound of two channels based on listener's position
CN102687522A (zh) * 2009-09-15 2012-09-19 索尼公司 显示装置和控制方法
CN102685419A (zh) * 2011-03-11 2012-09-19 索尼公司 音频装置和音频系统
CN103187080A (zh) * 2011-12-27 2013-07-03 启碁科技股份有限公司 电子装置及播放方法
US9277343B1 (en) * 2012-06-20 2016-03-01 Amazon Technologies, Inc. Enhanced stereo playback with listener position tracking
CN107484082A (zh) * 2017-08-08 2017-12-15 广东小天才科技有限公司 一种基于声道控制音频信号传输的方法及用户终端
CN112073804A (zh) * 2020-09-10 2020-12-11 深圳创维-Rgb电子有限公司 电视声音调整方法、电视及存储介质

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100416757B1 (ko) * 1999-06-10 2004-01-31 삼성전자주식회사 위치 조절이 가능한 가상 음상을 이용한 스피커 재생용 다채널오디오 재생 장치 및 방법
CN102064781B (zh) * 2010-10-29 2015-09-09 华为终端有限公司 一种终端音频的调整方法、装置和终端
US9456266B2 (en) * 2012-04-27 2016-09-27 Brül & Kjær Sound & Vibration Measurement A/S Human like ear simulator
CN103826194B (zh) * 2014-02-28 2015-06-03 武汉大学 一种多声道系统中声源方向和距离重建的方法与装置
CN108307261A (zh) * 2017-01-11 2018-07-20 中兴通讯股份有限公司 一种自适应耳机声道切换方法和装置
US9883278B1 (en) * 2017-04-18 2018-01-30 Nanning Fugui Precision Industrial Co., Ltd. System and method for detecting ear location of earphone and rechanneling connections accordingly and earphone using same
CN107071648B (zh) * 2017-06-19 2020-01-10 深圳市泰衡诺科技有限公司上海分公司 声音播放调节系统、装置及方法

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070154019A1 (en) * 2005-12-22 2007-07-05 Samsung Electronics Co., Ltd. Apparatus and method of reproducing virtual sound of two channels based on listener's position
CN102687522A (zh) * 2009-09-15 2012-09-19 索尼公司 显示装置和控制方法
CN102685419A (zh) * 2011-03-11 2012-09-19 索尼公司 音频装置和音频系统
CN103187080A (zh) * 2011-12-27 2013-07-03 启碁科技股份有限公司 电子装置及播放方法
US9277343B1 (en) * 2012-06-20 2016-03-01 Amazon Technologies, Inc. Enhanced stereo playback with listener position tracking
CN107484082A (zh) * 2017-08-08 2017-12-15 广东小天才科技有限公司 一种基于声道控制音频信号传输的方法及用户终端
CN112073804A (zh) * 2020-09-10 2020-12-11 深圳创维-Rgb电子有限公司 电视声音调整方法、电视及存储介质

Also Published As

Publication number Publication date
CN112073804A (zh) 2020-12-11
CN112073804B (zh) 2022-05-20

Similar Documents

Publication Publication Date Title
WO2022052833A1 (fr) Procédé de réglage de son de télévision, télévision et support d'enregistrement
US20210334066A1 (en) Devices with enhanced audio
US20160353223A1 (en) System and method for dynamic control of audio playback based on the position of a listener
EP4054177B1 (fr) Procédé et dispositif de traitement audio
US20150078595A1 (en) Audio accessibility
WO2018149275A1 (fr) Procédé et appareil d'ajustement d'une sortie audio par un haut-parleur
US9263044B1 (en) Noise reduction based on mouth area movement recognition
US8218033B2 (en) Sound corrector, sound recording device, sound reproducing device, and sound correcting method
EP2046032A1 (fr) Procede et dispositif pour obtenir des informations de position initiale acoustiques et systeme e communication multimedia
US9277343B1 (en) Enhanced stereo playback with listener position tracking
CN113676592B (zh) 录音方法、装置、电子设备及计算机可读介质
US20130329921A1 (en) Optically-controlled speaker system
US9351073B1 (en) Enhanced stereo playback
CN113014844A (zh) 一种音频处理方法、装置、存储介质及电子设备
CN111081285A (zh) 一种调整特效的方法、电子设备及存储介质
US11087435B1 (en) Adaptive dewarping of wide angle video frames
CN114157905B (zh) 基于图像识别的电视声音调节方法、装置、电视机
GB2610460A (en) Information processing method and electronic device
CN112333531A (zh) 音频数据播放方法、设备及可读存储介质
KR20090016289A (ko) 노트북 컴퓨터의 스피커 제어 장치 및 방법
US11227396B1 (en) Camera parameter control using face vectors for portal
KR102650763B1 (ko) 오디오 소스 지향성에 기초한 심리음향 강화
US10902864B2 (en) Mixed-reality audio intelligibility control
CN113709652B (zh) 音频播放控制方法和电子设备
US20230105785A1 (en) Video content providing method and video content providing device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21865888

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 18/07/2023)

122 Ep: pct application non-entry in european phase

Ref document number: 21865888

Country of ref document: EP

Kind code of ref document: A1