WO2022001407A1 - Camera control method and display device - Google Patents

Camera control method and display device Download PDF

Info

Publication number
WO2022001407A1
WO2022001407A1 PCT/CN2021/093589 CN2021093589W WO2022001407A1 WO 2022001407 A1 WO2022001407 A1 WO 2022001407A1 CN 2021093589 W CN2021093589 W CN 2021093589W WO 2022001407 A1 WO2022001407 A1 WO 2022001407A1
Authority
WO
WIPO (PCT)
Prior art keywords
camera
angle
area
portrait
image
Prior art date
Application number
PCT/CN2021/093589
Other languages
French (fr)
Chinese (zh)
Inventor
杨鲁明
王依林
朱铄
孙永江
Original Assignee
海信视像科技股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 海信视像科技股份有限公司 filed Critical 海信视像科技股份有限公司
Publication of WO2022001407A1 publication Critical patent/WO2022001407A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/67Focus control based on electronic image sensor signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects
    • H04N23/611Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/62Control of parameters via user interfaces
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/67Focus control based on electronic image sensor signals
    • H04N23/675Focus control based on electronic image sensor signals comprising setting of focusing regions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/69Control of means for changing angle of the field of view, e.g. optical zoom objectives or electronic zooming

Definitions

  • the present application relates to the technical field of television software, and in particular, to a control method of a camera and a display device.
  • the display device can implement functions such as network search, IP TV, BBTV, video on demand (VOD), digital music, network news, and network video telephony.
  • VOD video on demand
  • a camera needs to be installed on the display device to collect user images.
  • An embodiment of the present application provides a display device, including:
  • the camera is configured to capture a portrait and realize rotation within a preset angle range
  • the shooting angle of the camera is adjusted so that the portrait of the person is located in the center area of the designated image captured by the camera.
  • FIG. 1 exemplarily shows a schematic diagram of an operation scene between a display device and a control apparatus according to some embodiments
  • FIG. 2 exemplarily shows a hardware configuration block diagram of a display device 200 according to some embodiments
  • FIG. 3 exemplarily shows a hardware configuration block diagram of the control device 100 according to some embodiments
  • FIG. 4 exemplarily shows a schematic diagram of software configuration in the display device 200 according to some embodiments
  • FIG. 5 exemplarily shows a schematic diagram of displaying an icon control interface of an application in the display device 200 according to some embodiments
  • FIG. 6 exemplarily shows a structural block diagram of a display device according to some embodiments.
  • FIG. 7 exemplarily shows a schematic diagram of implementing a preset angle range for camera rotation according to some embodiments
  • FIG. 8 exemplarily shows a scene graph of camera rotation within a preset angle range according to some embodiments
  • FIG. 9 exemplarily shows a schematic diagram of a sound source angle range according to some embodiments.
  • FIG. 10 exemplarily shows a flowchart of a method for adjusting the shooting angle of a camera according to some embodiments
  • FIG. 11 exemplarily shows a flowchart of a wake-up text comparison method according to some embodiments
  • FIG. 12 exemplarily shows a flowchart of a method for performing sound source identification on character sound source information according to some embodiments
  • FIG. 13 exemplarily shows a flowchart of a method for determining a target rotation direction and a target rotation angle of a camera according to some embodiments
  • FIG. 14 exemplarily shows a scene diagram of adjusting the shooting angle of the camera according to some embodiments
  • FIG. 15 exemplarily shows another scene diagram of adjusting the shooting angle of the camera according to some embodiments.
  • FIG. 16 exemplarily shows a scene diagram of the position of a character when speaking according to some embodiments
  • FIG. 17 exemplarily shows another scene diagram in which the camera rotates within a preset angle range according to some embodiments
  • FIG. 18 exemplarily shows a flowchart of a method for controlling a camera according to some embodiments
  • FIG. 19 exemplarily shows an overall data flow diagram of a camera control method according to some embodiments.
  • FIG. 20 exemplarily shows a flowchart of a method for calculating an azimuth distance according to some embodiments
  • FIG. 21 exemplarily shows a schematic diagram of calculating azimuth distance according to some embodiments.
  • FIG. 22 exemplarily shows a schematic diagram of the horizontal viewing angle of a camera according to some embodiments.
  • Figure 23 exemplarily shows a schematic diagram of calculating a target horizontal adjustment angle according to some embodiments
  • FIG. 24 exemplarily shows a schematic diagram of a vertical viewing angle of a camera according to some embodiments.
  • FIG. 25 exemplarily shows a schematic diagram of calculating a vertical adjustment angle of a target according to some embodiments
  • FIG. 26 exemplarily shows a flowchart of a method for focusing and zooming in on a portrait display according to some embodiments
  • FIG. 27 exemplarily shows a schematic diagram of zoomed-in portrait display according to some embodiments.
  • module refers to any known or later developed hardware, software, firmware, artificial intelligence, fuzzy logic or combination of hardware or/and software code capable of performing the function associated with that element.
  • remote control refers to a component of an electronic device, such as the display device disclosed in this application, that can wirelessly control the electronic device, usually over a short distance.
  • infrared and/or radio frequency (RF) signals and/or Bluetooth are used to connect with electronic devices, and functional modules such as WiFi, wireless USB, Bluetooth, and motion sensors may also be included.
  • RF radio frequency
  • a hand-held touch remote control replaces most of the physical built-in hard keys in a general remote control device with a user interface in a touch screen.
  • gesture used in this application refers to a user's behavior that is used by a user to express an expected thought, action, purpose/or result through an action such as a change of hand shape or hand movement.
  • FIG. 1 is a schematic diagram of an operation scenario between a display device and a control apparatus according to an embodiment. As shown in FIG. 1 , a user can operate the display device 200 through the smart device 300 or the control device 100 .
  • the control apparatus 100 may be a remote controller, and the communication between the remote controller and the display device includes infrared protocol communication or Bluetooth protocol communication, and other short-distance communication methods, and the display device 200 is controlled wirelessly or wiredly.
  • the user can control the display device 200 by inputting user instructions through keys on the remote control, voice input, control panel input, and the like.
  • a smart device 300 eg, a mobile terminal, a tablet computer, a computer, a notebook computer, etc.
  • the display device 200 is controlled using an application running on the smart device.
  • the display device 200 can also be controlled in a manner other than the control apparatus 100 and the smart device 300.
  • the module for acquiring voice commands configured inside the display device 200 can directly receive the user's voice command for control.
  • the user's voice command control can also be received through a voice control device provided outside the display device 200 device.
  • the display device 200 is also in data communication with the server 400 .
  • the display device 200 may be allowed to communicate via local area network (LAN), wireless local area network (WLAN), and other networks.
  • the server 400 may provide various contents and interactions to the display device 200 .
  • FIG. 2 exemplarily shows a configuration block diagram of the control apparatus 100 according to an exemplary embodiment.
  • the control device 100 includes a controller 110 , a communication interface 130 , a user input/output interface 140 , a memory 190 , and a power supply 180 .
  • the control device 100 can receive the user's input operation instruction, and convert the operation instruction into an instruction that the display device 200 can recognize and respond to, and play an intermediary role between the user and the display device 200 .
  • FIG. 2 is a block diagram showing a hardware configuration of a display device 200 according to an exemplary embodiment.
  • the display device 200 includes at least one of a tuner and demodulator 210 , a communicator 220 , a detector 230 , an external device interface 240 , a controller 250 , a display 275 , an audio output interface 285 , a memory 260 , a power supply 290 , and a user interface 265 .
  • the display 275 includes a display screen component for presenting pictures, and a driving component for driving image display, for receiving image signals output from the controller, components for displaying video content, image content, and menu manipulation interfaces, and user manipulation UI interfaces .
  • the display 275 may be a liquid crystal display, an OLED display, and a projection display, and may also be a projection device and a projection screen.
  • the communicator 220 is a component for communicating with external devices or servers according to various communication protocol types.
  • the communicator may include at least one of a Wifi module, a Bluetooth module, a wired Ethernet module and other network communication protocol chips or near field communication protocol chips, and an infrared receiver.
  • the display device 200 may establish transmission and reception of control signals and data signals with the external control device 100 or the server 400 through the communicator 220 .
  • the user interface can be used to receive control signals from the control device 100 (eg, an infrared remote control, etc.).
  • control device 100 eg, an infrared remote control, etc.
  • the detector 230 is used to collect external environment or external interaction signals.
  • the detector 230 includes a light receiver, a sensor for collecting ambient light intensity; alternatively, the detector 230 includes an image collector, such as a camera, which can be used to collect external environmental scenes, user attributes or user interaction gestures, or , the detector 230 includes a sound collector, such as a microphone, for receiving external sound.
  • the external device interface 240 may include, but is not limited to, the following: any one of high-definition multimedia interface (HDMI), analog or data high-definition component input interface (component), composite video input interface (CVBS), USB input interface (USB), RGB port, etc. or multiple interfaces. It may also be a composite input/output interface formed by a plurality of the above-mentioned interfaces.
  • HDMI high-definition multimedia interface
  • component analog or data high-definition component input interface
  • CVBS composite video input interface
  • USB USB input interface
  • RGB port etc.
  • It may also be a composite input/output interface formed by a plurality of the above-mentioned interfaces.
  • the controller 250 and the tuner 210 may be located in different separate devices, that is, the tuner 210 may also be located in an external device of the main device where the controller 250 is located, such as an external set-top box.
  • the controller 250 controls the operation of the display device and responds to user operations through various software control programs stored in the memory 260 .
  • the controller 250 controls the overall operation of the display apparatus 200 . For example, in response to receiving a user command for selecting a UI object to be displayed on the display 275, the controller 250 may perform an operation related to the object selected by the user command.
  • Objects can be any of the optional objects, such as hyperlinks, icons, or other actionable controls.
  • the operations related to the selected object include: displaying operations connected to hyperlinked pages, documents, images, etc., or executing operations of programs corresponding to the icons.
  • the user may input user commands on a graphical user interface (GUI) displayed on the display 275, and the user input interface receives the user input commands through the graphical user interface (GUI).
  • GUI graphical user interface
  • the user may input a user command by inputting a specific sound or gesture, and the user input interface recognizes the sound or gesture through a sensor to receive the user input command.
  • GUI Graphical User Interface
  • control can include icons, buttons, menus, tabs, text boxes, dialog boxes, status bars, navigation bars, Widgets, etc. visual interface elements.
  • the system is divided into four layers, from top to bottom, they are an application layer (referred to as “application layer”), an application framework layer (referred to as “framework layer”) ”), the Android runtime and the system library layer (referred to as the “system runtime layer”), and the kernel layer.
  • application layer an application layer
  • frame layer an application framework layer
  • Android runtime the Android runtime
  • system library layer the system library layer
  • kernel layer the kernel layer
  • At least one application program runs in the application program layer, and these application programs may be a Window program, a system setting program, a clock program, a camera application, etc. built into the operating system; they may also be developed by a third party
  • the application programs developed by the author such as the Hijian program, the K song program, the magic mirror program, etc.
  • the application package in the application layer is not limited to the above examples, and may actually include other application packages, which is not limited in this embodiment of the present application.
  • the framework layer provides an application programming interface (API) and a programming framework for applications in the application layer.
  • the application framework layer includes some predefined functions.
  • the application framework layer is equivalent to a processing center, which decides to let the applications in the application layer take action.
  • the application program can access the resources in the system and obtain the services of the system during execution through the API interface.
  • the application framework layer in the embodiment of the present application includes managers (Managers), content providers (Content Provider), etc., wherein the manager includes at least one of the following modules: an activity manager (Activity Manager) uses Interacts with all activities running in the system; Location Manager is used to provide system services or applications with access to system location services; Package Manager is used to retrieve files currently installed on the device Various information related to the application package; Notification Manager (Notification Manager) is used to control the display and clearing of notification messages; Window Manager (Window Manager) is used to manage icons, windows, toolbars, wallpapers on the user interface and desktop widgets.
  • an activity manager uses Interacts with all activities running in the system
  • Location Manager is used to provide system services or applications with access to system location services
  • Package Manager is used to retrieve files currently installed on the device Various information related to the application package
  • Notification Manager Notification Manager
  • Window Manager Window Manager
  • the activity manager is used to: manage the life cycle of each application and the usual navigation and fallback functions, such as controlling the exit of the application (including switching the user interface currently displayed in the display window to the system desktop), opening the , back (including switching the currently displayed user interface in the display window to the upper-level user interface of the currently displayed user interface), and the like.
  • the window manager is used to manage all window programs, such as obtaining the size of the display screen, judging whether there is a status bar, locking the screen, taking screenshots, and controlling the change of the display window (for example, reducing the display window to display, shaking display, twisting deformation display, etc.), etc.
  • the system runtime layer provides support for the upper layer, that is, the framework layer.
  • the Android operating system will run the C/C++ library included in the system runtime layer to implement the functions to be implemented by the framework layer.
  • the kernel layer is the layer between hardware and software. As shown in Figure 4, the kernel layer at least includes at least one of the following drivers: audio driver, display driver, Bluetooth driver, camera driver, WIFI driver, USB driver, HDMI driver, sensor driver (such as fingerprint sensor, temperature sensor, touch sensors, pressure sensors, etc.), etc.
  • the kernel layer at least includes at least one of the following drivers: audio driver, display driver, Bluetooth driver, camera driver, WIFI driver, USB driver, HDMI driver, sensor driver (such as fingerprint sensor, temperature sensor, touch sensors, pressure sensors, etc.), etc.
  • the kernel layer further includes a power driver module for power management.
  • software programs and/or modules corresponding to the software architecture in FIG. 4 are stored in the first memory or the second memory shown in FIG. 2 or FIG. 3 .
  • the remote control receiving device receives the input operation of the remote control
  • the corresponding hardware interrupt is sent to the kernel layer.
  • the kernel layer processes the input operation into the original input event (including the value of the input operation, the timestamp of the input operation and other information).
  • Raw input events are stored at the kernel layer.
  • the application framework layer obtains the original input event from the kernel layer, identifies the control corresponding to the input event according to the current position of the focus, and regards the input operation as a confirmation operation, and the control corresponding to the confirmation operation is the control of the magic mirror application icon.
  • the mirror application calls the interface of the application framework layer, starts the mirror application, and then starts the camera driver by calling the kernel layer to capture still images or videos through the camera.
  • the display device receives an input operation (such as a split-screen operation) performed by the user on the display screen, and the kernel layer can generate corresponding input operations according to the input operation. Enter an event and report the event to the application framework layer.
  • the window mode (such as multi-window mode) and window position and size corresponding to the input operation are set by the activity manager of the application framework layer.
  • the window management of the application framework layer draws the window according to the settings of the activity manager, and then sends the drawn window data to the display driver of the kernel layer, and the display driver displays the corresponding application interface in different display areas of the display screen.
  • the application layer contains at least one application that can display corresponding icon controls in the display, such as: live TV application icon control, video on demand application icon control, media center application Program icon controls, application center icon controls, game application icon controls, etc.
  • the live TV application may provide live TV from different sources.
  • a live TV application may provide a TV signal using input from cable, over-the-air, satellite services, or other types of live TV services.
  • the live TV application may display the video of the live TV signal on the display device 200 .
  • a video-on-demand application may provide video from various storage sources. Unlike live TV applications, video-on-demand provides a display of video from certain storage sources. For example, video-on-demand can come from the server side of cloud storage, from local hard disk storage containing existing video programs.
  • the media center application may provide various multimedia content playback applications.
  • a media center may provide services other than live TV or video-on-demand, where users can access various images or audio through a media center application.
  • the application center may provide storage of various applications.
  • An application can be a game, an application, or some other application that is related to a computer system or other device but can be run on a Smart TV.
  • the application center can obtain these applications from various sources, store them in local storage, and then run them on the display device 200 .
  • the application programs that need to use the camera in the display device include “Hey see”, “Look in the mirror”, “Youxuemao”, “Fitness”, etc., which can realize “video chat", “watch while chatting” ” and “Fitness”.
  • “See you” is a video chat application that can realize one-click chat between mobile phone and TV, and between TV and TV.
  • "Looking in the Mirror” is an application that provides users with mirror services. By turning on the camera through the mirroring application, users can use the smart TV as a mirror.
  • “Youxuemao” is an application that provides learning functions.
  • the "fitness” function can simultaneously display the fitness instruction video and the image of the user following the fitness instruction video to perform corresponding actions on the display of the display device, so that users can check whether their actions are standard in real time.
  • the camera is fixedly installed on the display device, the center line of the viewing angle of the camera is perpendicular to the display, and the viewing angle of the camera is limited, usually between 60° and 75°, that is, the shooting area of the camera is The center line of the angle of view of the camera is synchronously spread to the left and right to form an area corresponding to an angle of 60° to 75°.
  • the camera cannot capture an image containing the user's portrait, so that the portrait cannot be displayed on the monitor. If in the video chat call scenario, the opposite end user who is in the video chat call with the local user will not be able to see the local user; if in the fitness scenario, the display will not be able to display the image of the user presenting the fitness action, and the user will not be able to see it. Your own fitness movements will not be able to judge whether they are standard or not, which will affect the user experience.
  • FIG. 6 exemplarily shows a structural block diagram of a display device according to some embodiments.
  • the camera is used to capture portraits.
  • the camera is no longer fixedly installed, but is rotatably installed on the display device.
  • the camera 232 is installed on the top of the display in a rotating form, and the camera 232 can rotate along the top of the display.
  • FIG. 7 exemplarily shows a schematic diagram of implementing a preset angle range of camera rotation according to some embodiments
  • FIG. 8 exemplarily shows a scene diagram of camera rotation within the preset angle range according to some embodiments.
  • the preset camera 232 can be rotated within a preset angle range and rotated in a horizontal direction.
  • the preset angle range is 0° ⁇ 120°, that is, at the position facing the display, the left side of the user is 0° and the right side of the user is 120°.
  • the camera can be rotated 60° to the left from the initial state, and 60° to the right from the initial state; the center line of the viewing angle of the camera is perpendicular to the display.
  • the position is the 60° position of the camera.
  • the display device provided by the embodiment of the present application realizes the use of sound source information to trigger the rotation of the camera, can automatically identify the real-time location of the user and adjust the shooting angle of the camera, so that the camera can always capture images including portraits.
  • the display device implements the collection of the sound source information of the person by setting the sound collector 231 .
  • multiple sets of sound collectors can be set in the display device.
  • four sets of sound collectors 231 are set in the display device, and the four sets of sound collectors 231 can be arranged in a linear positional relationship. set up.
  • the sound collector may be a microphone, and four groups of microphones are linearly arranged to form a microphone array. During sound collection, the four groups of sound collectors 231 receive sound information generated when the same user interacts with the display device through voice.
  • FIG. 9 A schematic diagram of a sound source angle range according to some embodiments is exemplarily shown in FIG. 9 .
  • the angle of the sound source generated by the user ranges from 0° to 180°.
  • the user is on the back of the display device, the user The generated sound source angle also ranges from 0° to 180°.
  • the user is located on the left side of the sound collector, which is 0° horizontally, and the user is located on the right side of the sound collector, which is 180° horizontally.
  • the 30° angular position of the sound source is equal to the 0° angular position of the camera
  • the 90° angular position of the sound source is equal to the 60° angular position of the camera
  • the 150° angular position of the sound source is equal to the 120° angular position of the camera corner position.
  • the controller 250 is connected with the camera 232 and the sound collector 231 respectively, and the controller is used to receive the sound source information of the character collected by the sound collector, identify the sound source information of the character, determine the azimuth angle of the position of the character, and then determine the camera. The angle that needs to be turned.
  • the controller adjusts the shooting angle of the camera according to the determined angle that the camera needs to rotate, so that the shooting area of the camera is facing the position of the voice of the character, and adjusts the shooting angle of the camera according to the position of the character to capture the image containing the character.
  • FIG. 10 exemplarily shows a flowchart of a method for adjusting the shooting angle of a camera according to some embodiments.
  • the controller when adjusting the shooting angle of the camera according to the position of the character, the controller is configured to execute the method for adjusting the shooting angle of the camera shown in FIG. 10 , including:
  • the controller in the display device drives the camera to rotate to adjust the shooting angle of the camera, it needs to determine the sound source information of the person generated when the person performs voice interaction with the display device at the location where the person is located.
  • the source information refers to the sound information generated when the character interacts with the display device through voice.
  • the sound source information of the person can determine the azimuth and angle of the person's position when speaking, and in order to accurately determine the angle that the camera needs to adjust, it is necessary to first obtain the current state of the camera, that is, the current shooting angle.
  • the current shooting angle of the camera needs to be acquired when the camera is in a stopped state, so as to ensure the accuracy of the current shooting angle of the camera, and thus to ensure the accuracy of determining the angle that the camera needs to adjust.
  • the controller before executing the acquisition of the current shooting angle of the camera, the controller is further configured to execute the following steps:
  • Step 111 query the current operating state of the camera.
  • Step 112 if the current operating state of the camera is in the rotating state, wait for the camera to rotate completely.
  • Step 113 If the current operating state of the camera is in the non-rotation state, obtain the current shooting angle of the camera.
  • a motor control service is configured in the controller, and the motor control service is used to drive the camera to rotate, obtain the running status of the camera and the orientation angle of the camera.
  • the motor control service monitors the running status of the camera in real time.
  • the controller queries the current running status of the camera by calling the motor control service.
  • the current running status of the camera can represent the current orientation angle of the camera and whether the camera is in a rotating state.
  • the current shooting angle of the camera cannot be obtained at this time, otherwise the exact value cannot be determined. Therefore, when the camera is in the rotating state, it is necessary to wait for the camera to execute the previous instruction to complete the rotation, and then perform the step of obtaining the current shooting angle of the camera in the stopped state.
  • the steps of obtaining the current shooting angle of the camera can be performed.
  • S12 Perform sound source identification on the person's sound source information, and determine the sound source angle information, where the sound source angle information is used to represent the azimuth angle of the person's position when speaking.
  • the controller After obtaining the character sound source information generated by the interaction between the character and the display device, the controller needs to perform sound source identification on the character sound source information to determine the position of the character when speaking, specifically the azimuth angle, that is, the character is located in The left and right sides of the sound collector are still facing the sound collector, and the shooting angle of the camera is adjusted according to the position of the character.
  • the character's voice may be in a dialogue with the opposite end user, while the character is still in the shooting area of the camera. If the controller executes the function of adjusting the shooting angle of the camera step, an invalid operation occurs.
  • the wake-up text for triggering the adjustment of the shooting angle of the camera may be stored in the controller in advance, for example, customizing "Hisense Small Gathering” as the wake-up text for sound source recognition.
  • the character uses the voice "Hisense Xiaoju” as the identification sound source to trigger the process of adjusting the camera's shooting angle.
  • the wake-up text can also be customized as other words, which are not specifically limited in this embodiment.
  • FIG. 11 exemplarily shows a flowchart of a wake-up text comparison method according to some embodiments. Specifically, referring to Figure 11, the controller is further configured to perform the following steps before performing sound source identification on the character sound source information and determining the sound source angle information:
  • the preset wake-up text refers to the text used to trigger the sound source identification process.
  • the controller after acquiring the sound source information of the character, the controller first performs text extraction, and extracts the voice interaction text when the character interacts with the display device through voice. Compare the extracted voice interaction text with the preset wake-up text. If the comparison is inconsistent, for example, the character's voice is not "Hisense Xiaoju", but other interactive content. At this time, it means that the current character's voice does not trigger the adjustment of the camera's shooting angle. voice, the controller does not need to perform the relevant steps to adjust the camera's shooting angle.
  • the controller can continue to perform the subsequent steps to adjust the camera's shooting angle.
  • the controller When judging that the person's sound source information is a wake-up voice, that is, a trigger voice for adjusting the shooting angle of the camera, the controller needs to perform a subsequent sound source recognition process.
  • the multiple groups of sound collectors can collect multiple groups of person sound source information when the same person's voice is spoken, so when the controller obtains the person sound source information collected by the sound collector, it can obtain Each sound collector collects the character sound source information generated when the character speaks, that is, the controller will acquire multiple sets of character sound source information.
  • FIG. 12 exemplarily shows a flow chart of a method for sound source identification for character sound source information according to some embodiments.
  • the controller is further configured to perform the following steps when performing sound source identification on the character sound source information and determining the sound source angle information:
  • S122 based on the time difference of speech, calculate the sound source angle information of the position where the character is at the time of speech.
  • each sound collector The frequency response of each sound collector is the same, and its sampling clock is also synchronized. However, because the distance between each sound collector and the character is not the same, the time when each sound collector can collect speech is not the same. There will be a difference in acquisition time between multiple groups of sound collectors.
  • the angle and distance of the sound source from the array can be calculated by the sound collector array, so as to realize the tracking of the sound source at the position of the character when speaking.
  • the time difference between the arrival of the signal between the two microphones is estimated, so as to obtain the equation set of the sound source position coordinates, and then the exact position of the sound source can be obtained by solving the equation set Coordinates, that is, sound source angle information.
  • step S121 the controller performs sound source identification for each of the character sound source information, and calculates the voices generated by the plurality of groups of the sound collectors when collecting the corresponding character sound source information.
  • the time difference is further configured to perform the following steps:
  • Step 1211 Extract the ambient noise, the sound source signal of the person's voice, and the propagation time of the person's voice to each sound collector from the person's sound source information.
  • Step 1212 Determine the received signal of each sound collector according to the environmental noise, the sound source signal and the propagation time.
  • Step 1213 using the cross-correlation time delay estimation algorithm to process the received signal of each sound collector to obtain the speech time difference generated when every two sound collectors collect the corresponding character sound source information.
  • the sound source array can be used to estimate the direction-of-arrival (DOA) estimation. time difference.
  • DOA direction-of-arrival
  • the target signal received by each element of the sound collector array comes from the same sound source. Therefore, there is a strong correlation between the signals of each channel.
  • the time delay between the signals observed by each two sound collectors that is, the speech time difference, can be determined.
  • the character sound source information generated by the character during the speech includes the ambient noise and the sound source signal of the character voice, and the propagation time of the character's voice transmitted to each sound collector can also be extracted from the character sound source information by identifying and extracting, and calculating The received signal of each sound collector.
  • x i (t) is the received signal of the i-th sound collector
  • s(t) is the sound source signal when the character's voice is spoken
  • ⁇ i is the propagation time of the character's voice propagating to the i-th sound collector
  • n i (t) is the environmental noise
  • ⁇ i is the correction coefficient
  • the cross-correlation delay estimation algorithm is used to process the received signal of each sound collector to estimate the delay, which is expressed as: In the formula, is the time delay between the i-th sound collector and the i+1-th sound collector, that is, the voice time difference.
  • the maximum value is the time delay of the two sound collectors, that is, the voice time difference.
  • the peak value of the cross-power spectrum can be weighted in the frequency domain according to the prior knowledge of signal and noise, so as to suppress noise and reverberation interference.
  • PHAT weighting is used to make the interaction rate spectrum between the signals smoother, and the final speech time difference generated by each two sound collectors when collecting the corresponding character sound source information is obtained.
  • the cross-power spectrum weighted by PHAT is similar to the expression of the unit impulse response, which highlights the peak value of the delay, which can effectively suppress the reverberation noise and improve the accuracy and accuracy of the delay (speech time difference) estimation.
  • step S122 the controller is further configured to perform the following steps when calculating the sound source angle information of the position of the character when speaking based on the speech time difference:
  • Step 1221 Acquire the speed of sound in the current environmental state, the coordinates of each sound collector, and the set number of sound collectors.
  • Step 1222 Determine the number of combined pairs of sound collectors according to the set number of sound collectors, where the number of combined pairs refers to the number of combinations obtained by combining two sound collectors.
  • Step 1223 according to the speech time difference, the sound speed and the coordinates of each sound collector corresponding to each two sound collectors, establish a vector relational equation set, the number of which is the same as the number of combination pairs.
  • Step 1224 Solve the vector relation equation system to obtain the vector value of the unit plane wave propagation vector of the sound source at the position of the person's speech.
  • Step 1225 Calculate, according to the vector value, the sound source angle information of the position where the character is speaking.
  • the sound source angle information of the position of the character when speaking can be calculated according to each voice time difference.
  • the number of equations can be set to be the same as the number of combinations obtained by combining the sound collectors in pairs. To this end, the set number N of the sound collectors is obtained, and there are N(N-1)/2 pairs of combinations between all the sound collectors.
  • the sound source angle information can be determined by solving the vector value of the sound source unit plane wave propagation vector at the character's voice position.
  • the sound source angle information of the azimuth angle of the position of the character when speaking is, the sound source angle information of the azimuth angle of the position of the character when speaking.
  • the controller determines the sound source angle information used to represent the azimuth angle of the person's position when speaking by performing sound source identification on the sound source information of the person.
  • the sound source angle information can identify the current position of the character, the current shooting angle of the camera can identify the current position of the camera, and the target rotation angle that the camera needs to rotate in the horizontal direction can be determined according to the difference angle between the two positions. , and the target rotation direction when the camera is rotated.
  • FIG. 13 exemplarily shows a flowchart of a method for determining a target rotation direction and a target rotation angle of a camera according to some embodiments.
  • the controller is further configured to perform the following steps when determining the target rotation direction and target rotation angle of the camera based on the current shooting angle and sound source angle information of the camera:
  • the sound source angle information represents the azimuth angle of the character
  • the sound source angle information of the character can be converted into the camera.
  • the coordinate angle that is, the coordinate angle of the camera is used to replace the sound source angle information of the character.
  • the controller converts the sound source angle information into the coordinate angle of the camera, the controller is further configured to perform the following steps:
  • Step 1311 Acquire the sound source angle range when the character is speaking and the preset angle range when the camera rotates.
  • Step 1312 Calculate the angle difference between the sound source angle range and the preset angle range, and use the half value of the angle difference as the conversion angle.
  • Step 1313 Calculate the angle difference between the angle corresponding to the sound source angle information and the conversion angle, and use the angle difference as the coordinate angle of the camera.
  • the preset angle range is 0° ⁇ 120°
  • the sound source angle range is 0° ⁇ 180°
  • the coordinate angle of the camera cannot directly replace the sound source angle information . Therefore, first calculate the angle difference between the sound source angle range and the preset angle range, then calculate the half value of the angle difference, and use the half value as the conversion angle when the sound source angle information is converted into the coordinate angle of the camera.
  • the angle difference between the sound source angle range and the preset angle range is 60°, the half value of the angle difference is 30°, and 30° is used as the conversion angle. Finally, the angle difference between the angle corresponding to the sound source angle information and the conversion angle is calculated, which is the coordinate angle of the camera converted from the sound source angle information.
  • the angle corresponding to the sound source angle information determined by the controller by acquiring the character sound source information collected by multiple sound collectors is 50°, and the conversion angle is 30°. Therefore, The calculated angle difference is 20°, that is, the 50° corresponding to the sound source angle information is replaced by the camera's coordinate angle of 20° to represent it.
  • the angle corresponding to the sound source angle information determined by the controller by acquiring the character sound source information collected by multiple sound collectors is 130°, and the conversion angle is 30°. Therefore, the calculated angle The difference is 100°, that is, the 130° corresponding to the sound source angle information is replaced by the camera's coordinate angle of 100° to represent it.
  • S132 Calculate the angle difference between the coordinate angle of the camera and the current shooting angle of the camera, and use the angle difference as the target rotation angle of the camera.
  • the coordinate angle of the camera is used to identify the angle of the person's position within the camera coordinates. Therefore, according to the angle difference between the current shooting angle of the camera and the coordinate angle of the camera, the target rotation angle that the camera needs to rotate can be determined.
  • the current shooting angle of the camera is 100° and the coordinate angle of the camera is 20°, it means that the current shooting area of the camera is not aimed at the position of the person, and the difference between the two is 80°. Therefore, it is necessary to rotate the camera 80° after , the shooting area of the camera can be aimed at the position of the person, that is, the target rotation angle of the camera is 80°.
  • S133 Determine the target rotation direction of the camera according to the angle difference.
  • the left side is taken as the 0° position of the camera
  • the right side is taken as the 120° position of the camera. Therefore, after the angle difference is determined according to the coordinate angle of the camera and the current shooting angle of the camera, if the If the angle is greater than the coordinate angle, it means that the camera's shooting angle is on the right side of the character's position, and the angle difference is a negative value; if the current shooting angle is less than the coordinate angle, it means that the camera's shooting angle is on the left side of the character's position. side, the angle difference is a positive value at this time.
  • the target rotation direction of the camera may be determined according to the positive or negative of the angle difference. If the angle difference is a positive value, it means that the shooting angle of the camera is on the left side of the position of the character. At this time, in order to make the image of the character captured by the camera, the shooting angle of the camera needs to be adjusted to the right, and the target rotation direction of the camera is determined. to turn right.
  • the angle difference is a negative value, it means that the shooting angle of the camera is located on the right side of the person's position. At this time, in order to make the camera capture the image of the person, it is necessary to adjust the shooting angle of the camera to the left, and then determine the target rotation direction of the camera. to turn left.
  • FIG. 14 exemplarily shows a scene graph for adjusting the shooting angle of the camera according to some embodiments.
  • the angle corresponding to the sound source angle information corresponding to the character is 50°
  • the converted coordinate angle of the camera is 20°
  • the current shooting angle of the camera is 100°, that is, the center line of the camera's viewing angle is located at the position of the character.
  • the calculated angle difference is -80°.
  • the visible angle difference is a negative value.
  • the camera needs to be adjusted to rotate 80° to the left.
  • FIG. 15 exemplarily shows another scene diagram for adjusting the shooting angle of the camera according to some embodiments.
  • the angle corresponding to the sound source angle information corresponding to the character is 120°
  • the converted coordinate angle of the camera is 90°
  • the current shooting angle of the camera is 40°, that is, the center line of the camera's viewing angle is located at the position of the character.
  • the calculated angle difference is 50°.
  • the visible angle difference is a positive value.
  • the camera needs to be adjusted to rotate 50° to the right.
  • the controller After the controller determines the target rotation direction and target rotation angle required when the camera needs to adjust the shooting angle, it can adjust the shooting angle of the camera according to the target rotation direction and target rotation angle, so that the shooting area of the camera is facing the position of the character. , so that the camera can capture images including characters, so that the shooting angle of the camera can be adjusted according to the position of the characters.
  • FIG. 16 exemplarily shows a scene graph of the position of the character when speaking according to some embodiments. Since the preset angle range of the camera is different from the sound source angle range of the human voice, if it is reflected in the angle diagram, see Figure 16, there is a 30° position between the 0° position of the preset angle range and the 0° position of the sound source angle range ° angle difference, similarly, there is also a 30° angle difference between the 120° position of the preset angle range and the 180° position of the sound source angle range.
  • the controller converts the sound source angle information into the coordinate angle of the camera in the aforementioned step S131, the coordinate angle of the camera converted from the sound source angle information of the character will be negative, or larger than the camera.
  • the maximum value of the preset angle range that is, the coordinate angle of the camera obtained by conversion is not within the preset angle range of the camera.
  • the calculated coordinate angle of the camera is -10°. If the sound source angle information corresponding to the position of the person (b) is 170°, and the conversion angle is 30°, the calculated coordinate angle of the camera is 140°. It can be seen that the coordinate angles of the camera respectively converted according to the position of the person (a) and the position of the person (b) are beyond the preset angle range of the camera.
  • the viewing angle range of the camera is between 60° and 75°, it means that when the camera is rotated to the 0° position or the 120° position, the viewing angle range of the camera can cover the 0° position of the preset angle range and the sound source.
  • the position of the character is within a 30° angle difference between the 0° position of the preset angle range and the 0° position of the sound source angle range, or, if the character is located at the 120° position of the preset angle range and the sound source If there is a 30° angle difference between the 180° positions of the source angle range, in order to capture images including people, adjust the camera’s shooting angle according to the position corresponding to the minimum or maximum value of the camera’s preset angle range .
  • the controller is further configured to perform the following steps: when the angle information of the sound source of the character is converted into the coordinate angle of the camera beyond the preset angle range of the camera, according to the current shooting angle of the camera and the preset angle range The angle difference between the minimum or maximum value determines the target rotation direction and target rotation angle of the camera.
  • the person (a) is located within a 30° angle difference between the 0° position of the preset angle range and the 0° position of the sound source angle range, that is, the sound source corresponding to the sound source angle information of the person (a)
  • the current shooting angle of the camera is 50°.
  • the angle difference is -50°
  • the target rotation direction of the camera is determined to be leftward
  • the target rotation angle is 50° .
  • the center line (a) of the viewing angle of the camera coincides with the 0° line of the camera.
  • the sound source angle corresponding to the sound source angle information of the person (b) is 170°
  • the current shooting angle of the camera is 50°.
  • the angle difference is 70°
  • the target rotation direction of the camera is determined to be rightward
  • the target rotation angle is 70°.
  • the center line (b) of the viewing angle of the camera coincides with the 120° line of the camera.
  • the display device provided by the embodiment of the present application can still rotate the camera to the preset angle range according to the position of the character.
  • the position of the minimum or maximum value of depending on the viewing angle coverage of the camera, an image containing a person is captured.
  • the camera can be rotated within a preset angle range
  • the controller is configured to obtain the sound source information of the characters collected by the sound collector and identify the sound source, and determine the sound source used for identification.
  • the sound source angle information of the azimuth angle of the character's position based on the current shooting angle and sound source angle information of the camera, determine the target rotation direction and target rotation angle of the camera; according to the target rotation direction and target rotation angle, adjust the camera's shooting angle, so that the shooting area of the camera is facing the position of the person's voice.
  • the display device provided by the present application can trigger the rotation of the camera by using the sound source information of the person, and can automatically identify the real-time position of the user and adjust the shooting angle of the camera, so that the camera can always capture images containing the portrait.
  • the display device provided by the foregoing embodiment when adjusting the shooting angle of the camera, is adjusted in the horizontal direction based on the sound source information when the character interacts with the display device, so that the portrait of the character can appear in the shooting area of the camera, and further. Images including portraits were captured.
  • the display device After adjusting the shooting angle of the camera, when shooting a portrait of a person, the center line of its angle of view may not be aligned with the person, which will make the image captured by the camera not located in the center of the image, and the portrait will deviate, affecting the visual effect. Therefore, after adjusting the shooting angle of the camera to capture the portrait, the display device can also locate the position of the portrait through automatic focusing, so as to display the portrait in the central area of the image.
  • the character may be in a standing posture or a sitting posture when interacting with the display device, there is a different height gap between the face of the character and the camera. Therefore, after adjusting the shooting angle of the camera using the sound source information of the person, the shooting area of the camera may be located above or below the head of the person, which will cause the camera to fail to completely capture the person's portrait.
  • FIG. 17 exemplarily shows another scene diagram in which the camera rotates within a preset angle range according to some embodiments.
  • the camera can be rotated horizontally as well as vertically. Therefore, the preset angle range of the camera includes 0 to 120° in the horizontal direction and 0 to 105° in the vertical direction.
  • 17 exemplarily shows the rotation angles of the camera in the vertical direction: pitch 0°, pitch 90°, and pitch 105°; the rotation angles of the camera in the horizontal direction: horizontal 0°, horizontal 60°, horizontal 120°.
  • the display device provided by the embodiment of the present application after adjusting the shooting angle of the camera based on the sound source information provided by the above-mentioned embodiment to include the portrait of the person, also accurately recognizes the position information of the person through the detection of the camera image, so as to calculate The difference between the portrait of the person and the image center of the camera is used to fine-tune the shooting angle of the camera from the horizontal and vertical directions again, so that the portrait of the person is in the center of the image captured by the camera, so as to ensure that the person in the display image is centered.
  • FIG. 18 exemplarily shows a flowchart of a camera control method according to some embodiments
  • FIG. 19 exemplarily shows an overall data flow diagram of a camera control method according to some embodiments.
  • an image detection method is used, by identifying the portrait of the person being photographed in the image, automatically focusing and positioning, and adjusting the shooting angle of the camera to display the portrait in the image. Central location.
  • the controller performs fine-tuning of the camera, it acquires the specified image captured by the camera in real time, and the specified image includes the portrait of the person in the shooting area of the camera.
  • the controller needs to obtain the shooting parameters of the camera after adjusting the shooting angle and the designated image of the captured person located in the shooting area of the camera.
  • the shooting parameters of the camera include the horizontal viewing angle of the camera, the horizontal width of the image, the vertical viewing angle of the camera, and the vertical height of the image.
  • the horizontal viewing angle of the camera means that the preset angle range of the camera is 0 to 120° in the horizontal direction
  • the vertical viewing angle of the camera means that the preset angle range of the camera is 0 to 105° in the vertical direction.
  • the image horizontal width and image vertical height are related to the resolution of the camera. If the camera supports 1080P image preview, the image horizontal width is 1920 pixels and the image vertical height is 1080 pixels.
  • the controller In order to perform positioning and focusing display based on the portrait, the controller recognizes the designated image collected by the camera, identifies the portrait in the image, and obtains the position of the head area as the position of the portrait area, so that the portrait can be accurately displayed on the designated image. the central area.
  • the designated image captured by the camera can be simultaneously displayed on the monitor for preview, and the position of the portrait area can be displayed in the designated image in the form of a face frame, and then the face frame is also displayed in the designated image displayed on the monitor.
  • a face frame is a rectangular or square frame that encloses the head and/or a few body parts of a portrait.
  • the specified image may include portraits of multiple people, so when determining the position of the portrait area, the portraits of multiple people need to be considered at the same time.
  • controller is further configured as:
  • Step 221 Perform identification processing on the designated image to obtain the position information of the head region corresponding to at least one person.
  • Step 222 Calculate the total area information of the head area position information corresponding to at least one character, and use the position corresponding to the total area information as the portrait area position corresponding to the character, and the portrait area position refers to the total area including the head image of at least one character.
  • the position information of the head area refers to the position information of the area framed by the face frame, which can exist in the form of coordinates. There is a corresponding face frame on the portrait of each person, and the portrait and the face frame are in a one-to-one relationship.
  • the total face frame refers to the total area framed by multiple face frames. The smallest rectangular area formed.
  • the position of the portrait region corresponding to the total face frame includes multiple head images of the people.
  • the position of the portrait area can also be used as the top boundary point of the total face frame by the position of the person's head at the top of the specified image, and the bottom boundary point of the total face frame by the position of the head of the person at the bottom of the specified image.
  • the position of the character's head located at the far left in the specified image is used as the left boundary point of the total face frame, and the position of the character's head at the far right in the specified image is used as the right boundary point of the total face frame.
  • Four boundary points are used to make parallel lines parallel to the corresponding sides of the display, and the four parallel lines are perpendicular to each other. After intersecting, a rectangular total face frame can be obtained.
  • the portrait may be offset in the horizontal direction and may also be cheap in the vertical direction. Therefore, in order to fine-tune the camera accurately, it can be adjusted in the horizontal and vertical directions, and the azimuth distance Including horizontal distance and vertical distance.
  • FIG. 20 exemplarily shows a flowchart of a method for calculating azimuth distance according to some embodiments
  • FIG. 21 exemplarily shows a schematic diagram of calculating azimuth distance according to some embodiments.
  • the controller is further configured as:
  • the calculation when calculating the azimuth distance, may be performed according to the coordinate position of the position of the portrait area and the coordinate position of the specified image.
  • the controller recognizes and detects the specified image, it can obtain the coordinate information of each vertex of the position of the portrait area, that is, the pixel coordinate values of the upper left vertex, the upper right vertex, the lower left vertex, and the lower right vertex.
  • the image center P 0 of the specified image is the center point of the image captured by the camera, that is, the center point of the display. Since the specified image is captured by the camera, the size of the specified image is the same as the resolution of the camera, that is, if the resolution of the camera is constant, the width and height of the image captured by the camera are also constant. Determine the image center coordinate information of the specified image according to the resolution of the camera.
  • the horizontal width of the image is 1920 pixels
  • the vertical height of the image is 1080 pixels.
  • the horizontal coordinate of the image center of the specified image is 960 pixels
  • the vertical coordinate is 540 pixels, that is, the image center P 0 coordinate information (x 0 , y 0 ) of the specified image is (960,540).
  • Coordinate position information area may be determined portrait portrait area location coordinate values of four vertices of the pixel, based on this calculated to obtain the center position of the area portrait region P and the horizontal coordinate of the coordinates 1 is a central vertical region.
  • the coordinate information of the position of the portrait area is: upper left vertex A (200, 100), upper right vertex B (500, 100), lower left vertex C (200, 400), and lower right vertex D (500, 400), then calculate the area center P of the position of the portrait area 1
  • the coordinates (x 1 , y 1 ) are (350, 250).
  • S233 Calculate the difference between the horizontal coordinate of the area center of the portrait area position and the image horizontal coordinate of the designated image, and obtain the horizontal distance between the area center of the portrait area position and the image center of the designated image.
  • the horizontal and vertical distances need to be calculated separately. Therefore, when calculating the distance D in the horizontal direction, it is determined by the difference between the horizontal coordinate x 1 of the area center of the portrait area and the horizontal coordinate x 0 of the image of the specified image; when calculating the distance H in the vertical direction, it is determined by the area of the portrait area position. The difference between the center vertical coordinate y 1 and the image vertical coordinate y 0 of the specified image is determined.
  • both the horizontal distance and the vertical distance are expressed in pixel coordinate values.
  • a threshold value can be set by a preset orientation to determine whether there is a distance difference between the area center of the portrait area and the image center of the designated image, thereby determining whether the portrait is located in the center area of the designated image.
  • the threshold is set, it means that the portrait is not located in the center area of the specified image, the camera is not focusing and shooting, and the display position of the portrait in the specified image is deviated. Therefore, it is necessary to control the shooting angle of the camera to adjust the portrait to the central area of the designated image.
  • the target adjustment angle of the camera In order to place the portrait captured by the camera in the center area of the specified image, the target adjustment angle of the camera needs to be determined first. Since the camera can be rotated in the horizontal direction and also in the vertical direction, the target adjustment angle of the camera includes the target horizontal adjustment angle and the target vertical adjustment angle.
  • the image horizontal coordinate of the specified image may be the same as the horizontal coordinate of the area center of the portrait area
  • the image vertical coordinate of the specified image may not be the same as the vertical coordinate of the area center of the portrait area, that is, the shooting angle of the camera may be positive in the horizontal direction.
  • the shooting angle of the camera may be positive in the horizontal direction.
  • the image vertical coordinate of the specified image may be the same as the vertical coordinate of the area center of the portrait area, and the image horizontal coordinate of the specified image may be different from the area center horizontal coordinate of the portrait area, that is, the shooting angle of the camera may be in The vertical direction is facing the person, and there is a deviation in the horizontal direction. At this time, there is no need to control the camera to adjust the shooting angle in the vertical direction, just adjust the shooting angle of the camera in the horizontal direction.
  • the orientation setting threshold used for determination includes the horizontal setting threshold. Threshold and Vertical Threshold.
  • the shooting parameters of the camera include the viewing angle of the camera and the width of the image.
  • the shooting parameters of the camera include the horizontal viewing angle of the camera, the horizontal width of the image, the vertical viewing angle of the camera, and the vertical height of the image.
  • the horizontal viewing angle of the camera ranges from 0 to 120°
  • the vertical viewing angle of the camera ranges from 0 to 105°. If the camera supports 1080P image preview, the horizontal width of the image is 1920 pixels, and the vertical height of the image is 1080 pixels.
  • the azimuth setting threshold is the horizontal setting threshold
  • the azimuth distance between the area center of the portrait area position and the image center of the designated image is the horizontal direction distance
  • the shooting parameters of the camera include the horizontal viewing angle of the camera and the horizontal width of the image.
  • the controller calculates the target adjustment angle of the camera according to the azimuth distance and the shooting parameters of the camera, and is further configured as: if the horizontal distance is greater than the horizontal setting threshold, according to The distance in the horizontal direction, the horizontal viewing angle of the camera, and the horizontal width of the image are used to calculate the target horizontal adjustment angle of the camera.
  • the horizontal distance between the area center of the portrait area position and the image center of the specified image is greater than or equal to the horizontal set threshold, it means that there is a deviation between the portrait position of the camera in the horizontal direction and the center position of the specified image, so that the area center of the portrait area position There is a certain distance from the image center of the specified image. Therefore, in order to make the portrait in the center of the specified image, it is necessary to control the shooting angle of the camera to adjust, according to the horizontal distance D between the area center of the portrait area and the image center of the specified image, the horizontal viewing angle of the camera ⁇ and the horizontal width of the image. IW, calculate the target horizontal adjustment angle ⁇ 1 of the camera.
  • FIG. 22 exemplarily shows a schematic diagram of a horizontal viewing angle of a camera according to some embodiments
  • FIG. 23 exemplarily shows a schematic diagram of calculating a target horizontal adjustment angle according to some embodiments.
  • the viewing angle is ⁇ .
  • ⁇ 1 atan(2*Math.abs(x 0 -x 1 )*tan( ⁇ /2)/IW).
  • the area center of the portrait area position is located to the left of the image center of the specified image, ie x 0 >x 1 (the state shown in Figure 21). And if the person is located on the right side facing the display, then the area center of the portrait area location is located to the right of the image center of the specified image, that is, x 0 ⁇ x 1 (the state shown in Figure 23). It can be seen that when calculating the horizontal distance D between the area center of the portrait area and the image center of the specified image, a negative value may occur. Therefore, in order to accurately obtain the target horizontal adjustment angle of the camera, when calculating the horizontal distance D , and calculate the absolute value of the difference between x 0 -x 1.
  • the azimuth setting threshold is the vertical setting threshold
  • the azimuth distance between the area center of the portrait area position and the image center of the designated image is the vertical direction distance
  • the shooting parameters of the camera include the vertical viewing angle of the camera and the vertical width of the image.
  • the controller executes that the azimuth distance exceeds the azimuth setting threshold, it calculates the target adjustment angle of the camera according to the distance and the shooting parameters of the camera, and is further configured as: if the vertical distance is greater than the vertical setting threshold, Then, according to the vertical distance, the vertical viewing angle of the camera and the vertical height of the image, the vertical adjustment angle of the target of the camera is calculated.
  • the vertical distance between the area center of the portrait area position and the image center of the specified image is greater than or equal to the vertical set threshold, it means that there is a deviation between the portrait position of the camera in the vertical direction and the center position of the specified image, so that the area center of the portrait area position There is a certain distance from the image center of the specified image. Therefore, in order to make the portrait in the center of the specified image, it is necessary to control the shooting angle of the camera to adjust, according to the vertical distance H between the area center of the portrait area and the image center of the specified image, the vertical viewing angle of the camera ⁇ and the vertical width of the image. IH, calculate the target vertical adjustment angle ⁇ 2 of the camera.
  • FIG. 24 exemplarily shows a schematic diagram of a vertical viewing angle of a camera according to some embodiments
  • FIG. 25 exemplarily shows a schematic diagram of calculating a vertical adjustment angle of a target according to some embodiments.
  • the image vertical coordinate y 0 540 of the specified image
  • the vertical coordinate of the area center of the portrait area position is y 1
  • the camera is vertical
  • the viewing angle is ⁇ .
  • ⁇ 2 atan(2*Math.abs(y 0 -y 1 )*tan( ⁇ /2)/IH).
  • the area center of the portrait area position is located at the upper side of the image center of the designated image, that is, y 0 >y 1 (the state shown in FIG. 21 ).
  • the area center of the portrait area position is located on the lower side of the image center of the specified image, that is, y 0 ⁇ y 1 (the state shown in FIG. 25 ). It can be seen that when calculating the vertical distance H between the area center of the portrait area and the image center of the specified image, a negative value may occur. Therefore, in order to accurately obtain the target vertical adjustment angle of the camera, when calculating the vertical distance H , and calculate the absolute value of the difference between y 0 -y 1.
  • the azimuth distance in the horizontal direction between the area center of the portrait area position and the image center of the designated image exceeds the horizontal set threshold, and/or, when the azimuth preset in the vertical direction exceeds the vertical set threshold, then according to the azimuth
  • the distance and the shooting parameters of the camera are used to calculate the target horizontal adjustment angle and/or the target vertical adjustment angle of the camera.
  • the controller controls the shooting angle of the camera to adjust according to the target horizontal adjustment angle and/or the target vertical adjustment angle, which can ensure that the portrait captured by the camera is located in the center area of the designated image.
  • the controller can send a control command to the motor control service, and the motor control service responds to the control command to control the camera to adjust the shooting angle.
  • the portrait can be placed in the center area of the specified image captured by the camera.
  • the camera In the direction facing the monitor, if the area center of the portrait area position is horizontally to the left of the image center of the specified image, the camera will be rotated to the right according to the target horizontal adjustment angle; otherwise, it will be rotated to the left. If the area center of the portrait area position is horizontally above the image center of the specified image, rotate the camera downward according to the target vertical adjustment angle; otherwise, rotate it upward.
  • the controller controls the camera to adjust the shooting angle, if the rotation speed is too fast, the image will shake, and it will stop unstable when it rotates to the specified angle. Therefore, in order to obtain a stable image, it is necessary to accurately determine the rotation direction and rotation speed of the camera when adjusting the shooting angle.
  • the controller is further configured to:
  • Step 251 Determine the target rotation speed and target adjustment direction of the camera according to the target adjustment angle of the camera.
  • the rotational speed of the camera is associated with the target adjustment angle.
  • the logic value of the maximum speed and the logic value of the minimum speed are set, so that the camera rotates within the speed range corresponding to the logic value of the maximum speed and the logic value of the minimum speed.
  • the default maximum speed logic value is 100, which is 100°/s
  • the minimum speed logic value is 10, which is 100°/s.
  • the maximum rotational speed logic value is used as the target rotational speed of the camera. If the target adjustment angle of the camera is greater than or equal to the maximum rotational speed logic value of 100, the target rotation speed of the camera is set to 100°/s.
  • the minimum rotational speed logic value is used as the target rotational speed of the camera. If the target adjustment angle of the camera is less than or equal to the minimum rotation speed logic value of 10, the target rotation speed of the camera is set to 10°/s.
  • the value of the target adjustment angle is used as the target rotation speed of the camera. If the target adjustment angle of the camera is between 100 and 10, the actual target adjustment angle is set as the target rotation speed of the camera. For example, if the target adjustment angle of the camera is 30, set the target rotation speed of the camera to 30°/s.
  • the corresponding camera rotation speed is set before the rotation, and then the camera rotation is performed. Therefore, when the adjustment angle is small, the rotation speed is relatively gentle. If the adjustment angle is large, the rotation is realized at a faster speed, so that the camera can adjust the shooting angle in a timely and stable manner, so that the portrait is located in the center area of the specified image. .
  • the target adjustment direction of the camera it can be determined according to the positive or negative value of the azimuth distance between the area center of the portrait area position and the image center of the specified image.
  • the camera captures the center area of the specified image it is necessary to adjust the shooting angle of the camera to the right, then determine that the target adjustment direction of the camera is to rotate to the right.
  • the target adjustment direction of the camera is determined to be downward rotation.
  • Step 252 Adjust the shooting angle of the camera according to the target adjustment angle, the target adjustment direction and the target rotation speed.
  • the camera can be controlled to perform the corresponding rotation to adjust the shooting angle and realize the focus positioning of the character position, so that the portrait captured by the camera is located in the center of the specified image. displayed in the center area of the display.
  • the camera when controlling the camera, based on the solution of roughly adjusting the shooting angle of the camera through the voice source information of the characters in the display device provided in the foregoing embodiment, the camera can be used again.
  • the image is recognized and detected to more accurately adjust the shooting angle of the camera, effectively locate the specific position of the person, and the camera captures images with high portrait detection accuracy.
  • the display device provided by this embodiment comprehensively utilizes sound source localization and camera image analysis, and takes advantage of the sound source localization's strong spatial perception ability. First, the approximate position of the character is confirmed, the camera is driven toward the sound source, and the camera is used at the same time.
  • the advantage of high accuracy of image analysis is to perform person detection on the captured image to determine the specific position, and drive the camera to perform fine-tuning, so as to achieve precise positioning, so that the person captured by the camera can be displayed in the center area of the designated image, and the focused display is realized on the display.
  • the display device provided in this embodiment is suitable for scenes such as video calls and fitness, and it is very effective for quickly and accurately locating the focused person if the person's standing position is not within the shooting area of the default camera.
  • the portrait is displayed in the center area of the display by fine-adjusting the shooting angle of the camera again.
  • the azimuth distance between the area center of the portrait area and the image center of the specified image in the horizontal direction does not exceed the azimuth setting threshold, it means that the display of the portrait in the specified image does not deviate, which is reflected in the display.
  • the portrait can be displayed in the center of the display. In this case, there is no need to fine-tune the shooting angle of the camera.
  • the display device provided by the embodiment of the present application can perform a portrait focus and magnification display on the position of the portrait area.
  • FIG. 26 exemplarily shows a flowchart of a method for focusing and zooming in on a portrait display according to some embodiments. Specifically, referring to FIG. 26 , based on the display device provided by the foregoing embodiment, the controller is further configured as:
  • the controller obtains The specified image for a preset number of frames.
  • the preset number of frames may be 20 frames.
  • the controller will recognize the image of the preset number of frames to determine that the position of the portrait area is smaller than the entire designated image area, and will automatically focus and enlarge the area where the person's head is located to adapt to the distance between the person and the display device. need.
  • the preset ratio can be set to one third. If the size of the portrait area position is less than or equal to one third of the specified image, it means that the portrait area position is displayed too small and needs to be focused and enlarged.
  • the proportional calculation method of the position of the portrait area can be calculated by the pixel area (number of pixels).
  • the zoom is performed by comparing the position of the portrait area with the aspect ratio of the monitor. Specifically, if the size of the position of the portrait area is smaller than or equal to the preset ratio of the specified image, the controller is further configured to:
  • Step 281 If the size of the portrait area position is smaller than or equal to the preset ratio of the specified image, calculate the aspect ratio value of the display and the aspect ratio value of the portrait area position.
  • FIG. 27 exemplarily shows a schematic diagram of zoomed-in portrait display according to some embodiments.
  • FIG. 27 when focusing and zooming in on the position of the portrait region in a specified image, it may be determined according to the ratio of the position of the portrait region to the aspect ratio of the display. Therefore, the aspect ratio of the display and the aspect ratio of the portrait area position need to be calculated separately.
  • the aspect ratio can be calculated according to the pixel coordinate value.
  • the aspect ratio of the display is the ratio of the width and height of the display, and the width and height of the display are the same as the resolution of the camera, that is, if the camera supports 1080P image preview, Then the horizontal width of the image is 1920 pixels, and the vertical height of the image is 1080 pixels, then the width value of the display is 1920 pixels, and the height value is 1080 pixels.
  • the aspect ratio value of the portrait area position refers to the ratio of the width value and the height value of the portrait area position.
  • the portrait area position may include only the position of the head of the character, or, the position of the head of the character and a few body parts.
  • the width value and height value of the position of the portrait area can be determined by imagining the coordinate value, and the specific method can refer to the method for determining the coordinate information of the position of the portrait area in the foregoing embodiment, which will not be repeated here.
  • Step 282 If the aspect ratio value of the display is inconsistent with the aspect ratio value of the portrait area position, adjust the aspect ratio value of the portrait area position, and the adjusted aspect ratio value of the portrait area position is the same as the aspect ratio value of the display.
  • the designated image captured by the camera will include portraits of multiple characters, and the position of the portrait area enclosed by the portraits of multiple characters may be a rectangle or a rectangle.
  • the aspect ratio of the portrait area needs to be the same as the aspect ratio of the display.
  • the aspect ratio value of the display is inconsistent with the aspect ratio value of the portrait area position, as shown in (a) in Figure 27, adjust the aspect ratio value of the portrait area position so that the portrait area position after adjusting the aspect ratio value
  • the aspect ratio value is the same as that of the display, as shown in Figure 27(b).
  • the aspect ratio of the display and the aspect ratio of the portrait area are inconsistent. One is that the aspect ratio of the portrait area is greater than the aspect ratio of the monitor, and the second is that the aspect ratio of the portrait area is smaller than that of the monitor. of the aspect ratio.
  • the aspect ratio value of the portrait area position is greater than the aspect ratio value of the display, the height value of the portrait area position is adjusted, and the aspect ratio value of the original width value of the portrait area position and the adjusted height value is the same as that of the display.
  • the aspect ratio values are the same.
  • the aspect ratio of the portrait area position is greater than the aspect ratio of the monitor, in order to keep the size of the portrait area position with the ratio of the monitor, you should use the area center point of the portrait area position to expand the upper and lower sides, and the height of the portrait area position should be expanded. value increases.
  • the expansion size of the upper and lower sides is (IH*pW/IW-pH)/2, where, IW is the width value of the display, IH is the height value of the display, pW is the width value of the portrait area position, and pH is the height value of the portrait area position.
  • the width value of the portrait area position is adjusted, and the aspect ratio value of the adjusted width value of the portrait area position and the original height value is the same as that of the display.
  • the aspect ratio values are the same.
  • the center point of the portrait area position should be used to expand the left and right sides, and the width of the portrait area position should be expanded. value increases.
  • the expansion size of the left and right sides is (PH*IW/IH-pW)/2, where, IW is the width value of the display, IH is the height value of the display, pW is the width value of the portrait area position, and pH is the height value of the portrait area position.
  • Step 283 Determine the target enlarged area for the position of the portrait area according to the position of the portrait area adjusted by the aspect ratio.
  • the position of the portrait area only includes the position of the head of the character, or includes the position of the head of the character and a few limbs, if the position of the portrait area is directly enlarged and displayed on the display, distortion will occur. Therefore, in order to prevent the image distortion from being seriously distorted due to the excessive enlargement ratio when the position of the small portrait area is enlarged to full-screen display, it is necessary to determine the target enlargement area.
  • the target enlargement area is the area to be displayed in the display, and the target enlargement area includes the position of the portrait area and the surrounding area. In some embodiments, the target enlargement area is about 1.5 times the position of the portrait area.
  • the position of the portrait area adjusted based on the aspect ratio is enlarged by 1.5 times, and the target enlarged area can be obtained, such as the dotted rectangle area shown in (c) in Figure 27. Enlarge the image corresponding to the target zoom area to full screen display without causing image distortion.
  • Step 284 focus and enlarge the portrait corresponding to the target enlargement area, and display it on the display in full screen.
  • the target magnification area is the area of the portrait area adjusted in the same proportion of the monitor, and the center of the area is the center point and the area is enlarged in the same proportion, some areas may exceed the boundary due to being close to the edge of the image. If the image cannot be displayed on the monitor, adjust the target zoom area according to the part beyond the boundary, so that the part beyond the boundary coincides with the edge corresponding to the specified image.
  • the controller is further configured to perform the following steps when performing focusing and zooming in on the portrait corresponding to the target zoom area and displaying it on the display in full screen:
  • Step 2841 Obtain the coordinates of the center point of the target zoom-in area.
  • the target magnified area is the center point of the area center of the portrait area, the area obtained by enlarging according to the magnification ratio, therefore, the center of the target magnified area is the same as the area center of the portrait area, and the coordinates of the center point of the target magnified area It is the area center coordinate of the portrait area position.
  • Step 2842 calculate the first distance between the center point coordinates and any border of the target zoom area, and the second distance between the center point coordinates and any border of the display, and the position of any border of the target zoom area and any border of the display Corresponding.
  • a certain side of the obtained target enlarged area may exceed a certain boundary of the display.
  • the left boundary of the target enlarged area exceeds the left boundary of the display.
  • calculate the first distance L 11 between the coordinates of the center point of the target enlargement area and the left border of the target enlargement area calculate the first distance L 12 between the center point coordinates of the target enlargement area and the upper boundary of the target enlargement area, and calculate the target enlargement area
  • the first distance L 13 between the coordinates of the center point of the target enlargement area and the right border of the target enlargement area is calculated
  • the first distance L 14 between the center point coordinates of the target enlargement area and the lower boundary of the target enlargement area is calculated.
  • Calculate the second distance L 21 between the coordinates of the center point of the target enlargement area and the left border of the display calculate the second distance L 22 between the coordinates of the center point of the target enlargement area and the upper border of the display, and calculate the coordinates of the center point of the target enlargement area and the display.
  • the second distance L 23 of the right border is calculated, and the second distance L 24 between the coordinates of the center point of the target enlargement area and the lower border of the display is calculated.
  • Step 2843 If the distance difference between the second distance and the first distance is less than zero, adjust the position of the target enlarged area according to the distance difference.
  • the judgment is based on the difference between the first distance corresponding to the side of the target enlargement area located on the same side and the second distance corresponding to the side of the display.
  • the entire target enlarged area is moved in the opposite direction of the side of the target enlarged area beyond the display, so that the side of the target enlarged area beyond the display coincides with the side of the display. For example, as shown in FIG. 27(d), if the left border of the target enlargement area exceeds the left border of the display, the entire target enlargement area is moved to the right so that the left border of the target enlargement area coincides with the left border of the display.
  • the entire target enlargement area is shifted to the left, so that the right border coincides with the right border of the display;
  • the upper boundary coincides; if the lower boundary of the target enlargement area exceeds, the entire target enlargement area is shifted upward, so that the lower boundary coincides with the lower boundary of the display.
  • Step 2844 focus and zoom in on the portrait corresponding to the target zoom-in area whose position has been adjusted, and display it on the display in full screen.
  • the display device when the camera captures the portrait of the person, if the portrait is located in the central area of the display, and the person never changes his position, at this time, the display device does not need to control the camera to adjust the shooting angle, and continues to shoot the portrait of the person at the current shooting angle . Accumulate the designated images of the preset number of frames, and there is no change in the position of the person. When the proportion of the portrait area in the designated image in the designated image is small, the portrait area will be displayed in a zoomed-in portrait focus, so as to display the position corresponding to the portrait area. The image is displayed full screen in the monitor.
  • the display device needs to re-determine the area center of the portrait area position. Then you need to control the camera to adjust the shooting angle to ensure that the portrait is always in the center of the specified image and displayed in the center area of the monitor.
  • the display shows the position of the portrait area displayed in focus and magnification. Therefore, in order to ensure the accuracy of judging the camera to adjust the shooting angle based on the image detection and recognition method, the monitor needs to be adjusted. The position of the portrait area that is currently in focus and magnified display is restored to the original state, and then the subsequent steps of calculating the target adjustment angle of the camera are performed.
  • the controller before the controller performs the calculation of the target adjustment angle of the camera, it is further configured to perform the following steps:
  • Step 0241 Determine whether the specified image is subjected to a portrait focus zoom display operation.
  • Step 0242 If the specified image has not been subjected to the portrait focus and zoom-in display operation, execute the step of calculating the target adjustment angle of the camera.
  • Step 0243 If the specified image has been displayed by focusing and zooming in on the portrait, restore the display of the specified image, and perform the step of calculating the target adjustment angle of the camera.
  • the controller performs the portrait focus magnification display operation on the portrait area position in the specified image, a magnification mark will be generated on the current specified image. If the controller detects that there is a magnifying mark on the currently designated image, it can determine that the designated image is subjected to the portrait focus zoom display operation; if no zoom mark is detected, it is determined that the designated image has not been subjected to the portrait focus zoom display operation.
  • the controller determines that the specified image has not been subjected to the portrait focus and magnification display operation, it can directly perform image detection and analysis on the specified image, and continue to perform the subsequent steps of calculating the target adjustment angle of the camera.
  • the zoomed-in specified image will affect the accuracy of image detection and analysis. Therefore, it is necessary to restore the specified image to its original state and cancel the zoomed in portrait display operation. At the time, the specified image in the original state is displayed on the display, and then the subsequent steps of calculating the target adjustment angle of the camera are continued.
  • the controller performs recognition processing on the designated image collected by the camera, obtains the position of the portrait area, and calculates the azimuth distance between the area center of the portrait area position and the image center of the designated image. ; If the azimuth distance exceeds the azimuth setting threshold, calculate the target adjustment angle of the camera according to the azimuth distance and the shooting parameters of the camera; adjust the shooting angle of the camera based on the target adjustment angle of the camera, so that the portrait of the person is located at the specified location captured by the camera. the central area of the image.
  • the display device provided by the embodiment of the present application accurately recognizes the position information of the person through the image detection of the camera, and automatically focuses and locates the position of the portrait, so as to finely adjust the shooting angle of the camera from the horizontal direction and the vertical direction, so that the portrait of the person is captured by the camera.
  • the center of the image thus ensuring that the display image character is centered.
  • FIG. 18 exemplarily shows a flowchart of a control method of a camera according to some embodiments.
  • the present application also provides a method for controlling a camera, the method comprising:
  • the present application further provides a computer storage medium, wherein the computer storage medium can store a program, and when the program is executed, it can include some or all of the steps in each embodiment of the camera control method provided by the present application.
  • the storage medium may be a magnetic disk, an optical disk, a read-only memory (English: read-only memory, abbreviated as: ROM) or a random access memory (English: random access memory, abbreviated as: RAM) and the like.
  • the technology in the embodiments of the present application can be implemented by means of software plus a necessary general hardware platform.
  • the technical solutions in the embodiments of the present application can be embodied in the form of software products in essence or in the parts that make contributions to related technologies, and the computer software products can be stored in storage media, such as ROM/RAM, A magnetic disk, an optical disk, etc., includes several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to execute the methods described in various embodiments or some parts of the embodiments of the present application.
  • a computer device which may be a personal computer, a server, or a network device, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Studio Devices (AREA)

Abstract

Disclosed in the present application are a camera control method and a display device. The method comprises: a controller performing recognition processing on a designated image acquired by a camera to obtain a portrait area position, and calculating the azimuth distance between the area center of the portrait area position and the image center of the designated image; if the azimuth distance exceeds an azimuth set threshold, calculating, according to the azimuth distance and shooting parameters of the camera, a target adjustment angle of the camera; and on the basis of the target adjustment angle of the camera, adjusting the shooting angle of the camera, so that the portrait of a person is located in the center area of the designated image acquired by the camera.

Description

一种摄像头的控制方法及显示设备A camera control method and display device
本申请要求在2020年7月1日提交中国专利局、申请号为202010628749.6申请名称为“一种摄像头的控制方法及显示设备”的优先权;其全部内容通过引用结合在本申请中。This application claims the priority of the application titled "A Camera Control Method and Display Device" filed with the China Patent Office on July 1, 2020, with the application number of 202010628749.6; the entire contents of which are incorporated into this application by reference.
技术领域technical field
本申请涉及电视软件技术领域,尤其涉及一种摄像头的控制方法及显示设备。The present application relates to the technical field of television software, and in particular, to a control method of a camera and a display device.
背景技术Background technique
随着显示设备的快速发展,显示设备功能将越来越丰富,性能也越来越强大。例如,显示设备可实现网络搜索、IP电视、BBTV网视通、视频点播(VOD)、数字音乐、网络新闻、网络视频电话等功能。而在利用显示设备实现网络视频通话功能时,需在显示设备上安装摄像头,实现用户形象的采集。With the rapid development of display devices, the functions of the display devices will become more and more abundant, and the performance of the display devices will become more and more powerful. For example, the display device can implement functions such as network search, IP TV, BBTV, video on demand (VOD), digital music, network news, and network video telephony. When using a display device to implement the network video call function, a camera needs to be installed on the display device to collect user images.
发明内容SUMMARY OF THE INVENTION
本申请实施例提供了一种显示设备,包括:An embodiment of the present application provides a display device, including:
摄像头,所述摄像头被配置为采集人像以及实现在预设角度范围内的转动;a camera, the camera is configured to capture a portrait and realize rotation within a preset angle range;
与所述摄像头连接的控制器,所述控制器被配置为:A controller connected to the camera, the controller being configured to:
获取所述摄像头的拍摄参数和采集的人物位于摄像头拍摄区域内的指定图像;Acquiring the shooting parameters of the camera and the designated images of the collected characters located in the shooting area of the camera;
对所述指定图像进行识别处理,得到所述人物对应的人像区域位置,所述人像区域位置是指包括人物头部图像的区域;Performing identification processing on the designated image to obtain a portrait area position corresponding to the person, where the portrait area position refers to an area including the head image of the person;
计算所述人像区域位置的区域中心与所述指定图像的图像中心的方位距离,所述方位距离用于标识水平方向距离和垂直方向距离;Calculate the azimuth distance between the area center of the portrait area position and the image center of the designated image, and the azimuth distance is used to identify the horizontal direction distance and the vertical direction distance;
如果所述方位距离超过方位设定阈值,则根据所述方位距离和摄像头的拍摄参数,计算摄像头的目标调整角度;If the azimuth distance exceeds the azimuth setting threshold, calculate the target adjustment angle of the camera according to the azimuth distance and the shooting parameters of the camera;
基于所述摄像头的目标调整角度,调整所述摄像头的拍摄角度,以使所述人物的人像位于摄像头采集的指定图像的中心区域。Based on the target adjustment angle of the camera, the shooting angle of the camera is adjusted so that the portrait of the person is located in the center area of the designated image captured by the camera.
附图说明Description of drawings
为了更清楚地说明本申请的技术方案,下面将对实施例中所需要使用的附图作简单地介绍,显而易见地,对于本领域普通技术人员而言,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to illustrate the technical solutions of the present application more clearly, the accompanying drawings that need to be used in the embodiments will be briefly introduced below. Other drawings can also be obtained from these drawings.
图1中示例性示出了根据一些实施例的显示设备与控制装置之间操作场景的示意图;FIG. 1 exemplarily shows a schematic diagram of an operation scene between a display device and a control apparatus according to some embodiments;
图2中示例性示出了根据一些实施例的显示设备200的硬件配置框图;FIG. 2 exemplarily shows a hardware configuration block diagram of a display device 200 according to some embodiments;
图3中示例性示出了根据一些实施例的控制设备100的硬件配置框图;FIG. 3 exemplarily shows a hardware configuration block diagram of the control device 100 according to some embodiments;
图4中示例性示出了根据一些实施例的显示设备200中软件配置示意图;FIG. 4 exemplarily shows a schematic diagram of software configuration in the display device 200 according to some embodiments;
图5中示例性示出了根据一些实施例的显示设备200中应用程序的图标控件界面 显示示意图;FIG. 5 exemplarily shows a schematic diagram of displaying an icon control interface of an application in the display device 200 according to some embodiments;
图6中示例性示出了根据一些实施例的显示设备的结构框图;FIG. 6 exemplarily shows a structural block diagram of a display device according to some embodiments;
图7中示例性示出了根据一些实施例的实现摄像头转动的预设角度范围的示意图;FIG. 7 exemplarily shows a schematic diagram of implementing a preset angle range for camera rotation according to some embodiments;
图8中示例性示出了根据一些实施例的在预设角度范围内摄像头转动的场景图;FIG. 8 exemplarily shows a scene graph of camera rotation within a preset angle range according to some embodiments;
图9中示例性示出了根据一些实施例的声源角度范围的示意图;FIG. 9 exemplarily shows a schematic diagram of a sound source angle range according to some embodiments;
图10中示例性示出了根据一些实施例的摄像头拍摄角度的调整方法的流程图;FIG. 10 exemplarily shows a flowchart of a method for adjusting the shooting angle of a camera according to some embodiments;
图11中示例性示出了根据一些实施例的唤醒文本的对比方法的流程图;FIG. 11 exemplarily shows a flowchart of a wake-up text comparison method according to some embodiments;
图12中示例性示出了根据一些实施例的对人物声源信息进行声源识别的方法流程图;FIG. 12 exemplarily shows a flowchart of a method for performing sound source identification on character sound source information according to some embodiments;
图13中示例性示出了根据一些实施例的确定摄像头的目标转动方向和目标转动角度的方法流程图;FIG. 13 exemplarily shows a flowchart of a method for determining a target rotation direction and a target rotation angle of a camera according to some embodiments;
图14中示例性示出了根据一些实施例的调整摄像头拍摄角度的一种场景图;FIG. 14 exemplarily shows a scene diagram of adjusting the shooting angle of the camera according to some embodiments;
图15中示例性示出了根据一些实施例的调整摄像头拍摄角度的另一种场景图;FIG. 15 exemplarily shows another scene diagram of adjusting the shooting angle of the camera according to some embodiments;
图16中示例性示出了根据一些实施例的人物语音时所处位置的场景图;FIG. 16 exemplarily shows a scene diagram of the position of a character when speaking according to some embodiments;
图17中示例性示出了根据一些实施例的在预设角度范围内摄像头转动的另一场景图;FIG. 17 exemplarily shows another scene diagram in which the camera rotates within a preset angle range according to some embodiments;
图18中示例性示出了根据一些实施例的摄像头的控制方法的流程图;FIG. 18 exemplarily shows a flowchart of a method for controlling a camera according to some embodiments;
图19中示例性示出了根据一些实施例的摄像头的控制方法的整体数据流图;FIG. 19 exemplarily shows an overall data flow diagram of a camera control method according to some embodiments;
图20中示例性示出了根据一些实施例的计算方位距离的方法流程图;FIG. 20 exemplarily shows a flowchart of a method for calculating an azimuth distance according to some embodiments;
图21中示例性示出了根据一些实施例的计算方位距离的示意图;FIG. 21 exemplarily shows a schematic diagram of calculating azimuth distance according to some embodiments;
图22中示例性示出了根据一些实施例的摄像头水平视角角度的示意图;FIG. 22 exemplarily shows a schematic diagram of the horizontal viewing angle of a camera according to some embodiments;
图23中示例性示出了根据一些实施例的计算目标水平调整角度的示意图;Figure 23 exemplarily shows a schematic diagram of calculating a target horizontal adjustment angle according to some embodiments;
图24中示例性示出了根据一些实施例的摄像头垂直视角角度的示意图;FIG. 24 exemplarily shows a schematic diagram of a vertical viewing angle of a camera according to some embodiments;
图25中示例性示出了根据一些实施例的计算目标垂直调整角度的示意图;FIG. 25 exemplarily shows a schematic diagram of calculating a vertical adjustment angle of a target according to some embodiments;
图26中示例性示出了根据一些实施例的人像聚焦放大显示的方法流程图;FIG. 26 exemplarily shows a flowchart of a method for focusing and zooming in on a portrait display according to some embodiments;
图27中示例性示出了根据一些实施例的人像聚焦放大显示的示意图。FIG. 27 exemplarily shows a schematic diagram of zoomed-in portrait display according to some embodiments.
具体实施方式detailed description
为使本申请的目的、实施方式和优点更加清楚,下面将结合本申请示例性实施例中的附图,对本申请示例性实施方式进行清楚、完整地描述,显然,所描述的示例性实施例仅是本申请一部分实施例,而不是全部的实施例。In order to make the objectives, implementations and advantages of the present application clearer, the exemplary embodiments of the present application will be described clearly and completely below with reference to the accompanying drawings in the exemplary embodiments of the present application. Obviously, the exemplary embodiments described It is only a part of the embodiments of the present application, but not all of the embodiments.
基于本申请描述的示例性实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本申请所附权利要求保护的范围。此外,虽然本申请中公开内容按照示范性一个或几个实例来介绍,但应理解,可以就这些公开内容的各个方面也可以单独构成一个完整实施方式。Based on the exemplary embodiments described in this application, all other embodiments obtained by those of ordinary skill in the art without creative efforts shall fall within the protection scope of the appended claims of this application. Furthermore, although the disclosures in this application have been presented in terms of illustrative example or instances, it should be understood that various aspects of this disclosure may also constitute a complete embodiment in isolation.
需要说明的是,本申请中对于术语的简要说明,仅是为了方便理解接下来描述的实施方式,而不是意图限定本申请的实施方式。除非另有说明,这些术语应当按照其普通和通常的含义理解。It should be noted that the brief description of the terms in the present application is only for the convenience of understanding the embodiments described below, rather than intended to limit the embodiments of the present application. Unless otherwise specified, these terms are to be understood according to their ordinary and ordinary meanings.
本申请中说明书和权利要求书及上述附图中的术语“第一”、“第二”、“第三”等 是用于区别类似或同类的对象或实体,而不必然意味着限定特定的顺序或先后次序,除非另外注明(Unless otherwise indicated)。应该理解这样使用的用语在适当情况下可以互换,例如能够根据本申请实施例图示或描述中给出那些以外的顺序实施。The terms "first", "second", "third", etc. in the description and claims of this application and the above drawings are used to distinguish similar or similar objects or entities, and are not necessarily meant to limit specific Order or precedence unless otherwise indicated. It should be understood that the terms so used are interchangeable under appropriate circumstances, eg, can be implemented in an order other than those given in the illustration or description of the embodiments of the present application.
此外,术语“包括”和“具有”以及他们的任何变形,意图在于覆盖但不排他的包含,例如,包含了一系列组件的产品或设备不必限于清楚地列出的那些组件,而是可包括没有清楚地列出的或对于这些产品或设备固有的其它组件。Furthermore, the terms "comprising" and "having" and any variations thereof, are intended to cover but not exclusively include, for example, a product or device incorporating a series of components is not necessarily limited to those explicitly listed, but may include No other components are expressly listed or inherent to these products or devices.
本申请中使用的术语“模块”,是指任何已知或后来开发的硬件、软件、固件、人工智能、模糊逻辑或硬件或/和软件代码的组合,能够执行与该元件相关的功能。The term "module" as used in this application refers to any known or later developed hardware, software, firmware, artificial intelligence, fuzzy logic or combination of hardware or/and software code capable of performing the function associated with that element.
本申请中使用的术语“遥控器”,是指电子设备(如本申请中公开的显示设备)的一个组件,通常可在较短的距离范围内无线控制电子设备。一般使用红外线和/或射频(RF)信号和/或蓝牙与电子设备连接,也可以包括WiFi、无线USB、蓝牙、动作传感器等功能模块。例如:手持式触摸遥控器,是以触摸屏中用户界面取代一般遥控装置中的大部分物理内置硬键。As used in this application, the term "remote control" refers to a component of an electronic device, such as the display device disclosed in this application, that can wirelessly control the electronic device, usually over a short distance. Generally, infrared and/or radio frequency (RF) signals and/or Bluetooth are used to connect with electronic devices, and functional modules such as WiFi, wireless USB, Bluetooth, and motion sensors may also be included. For example, a hand-held touch remote control replaces most of the physical built-in hard keys in a general remote control device with a user interface in a touch screen.
本申请中使用的术语“手势”,是指用户通过一种手型的变化或手部运动等动作,用于表达预期想法、动作、目的/或结果的用户行为。The term "gesture" used in this application refers to a user's behavior that is used by a user to express an expected thought, action, purpose/or result through an action such as a change of hand shape or hand movement.
图1为根据实施例中显示设备与控制装置之间操作场景的示意图。如图1所示,用户可通过智能设备300或控制装置100操作显示设备200。FIG. 1 is a schematic diagram of an operation scenario between a display device and a control apparatus according to an embodiment. As shown in FIG. 1 , a user can operate the display device 200 through the smart device 300 or the control device 100 .
控制装置100可以是遥控器,遥控器和显示设备的通信包括红外协议通信或蓝牙协议通信,及其他短距离通信方式,通过无线或有线方式来控制显示设备200。用户可以通过遥控器上按键、语音输入、控制面板输入等输入用户指令,来控制显示设备200。The control apparatus 100 may be a remote controller, and the communication between the remote controller and the display device includes infrared protocol communication or Bluetooth protocol communication, and other short-distance communication methods, and the display device 200 is controlled wirelessly or wiredly. The user can control the display device 200 by inputting user instructions through keys on the remote control, voice input, control panel input, and the like.
在一些实施例中,也可以使用智能设备300(如移动终端、平板电脑、计算机、笔记本电脑等)以控制显示设备200。例如,使用在智能设备上运行的应用程序控制显示设备200。In some embodiments, a smart device 300 (eg, a mobile terminal, a tablet computer, a computer, a notebook computer, etc.) can also be used to control the display device 200 . For example, the display device 200 is controlled using an application running on the smart device.
在一些实施例中,显示设备200还可以采用除了控制装置100和智能设备300之外的方式进行控制,例如,可以通过显示设备200设备内部配置的获取语音指令的模块直接接收用户的语音指令控制,也可以通过显示设备200设备外部设置的语音控制设备来接收用户的语音指令控制。In some embodiments, the display device 200 can also be controlled in a manner other than the control apparatus 100 and the smart device 300. For example, the module for acquiring voice commands configured inside the display device 200 can directly receive the user's voice command for control. , the user's voice command control can also be received through a voice control device provided outside the display device 200 device.
在一些实施例中,显示设备200还与服务器400进行数据通信。可允许显示设备200通过局域网(LAN)、无线局域网(WLAN)和其他网络进行通信连接。服务器400可以向显示设备200提供各种内容和互动。In some embodiments, the display device 200 is also in data communication with the server 400 . The display device 200 may be allowed to communicate via local area network (LAN), wireless local area network (WLAN), and other networks. The server 400 may provide various contents and interactions to the display device 200 .
图2示例性示出了根据示例性实施例中控制装置100的配置框图。如图3所示,控制装置100包括控制器110、通信接口130、用户输入/输出接口140、存储器190、供电电源180。控制装置100可接收用户的输入操作指令,且将操作指令转换为显示设备200可识别和响应的指令,起用用户与显示设备200之间交互中介作用。FIG. 2 exemplarily shows a configuration block diagram of the control apparatus 100 according to an exemplary embodiment. As shown in FIG. 3 , the control device 100 includes a controller 110 , a communication interface 130 , a user input/output interface 140 , a memory 190 , and a power supply 180 . The control device 100 can receive the user's input operation instruction, and convert the operation instruction into an instruction that the display device 200 can recognize and respond to, and play an intermediary role between the user and the display device 200 .
图2示出了根据示例性实施例中显示设备200的硬件配置框图。FIG. 2 is a block diagram showing a hardware configuration of a display device 200 according to an exemplary embodiment.
显示设备200包括调谐解调器210、通信器220、检测器230、外部装置接口240、控制器250、显示器275、音频输出接口285、存储器260、供电电源290、用户接口265中的至少一种。The display device 200 includes at least one of a tuner and demodulator 210 , a communicator 220 , a detector 230 , an external device interface 240 , a controller 250 , a display 275 , an audio output interface 285 , a memory 260 , a power supply 290 , and a user interface 265 .
显示器275包括用于呈现画面的显示屏组件,以及驱动图像显示的驱动组件,用于接收源自控制器输出的图像信号,进行显示视频内容、图像内容以及菜单操控界面的组件以及用户操控UI界面。The display 275 includes a display screen component for presenting pictures, and a driving component for driving image display, for receiving image signals output from the controller, components for displaying video content, image content, and menu manipulation interfaces, and user manipulation UI interfaces .
显示器275可为液晶显示器、OLED显示器、以及投影显示器,还可以为一种投影装置和投影屏幕。The display 275 may be a liquid crystal display, an OLED display, and a projection display, and may also be a projection device and a projection screen.
通信器220是用于根据各种通信协议类型与外部设备或服务器进行通信的组件。例如:通信器可以包括Wifi模块,蓝牙模块,有线以太网模块等其他网络通信协议芯片或近场通信协议芯片,以及红外接收器中的至少一种。显示设备200可以通过通信器220与外部控制设备100或服务器400建立控制信号和数据信号的发送和接收。The communicator 220 is a component for communicating with external devices or servers according to various communication protocol types. For example, the communicator may include at least one of a Wifi module, a Bluetooth module, a wired Ethernet module and other network communication protocol chips or near field communication protocol chips, and an infrared receiver. The display device 200 may establish transmission and reception of control signals and data signals with the external control device 100 or the server 400 through the communicator 220 .
用户接口,可用于接收控制装置100(如:红外遥控器等)的控制信号。The user interface can be used to receive control signals from the control device 100 (eg, an infrared remote control, etc.).
检测器230用于采集外部环境或与外部交互的信号。例如,检测器230包括光接收器,用于采集环境光线强度的传感器;或者,检测器230包括图像采集器,如摄像头,可以用于采集外部环境场景、用户的属性或用户交互手势,再或者,检测器230包括声音采集器,如麦克风等,用于接收外部声音。The detector 230 is used to collect external environment or external interaction signals. For example, the detector 230 includes a light receiver, a sensor for collecting ambient light intensity; alternatively, the detector 230 includes an image collector, such as a camera, which can be used to collect external environmental scenes, user attributes or user interaction gestures, or , the detector 230 includes a sound collector, such as a microphone, for receiving external sound.
外部装置接口240可以包括但不限于如下:高清多媒体接口(HDMI)、模拟或数据高清分量输入接口(分量)、复合视频输入接口(CVBS)、USB输入接口(USB)、RGB端口等任一个或多个接口。也可以是上述多个接口形成的复合性的输入/输出接口。The external device interface 240 may include, but is not limited to, the following: any one of high-definition multimedia interface (HDMI), analog or data high-definition component input interface (component), composite video input interface (CVBS), USB input interface (USB), RGB port, etc. or multiple interfaces. It may also be a composite input/output interface formed by a plurality of the above-mentioned interfaces.
控制器250和调谐解调器210可以位于不同的分体设备中,即调谐解调器210也可在控制器250所在的主体设备的外置设备中,如外置机顶盒等。The controller 250 and the tuner 210 may be located in different separate devices, that is, the tuner 210 may also be located in an external device of the main device where the controller 250 is located, such as an external set-top box.
控制器250,通过存储在存储器260上中各种软件控制程序,来控制显示设备的工作和响应用户的操作。控制器250控制显示设备200的整体操作。例如:响应于接收到用于选择在显示器275上显示UI对象的用户命令,控制器250便可以执行与由用户命令选择的对象有关的操作。The controller 250 controls the operation of the display device and responds to user operations through various software control programs stored in the memory 260 . The controller 250 controls the overall operation of the display apparatus 200 . For example, in response to receiving a user command for selecting a UI object to be displayed on the display 275, the controller 250 may perform an operation related to the object selected by the user command.
对象可以是可选对象中的任何一个,例如超链接、图标或其他可操作的控件。与所选择的对象有关操作有:显示连接到超链接页面、文档、图像等操作,或者执行与所述图标相对应程序的操作。Objects can be any of the optional objects, such as hyperlinks, icons, or other actionable controls. The operations related to the selected object include: displaying operations connected to hyperlinked pages, documents, images, etc., or executing operations of programs corresponding to the icons.
在一些实施例中,用户可在显示器275上显示的图形用户界面(GUI)输入用户命令,则用户输入接口通过图形用户界面(GUI)接收用户输入命令。或者,用户可通过输入特定的声音或手势进行输入用户命令,则用户输入接口通过传感器识别出声音或手势,来接收用户输入命令。In some embodiments, the user may input user commands on a graphical user interface (GUI) displayed on the display 275, and the user input interface receives the user input commands through the graphical user interface (GUI). Alternatively, the user may input a user command by inputting a specific sound or gesture, and the user input interface recognizes the sound or gesture through a sensor to receive the user input command.
“用户界面”可以指应用程序或操作系统与用户之间进行交互和信息交换的介质接口,它实现信息的内部形式与用户可以接受形式之间的转换。用户界面常用的表现形式是图形用户界面(Graphic User Interface,GUI),是指采用图形方式显示的与计算机操作相关的用户界面。它可以是在电子设备的显示屏中显示的一个图标、窗口、控件等界面元素,其中控件可以包括图标、按钮、菜单、选项卡、文本框、对话框、状态栏、导航栏、Widget等可视的界面元素。"User interface" can refer to the medium interface for interaction and information exchange between application programs or operating systems and users, which realizes the conversion between the internal form of information and the form acceptable to users. The commonly used form of user interface is Graphical User Interface (GUI), which refers to a user interface related to computer operations displayed in a graphical manner. It can be an icon, window, control and other interface elements displayed on the display screen of the electronic device, wherein the control can include icons, buttons, menus, tabs, text boxes, dialog boxes, status bars, navigation bars, Widgets, etc. visual interface elements.
参见图4,在一些实施例中,将系统分为四层,从上至下分别为应用程序(Applications)层(简称“应用层”),应用程序框架(Application Framework)层(简称“框架层”),安卓运行时(Android runtime)和系统库层(简称“系统运行库层”),以及内 核层。Referring to FIG. 4 , in some embodiments, the system is divided into four layers, from top to bottom, they are an application layer (referred to as “application layer”), an application framework layer (referred to as “framework layer”) ”), the Android runtime and the system library layer (referred to as the “system runtime layer”), and the kernel layer.
在一些实施例中,应用程序层中运行有至少一个应用程序,这些应用程序可以是操作系统自带的窗口(Window)程序、系统设置程序、时钟程序、相机应用等;也可以是第三方开发者所开发的应用程序,比如嗨见程序、K歌程序、魔镜程序等。在具体实施时,应用程序层中的应用程序包不限于以上举例,实际还可以包括其它应用程序包,本申请实施例对此不做限制。In some embodiments, at least one application program runs in the application program layer, and these application programs may be a Window program, a system setting program, a clock program, a camera application, etc. built into the operating system; they may also be developed by a third party The application programs developed by the author, such as the Hijian program, the K song program, the magic mirror program, etc. During specific implementation, the application package in the application layer is not limited to the above examples, and may actually include other application packages, which is not limited in this embodiment of the present application.
框架层为应用程序层的应用程序提供应用编程接口(application programming interface,API)和编程框架。应用程序框架层包括一些预先定义的函数。应用程序框架层相当于一个处理中心,这个中心决定让应用层中的应用程序做出动作。应用程序通过API接口,可在执行中访问系统中的资源和取得系统的服务。The framework layer provides an application programming interface (API) and a programming framework for applications in the application layer. The application framework layer includes some predefined functions. The application framework layer is equivalent to a processing center, which decides to let the applications in the application layer take action. The application program can access the resources in the system and obtain the services of the system during execution through the API interface.
如图4所示,本申请实施例中应用程序框架层包括管理器(Managers),内容提供者(Content Provider)等,其中管理器包括以下模块中的至少一个:活动管理器(Activity Manager)用与和系统中正在运行的所有活动进行交互;位置管理器(Location Manager)用于给系统服务或应用提供了系统位置服务的访问;文件包管理器(Package Manager)用于检索当前安装在设备上的应用程序包相关的各种信息;通知管理器(Notification Manager)用于控制通知消息的显示和清除;窗口管理器(Window Manager)用于管理用户界面上的括图标、窗口、工具栏、壁纸和桌面部件。As shown in FIG. 4 , the application framework layer in the embodiment of the present application includes managers (Managers), content providers (Content Provider), etc., wherein the manager includes at least one of the following modules: an activity manager (Activity Manager) uses Interacts with all activities running in the system; Location Manager is used to provide system services or applications with access to system location services; Package Manager is used to retrieve files currently installed on the device Various information related to the application package; Notification Manager (Notification Manager) is used to control the display and clearing of notification messages; Window Manager (Window Manager) is used to manage icons, windows, toolbars, wallpapers on the user interface and desktop widgets.
在一些实施例中,活动管理器用于:管理各个应用程序的生命周期以及通常的导航回退功能,比如控制应用程序的退出(包括将显示窗口中当前显示的用户界面切换到系统桌面)、打开、后退(包括将显示窗口中当前显示的用户界面切换到当前显示的用户界面的上一级用户界面)等。In some embodiments, the activity manager is used to: manage the life cycle of each application and the usual navigation and fallback functions, such as controlling the exit of the application (including switching the user interface currently displayed in the display window to the system desktop), opening the , back (including switching the currently displayed user interface in the display window to the upper-level user interface of the currently displayed user interface), and the like.
在一些实施例中,窗口管理器用于管理所有的窗口程序,比如获取显示屏大小,判断是否有状态栏,锁定屏幕,截取屏幕,控制显示窗口变化(例如将显示窗口缩小显示、抖动显示、扭曲变形显示等)等。In some embodiments, the window manager is used to manage all window programs, such as obtaining the size of the display screen, judging whether there is a status bar, locking the screen, taking screenshots, and controlling the change of the display window (for example, reducing the display window to display, shaking display, twisting deformation display, etc.), etc.
在一些实施例中,系统运行库层为上层即框架层提供支撑,当框架层被使用时,安卓操作系统会运行系统运行库层中包含的C/C++库以实现框架层要实现的功能。In some embodiments, the system runtime layer provides support for the upper layer, that is, the framework layer. When the framework layer is used, the Android operating system will run the C/C++ library included in the system runtime layer to implement the functions to be implemented by the framework layer.
在一些实施例中,内核层是硬件和软件之间的层。如图4所示,内核层至少包含以下驱动中的至少一种:音频驱动、显示驱动、蓝牙驱动、摄像头驱动、WIFI驱动、USB驱动、HDMI驱动、传感器驱动(如指纹传感器,温度传感器,触摸传感器、压力传感器等)等。In some embodiments, the kernel layer is the layer between hardware and software. As shown in Figure 4, the kernel layer at least includes at least one of the following drivers: audio driver, display driver, Bluetooth driver, camera driver, WIFI driver, USB driver, HDMI driver, sensor driver (such as fingerprint sensor, temperature sensor, touch sensors, pressure sensors, etc.), etc.
在一些实施例中,内核层还包括用于进行电源管理的电源驱动模块。In some embodiments, the kernel layer further includes a power driver module for power management.
在一些实施例中,图4中的软件架构对应的软件程序和/或模块存储在图2或图3所示的第一存储器或第二存储器中。In some embodiments, software programs and/or modules corresponding to the software architecture in FIG. 4 are stored in the first memory or the second memory shown in FIG. 2 or FIG. 3 .
在一些实施例中,以魔镜应用(拍照应用)为例,当遥控接收装置接收到遥控器输入操作,相应的硬件中断被发给内核层。内核层将输入操作加工成原始输入事件(包括输入操作的值,输入操作的时间戳等信息)。原始输入事件被存储在内核层。应用程序框架层从内核层获取原始输入事件,根据焦点当前的位置识别该输入事件所对应的控件以及以该输入操作是确认操作,该确认操作所对应的控件为魔镜应用图标的控件,魔镜应用调用应用框架层的接口,启动魔镜应用,进而通过调用内核层启动摄像头驱 动,实现通过摄像头捕获静态图像或视频。In some embodiments, taking the magic mirror application (photography application) as an example, when the remote control receiving device receives the input operation of the remote control, the corresponding hardware interrupt is sent to the kernel layer. The kernel layer processes the input operation into the original input event (including the value of the input operation, the timestamp of the input operation and other information). Raw input events are stored at the kernel layer. The application framework layer obtains the original input event from the kernel layer, identifies the control corresponding to the input event according to the current position of the focus, and regards the input operation as a confirmation operation, and the control corresponding to the confirmation operation is the control of the magic mirror application icon. The mirror application calls the interface of the application framework layer, starts the mirror application, and then starts the camera driver by calling the kernel layer to capture still images or videos through the camera.
在一些实施例中,对于具备触控功能的显示设备,以分屏操作为例,显示设备接收用户作用于显示屏上的输入操作(如分屏操作),内核层可以根据输入操作产生相应的输入事件,并向应用程序框架层上报该事件。由应用程序框架层的活动管理器设置与该输入操作对应的窗口模式(如多窗口模式)以及窗口位置和大小等。应用程序框架层的窗口管理根据活动管理器的设置绘制窗口,然后将绘制的窗口数据发送给内核层的显示驱动,由显示驱动在显示屏的不同显示区域显示与之对应的应用界面。In some embodiments, for a display device with a touch function, taking a split-screen operation as an example, the display device receives an input operation (such as a split-screen operation) performed by the user on the display screen, and the kernel layer can generate corresponding input operations according to the input operation. Enter an event and report the event to the application framework layer. The window mode (such as multi-window mode) and window position and size corresponding to the input operation are set by the activity manager of the application framework layer. The window management of the application framework layer draws the window according to the settings of the activity manager, and then sends the drawn window data to the display driver of the kernel layer, and the display driver displays the corresponding application interface in different display areas of the display screen.
在一些实施例中,如图5中所示,应用程序层包含至少一个应用程序可以在显示器中显示对应的图标控件,如:直播电视应用程序图标控件、视频点播应用程序图标控件、媒体中心应用程序图标控件、应用程序中心图标控件、游戏应用图标控件等。In some embodiments, as shown in FIG. 5, the application layer contains at least one application that can display corresponding icon controls in the display, such as: live TV application icon control, video on demand application icon control, media center application Program icon controls, application center icon controls, game application icon controls, etc.
在一些实施例中,直播电视应用程序,可以通过不同的信号源提供直播电视。例如,直播电视应用程可以使用来自有线电视、无线广播、卫星服务或其他类型的直播电视服务的输入提供电视信号。以及,直播电视应用程序可在显示设备200上显示直播电视信号的视频。In some embodiments, the live TV application may provide live TV from different sources. For example, a live TV application may provide a TV signal using input from cable, over-the-air, satellite services, or other types of live TV services. And, the live TV application may display the video of the live TV signal on the display device 200 .
在一些实施例中,视频点播应用程序,可以提供来自不同存储源的视频。不同于直播电视应用程序,视频点播提供来自某些存储源的视频显示。例如,视频点播可以来自云存储的服务器端、来自包含已存视频节目的本地硬盘储存器。In some embodiments, a video-on-demand application may provide video from various storage sources. Unlike live TV applications, video-on-demand provides a display of video from certain storage sources. For example, video-on-demand can come from the server side of cloud storage, from local hard disk storage containing existing video programs.
在一些实施例中,媒体中心应用程序,可以提供各种多媒体内容播放的应用程序。例如,媒体中心,可以为不同于直播电视或视频点播,用户可通过媒体中心应用程序访问各种图像或音频所提供服务。In some embodiments, the media center application may provide various multimedia content playback applications. For example, a media center may provide services other than live TV or video-on-demand, where users can access various images or audio through a media center application.
在一些实施例中,应用程序中心,可以提供储存各种应用程序。应用程序可以是一种游戏、应用程序,或某些和计算机系统或其他设备相关但可以在智能电视中运行的其他应用程序。应用程序中心可从不同来源获得这些应用程序,将它们储存在本地储存器中,然后在显示设备200上可运行。In some embodiments, the application center may provide storage of various applications. An application can be a game, an application, or some other application that is related to a computer system or other device but can be run on a Smart TV. The application center can obtain these applications from various sources, store them in local storage, and then run them on the display device 200 .
在一些实施例中,显示设备中需要利用到摄像头的应用程序包括“嗨见”、“照镜子”、“优学猫”、“健身”等,可实现“视频聊天”、“边看边聊”和“健身”等功能。“嗨见”是一款视频聊天应用,可实现手机与电视之间,电视与电视之间的一键聊天。“照镜子”是为用户提供镜子服务的应用,通过照镜子应用打开摄像头,用户可将智能电视作为镜子使用。“优学猫”是提供学习功能的应用。在实现“边聊边看”功能时,用户在启动“嗨见”应用进行视频通话的场景下,同时观看视频节目。“健身”功能可在显示设备的显示器上同步显示健身指导视频和摄像头拍摄的用户跟随健身指导视频做相应动作的图像,实现用户实时查看自身的动作是否标准。In some embodiments, the application programs that need to use the camera in the display device include "Hey see", "Look in the mirror", "Youxuemao", "Fitness", etc., which can realize "video chat", "watch while chatting" ” and “Fitness”. "See you" is a video chat application that can realize one-click chat between mobile phone and TV, and between TV and TV. "Looking in the Mirror" is an application that provides users with mirror services. By turning on the camera through the mirroring application, users can use the smart TV as a mirror. "Youxuemao" is an application that provides learning functions. When the "watch while chatting" function is implemented, the user simultaneously watches the video program under the scenario of launching the "Hi See" application to make a video call. The "fitness" function can simultaneously display the fitness instruction video and the image of the user following the fitness instruction video to perform corresponding actions on the display of the display device, so that users can check whether their actions are standard in real time.
由于用户在利用显示设备进行“视频聊天”、“边看边聊”或“健身”时,可能不会固定在一个位置不动,用户还可以边走边进行上述功能实现。但是现有的显示设备中,摄像头固定安装在显示设备上,摄像头的视角中心线与显示器垂直,且摄像头的可视角度有限,通常位于60°~75°之间,即摄像头的拍摄区域为以摄像头的视角中心线向左和向右同步扩散形成60°~75°角度对应的区域。Since the user may not be fixed in one position when using the display device for "video chat", "watch while chatting" or "exercise", the user can also perform the above functions while walking. However, in the existing display device, the camera is fixedly installed on the display device, the center line of the viewing angle of the camera is perpendicular to the display, and the viewing angle of the camera is limited, usually between 60° and 75°, that is, the shooting area of the camera is The center line of the angle of view of the camera is synchronously spread to the left and right to form an area corresponding to an angle of 60° to 75°.
如果用户走出摄像头的拍摄区域,摄像头将无法拍摄到包含用户人像的图像,使得显示器上无法显示人像。若在视频聊天通话场景下,与本端用户进行视频聊天通话的对端用户将无法看到本端用户;若在健身场景下,显示器上将无法显示用户呈现健 身动作的图像,用户看不到自己的健身动作,将无法判断是否标准,影响用户体验。If the user walks out of the shooting area of the camera, the camera cannot capture an image containing the user's portrait, so that the portrait cannot be displayed on the monitor. If in the video chat call scenario, the opposite end user who is in the video chat call with the local user will not be able to see the local user; if in the fitness scenario, the display will not be able to display the image of the user presenting the fitness action, and the user will not be able to see it. Your own fitness movements will not be able to judge whether they are standard or not, which will affect the user experience.
图6中示例性示出了根据一些实施例的显示设备的结构框图。为了在用户走出摄像头的拍摄区域,使得摄像头仍然可以拍摄到用户的图像,参见图6,本申请实施例提供了一种显示设备,包括摄像头232、声音采集器231和控制器250。摄像头用于采集人像,摄像头不再采用固定安装方式,而是以可转动的方式安装在显示设备上,具体地,摄像头232以转动形式安装在显示器的顶部,摄像头232可沿显示器的顶部转动。FIG. 6 exemplarily shows a structural block diagram of a display device according to some embodiments. In order to allow the camera to still capture the user's image when the user walks out of the camera's shooting area, referring to FIG. The camera is used to capture portraits. The camera is no longer fixedly installed, but is rotatably installed on the display device. Specifically, the camera 232 is installed on the top of the display in a rotating form, and the camera 232 can rotate along the top of the display.
图7中示例性示出了根据一些实施例的实现摄像头转动的预设角度范围的示意图;图8中示例性示出了根据一些实施例的在预设角度范围内摄像头转动的场景图。参见图7和图8,预设摄像头232可在预设角度范围内转动,且在水平方向上转动。在一些实施例中,预设角度范围为0°~120°,即在面对显示器的位置,以用户的左侧为0°,用户的右侧为120°。以摄像头232的视角中心线垂直于显示器时的状态为初始状态,摄像头可实现由初始状态向左转动60°,以及,由初始状态向右转动60°;摄像头的视角中心线与显示器垂直的位置为摄像头60°的位置。FIG. 7 exemplarily shows a schematic diagram of implementing a preset angle range of camera rotation according to some embodiments; FIG. 8 exemplarily shows a scene diagram of camera rotation within the preset angle range according to some embodiments. Referring to FIG. 7 and FIG. 8 , the preset camera 232 can be rotated within a preset angle range and rotated in a horizontal direction. In some embodiments, the preset angle range is 0°˜120°, that is, at the position facing the display, the left side of the user is 0° and the right side of the user is 120°. Taking the state when the center line of the viewing angle of the camera 232 is perpendicular to the display as the initial state, the camera can be rotated 60° to the left from the initial state, and 60° to the right from the initial state; the center line of the viewing angle of the camera is perpendicular to the display. The position is the 60° position of the camera.
本申请实施例提供的显示设备,实现利用声源信息触发摄像头的转动,能够自动识别用户的实时所处位置并调整摄像头的拍摄角度,使得摄像头始终能够拍摄到包含人像的图像。为此,在一些实施例中,显示设备通过设置声音采集器231实现人物声源信息的采集。The display device provided by the embodiment of the present application realizes the use of sound source information to trigger the rotation of the camera, can automatically identify the real-time location of the user and adjust the shooting angle of the camera, so that the camera can always capture images including portraits. To this end, in some embodiments, the display device implements the collection of the sound source information of the person by setting the sound collector 231 .
为保证声源采集的准确性,显示设备中可设置多组声音采集器,在一些实施例中,显示设备中设置四组声音采集器231,四组声音采集器231可以线性的位置关系进行排列设置。在一些实施例中,声音采集器可为麦克风,四组麦克风线性排列形成麦克风阵列。在声音采集时,四组声音采集器231接收同一用户通过语音与显示设备交互时产生的声音信息。In order to ensure the accuracy of sound source collection, multiple sets of sound collectors can be set in the display device. In some embodiments, four sets of sound collectors 231 are set in the display device, and the four sets of sound collectors 231 can be arranged in a linear positional relationship. set up. In some embodiments, the sound collector may be a microphone, and four groups of microphones are linearly arranged to form a microphone array. During sound collection, the four groups of sound collectors 231 receive sound information generated when the same user interacts with the display device through voice.
图9中示例性示出了根据一些实施例的声源角度范围的示意图。用户在语音时,产生的声音会被360°接收到,因此,在用户位于显示设备正面时,用户产生的声源角度范围为0°~180°,同样的,在用户位于显示设备背面,用户产生的声源角度范围也为0°~180°。参见图9,以用户面对显示设备的位置为例,用户位于声音采集器左侧为水平0°,用户位于声音采集器右侧为水平180°。A schematic diagram of a sound source angle range according to some embodiments is exemplarily shown in FIG. 9 . When the user speaks, the generated sound will be received by 360°. Therefore, when the user is on the front of the display device, the angle of the sound source generated by the user ranges from 0° to 180°. Similarly, when the user is on the back of the display device, the user The generated sound source angle also ranges from 0° to 180°. Referring to FIG. 9 , taking the position of the user facing the display device as an example, the user is located on the left side of the sound collector, which is 0° horizontally, and the user is located on the right side of the sound collector, which is 180° horizontally.
再次参见图7和图9,声源的30°角位置等于摄像头的0°角位置,声源的90°角位置等于摄像头的60°角位置,声源的150°角位置等于摄像头的120°角位置。Referring again to Figures 7 and 9, the 30° angular position of the sound source is equal to the 0° angular position of the camera, the 90° angular position of the sound source is equal to the 60° angular position of the camera, and the 150° angular position of the sound source is equal to the 120° angular position of the camera corner position.
控制器250分别与摄像头232和声音采集器231连接,控制器用于接收声音采集器采集到的人物声源信息,并对人物声源信息进行识别,确定出人物所在位置的方位角度,进而确定摄像头需要转动的角度。控制器按照确定出的摄像头需要转动的角度调整摄像头的拍摄角度,使得摄像头的拍摄区域正对人物语音时的所处位置,实现根据人物的位置调整摄像头的拍摄角度以拍摄到包含人物的图像。The controller 250 is connected with the camera 232 and the sound collector 231 respectively, and the controller is used to receive the sound source information of the character collected by the sound collector, identify the sound source information of the character, determine the azimuth angle of the position of the character, and then determine the camera. The angle that needs to be turned. The controller adjusts the shooting angle of the camera according to the determined angle that the camera needs to rotate, so that the shooting area of the camera is facing the position of the voice of the character, and adjusts the shooting angle of the camera according to the position of the character to capture the image containing the character.
图10中示例性示出了根据一些实施例的摄像头拍摄角度的调整方法的流程图。本申请实施例提供的一种显示设备,在根据人物的位置调整摄像头的拍摄角度时,控制器被配置为执行图10所示的摄像头拍摄角度的调整方法,包括:FIG. 10 exemplarily shows a flowchart of a method for adjusting the shooting angle of a camera according to some embodiments. In a display device provided by an embodiment of the present application, when adjusting the shooting angle of the camera according to the position of the character, the controller is configured to execute the method for adjusting the shooting angle of the camera shown in FIG. 10 , including:
S11、获取声音采集器采集的人物声源信息和摄像头的当前拍摄角度。S11. Acquire the character sound source information collected by the sound collector and the current shooting angle of the camera.
在一些实施例中,显示设备中的控制器在驱动摄像头转动,以调整摄像头的拍摄 角度时,需根据人物在所处位置与显示设备进行语音交互时产生的人物声源信息来确定,人物声源信息是指人物通过语音与显示设备交互时产生的声音信息。In some embodiments, when the controller in the display device drives the camera to rotate to adjust the shooting angle of the camera, it needs to determine the sound source information of the person generated when the person performs voice interaction with the display device at the location where the person is located. The source information refers to the sound information generated when the character interacts with the display device through voice.
人物声源信息可确定出人物在语音时所处位置的方位角度,而为准确确定摄像头需要进行调整的角度,需要先获取摄像头的当前状态,即当前拍摄角度。摄像头的当前拍摄角度需要在摄像头处于停止状态时才可被获取,以保证摄像头的当前拍摄角度的准确性,进而保证确定摄像头需要进行调整角度的准确性。The sound source information of the person can determine the azimuth and angle of the person's position when speaking, and in order to accurately determine the angle that the camera needs to adjust, it is necessary to first obtain the current state of the camera, that is, the current shooting angle. The current shooting angle of the camera needs to be acquired when the camera is in a stopped state, so as to ensure the accuracy of the current shooting angle of the camera, and thus to ensure the accuracy of determining the angle that the camera needs to adjust.
因此,控制器在执行获取摄像头的当前拍摄角度之前,被进一步配置为执行下述步骤:Therefore, before executing the acquisition of the current shooting angle of the camera, the controller is further configured to execute the following steps:
步骤111、查询摄像头的当前运行状态。Step 111 , query the current operating state of the camera.
步骤112、如果摄像头的当前运行状态为处于旋转状态,则等待摄像头旋转完毕。 Step 112 , if the current operating state of the camera is in the rotating state, wait for the camera to rotate completely.
步骤113、如果摄像头的当前运行状态为处于未旋转状态,则获取摄像头的当前拍摄角度。Step 113: If the current operating state of the camera is in the non-rotation state, obtain the current shooting angle of the camera.
控制器内配置有马达控制服务,马达控制服务用于驱动摄像头转动、获取摄像头的运行状态和摄像头朝向角度。A motor control service is configured in the controller, and the motor control service is used to drive the camera to rotate, obtain the running status of the camera and the orientation angle of the camera.
马达控制服务实时监控摄像头的运行状态,控制器通过调用马达控制服务查询摄像头的当前运行状态,摄像头的当前运行状态可表征当前摄像头的朝向角度以及摄像头是否处于旋转状态。The motor control service monitors the running status of the camera in real time. The controller queries the current running status of the camera by calling the motor control service. The current running status of the camera can represent the current orientation angle of the camera and whether the camera is in a rotating state.
如果摄像头正处于旋转状态,此时不能获取摄像头的当前拍摄角度,否则无法确定准确的数值。因此,在摄像头处于旋转状态时,需先等待摄像头执行前一指令完成转动后,在停止状态下,再执行获取摄像头的当前拍摄角度的步骤。If the camera is in a rotating state, the current shooting angle of the camera cannot be obtained at this time, otherwise the exact value cannot be determined. Therefore, when the camera is in the rotating state, it is necessary to wait for the camera to execute the previous instruction to complete the rotation, and then perform the step of obtaining the current shooting angle of the camera in the stopped state.
如果摄像头正处于未旋转状态,即摄像头处于停止状态,则可执行获取摄像头的当前拍摄角度的步骤。If the camera is in a non-rotating state, that is, the camera is in a stopped state, the steps of obtaining the current shooting angle of the camera can be performed.
S12、对人物声源信息进行声源识别,确定声源角度信息,声源角度信息用于表征人物在语音时所处位置的方位角度。S12. Perform sound source identification on the person's sound source information, and determine the sound source angle information, where the sound source angle information is used to represent the azimuth angle of the person's position when speaking.
在获取到人物与显示设备交互产生的人物声源信息后,控制器需对人物声源信息进行声源识别,以判断出人物在语音时的所处位置,具体为方位角度,即人物是位于声音采集器的左侧、右侧还是正对声音采集器的位置,进而根据人物的所处位置调整摄像头的拍摄角度。After obtaining the character sound source information generated by the interaction between the character and the display device, the controller needs to perform sound source identification on the character sound source information to determine the position of the character when speaking, specifically the azimuth angle, that is, the character is located in The left and right sides of the sound collector are still facing the sound collector, and the shooting angle of the camera is adjusted according to the position of the character.
由于人物在与显示设备交互时,例如在视频通话场景中,人物语音可能是在与对端用户进行对话,而自身仍位于摄像头的拍摄区域内,若此时控制器执行调整摄像头的拍摄角度的步骤,则会出现无效操作。When the character interacts with the display device, such as in a video call scenario, the character's voice may be in a dialogue with the opposite end user, while the character is still in the shooting area of the camera. If the controller executes the function of adjusting the shooting angle of the camera step, an invalid operation occurs.
因此,为准确的根据人物声源信息确定是否需要对摄像头的拍摄角度进行调整,需要先对人物产生的人物声源信息进行分析,判断人物声源信息是否为触发摄像头调整的信息。Therefore, in order to accurately determine whether the shooting angle of the camera needs to be adjusted according to the sound source information of the person, it is necessary to analyze the sound source information of the person generated by the person to determine whether the sound source information of the person is the information that triggers the adjustment of the camera.
在一些实施例中,可预先在控制器内存储用于触发摄像头拍摄角度调整的唤醒文本,例如,定制“海信小聚”作为声源识别的唤醒文本。人物通过语音“海信小聚”作为识别声源,以触发调整摄像头拍摄角度的过程。唤醒文本也可定制为其他词语,本实施例中不做具体限定。In some embodiments, the wake-up text for triggering the adjustment of the shooting angle of the camera may be stored in the controller in advance, for example, customizing "Hisense Small Gathering" as the wake-up text for sound source recognition. The character uses the voice "Hisense Xiaoju" as the identification sound source to trigger the process of adjusting the camera's shooting angle. The wake-up text can also be customized as other words, which are not specifically limited in this embodiment.
图11中示例性示出了根据一些实施例的唤醒文本的对比方法的流程图。具体地,参见图11,控制器在执行对人物声源信息进行声源识别,确定声源角度信息之前,被 进一步配置为执行下述步骤:FIG. 11 exemplarily shows a flowchart of a wake-up text comparison method according to some embodiments. Specifically, referring to Figure 11, the controller is further configured to perform the following steps before performing sound source identification on the character sound source information and determining the sound source angle information:
S1021、对人物声源信息进行文本提取,得到语音交互文本。S1021. Perform text extraction on the character sound source information to obtain a voice interaction text.
S1022、对比语音交互文本和预置唤醒文本,预置唤醒文本是指用于触发声源识别过程的文本。S1022 , comparing the voice interaction text and the preset wake-up text, the preset wake-up text refers to the text used to trigger the sound source identification process.
S1023、如果语音交互文本与预置唤醒文本对比一致,则执行对人物声源信息进行声源识别的步骤。S1023. If the voice interaction text is consistent with the preset wake-up text, perform the step of performing sound source recognition on the person's sound source information.
在一些实施例中,控制器在获取到人物声源信息后,先进行文本提取,提取出人物通过语音与显示设备交互时的语音交互文本。将提取出的语音交互文本与预置唤醒文本进行对比,如果对比不一致,例如,人物语音并非“海信小聚”,而是其他交互内容,此时,说明当前人物的语音并非触发调整摄像头拍摄角度的语音,控制器无需执行调整摄像头拍摄角度的相关步骤。In some embodiments, after acquiring the sound source information of the character, the controller first performs text extraction, and extracts the voice interaction text when the character interacts with the display device through voice. Compare the extracted voice interaction text with the preset wake-up text. If the comparison is inconsistent, for example, the character's voice is not "Hisense Xiaoju", but other interactive content. At this time, it means that the current character's voice does not trigger the adjustment of the camera's shooting angle. voice, the controller does not need to perform the relevant steps to adjust the camera's shooting angle.
如果对比一致,则说明当前人物的语音为触发调整摄像头拍摄角度的语音,例如,人物语音为预先设置的“海信小聚”,此时,控制器可继续执行后续调整摄像头拍摄角度的步骤。If the comparison is consistent, it means that the current character's voice is the voice that triggers the adjustment of the camera's shooting angle. For example, the character's voice is the preset "Hisense Xiaoju". At this time, the controller can continue to perform the subsequent steps to adjust the camera's shooting angle.
在判断出人物声源信息为唤醒语音,即调整摄像头拍摄角度的触发语音时,控制器需执行后续声源识别的过程。When judging that the person's sound source information is a wake-up voice, that is, a trigger voice for adjusting the shooting angle of the camera, the controller needs to perform a subsequent sound source recognition process.
由于显示设备中设置多组声音采集器,多组声音采集器可采集到同一人物语音时的多组人物声源信息,那么控制器在获取声音采集器采集的人物声源信息时,可获取到每个声音采集器采集的人物在语音时产生的人物声源信息,即控制器会获取到多组人物声源信息。Since multiple groups of sound collectors are set in the display device, the multiple groups of sound collectors can collect multiple groups of person sound source information when the same person's voice is spoken, so when the controller obtains the person sound source information collected by the sound collector, it can obtain Each sound collector collects the character sound source information generated when the character speaks, that is, the controller will acquire multiple sets of character sound source information.
图12中示例性示出了根据一些实施例的对人物声源信息进行声源识别的方法流程图。多组声音采集器采集同一唤醒文本时,由于每个声音采集器与人物之间的距离并不相同,因此,可对每个人物声源信息进行识别,以确定人物语音时的方位角度,即声源角度信息。具体地,参见图12,制器在执行对人物声源信息进行声源识别,确定声源角度信息,被进一步配置为执行下述步骤:FIG. 12 exemplarily shows a flow chart of a method for sound source identification for character sound source information according to some embodiments. When multiple groups of sound collectors collect the same wake-up text, since the distance between each sound collector and the character is not the same, the sound source information of each character can be identified to determine the azimuth angle of the character’s voice, that is, Sound source angle information. Specifically, referring to FIG. 12 , the controller is further configured to perform the following steps when performing sound source identification on the character sound source information and determining the sound source angle information:
S121、对每个人物声源信息分别进行声源识别,计算多组声音采集器在采集对应的人物声源信息时产生的语音时间差。S121. Perform sound source identification on each character sound source information, and calculate the voice time difference generated when multiple groups of sound collectors collect the corresponding character sound source information.
S122、基于语音时间差,计算人物在语音时所处位置的声源角度信息。S122 , based on the time difference of speech, calculate the sound source angle information of the position where the character is at the time of speech.
每个声音采集器的频率响应一致,其采样时钟也同步,但由于每个声音采集器与人物之间的距离并不相同,因此,每个声音采集器能够采集到语音的时刻也并非相同,多组声音采集器之间会存在采集时间差。The frequency response of each sound collector is the same, and its sampling clock is also synchronized. However, because the distance between each sound collector and the character is not the same, the time when each sound collector can collect speech is not the same. There will be a difference in acquisition time between multiple groups of sound collectors.
在一些实施例中,可以通过声音采集器阵列计算声源距离阵列的角度和距离,实现对人物语音时所处位置的声源进行跟踪。基于TDOA(Time Difference Of Arrival,到达时间差)的声源定位技术,估计信号到达两两麦克风之间的时间差,从而得到声源位置坐标的方程组,然后求解方程组即可得到声源的精确方位坐标,即声源角度信息。In some embodiments, the angle and distance of the sound source from the array can be calculated by the sound collector array, so as to realize the tracking of the sound source at the position of the character when speaking. Based on the TDOA (Time Difference Of Arrival) sound source localization technology, the time difference between the arrival of the signal between the two microphones is estimated, so as to obtain the equation set of the sound source position coordinates, and then the exact position of the sound source can be obtained by solving the equation set Coordinates, that is, sound source angle information.
在一些实施例中,在步骤S121中,控制器在执行对每个所述人物声源信息分别进行声源识别,计算多组所述声音采集器在采集对应的人物声源信息时产生的语音时间差,被进一步配置为执行下述步骤:In some embodiments, in step S121, the controller performs sound source identification for each of the character sound source information, and calculates the voices generated by the plurality of groups of the sound collectors when collecting the corresponding character sound source information. The time difference is further configured to perform the following steps:
步骤1211、在人物声源信息中提取环境噪声、人物语音时的声源信号和人物的语 音传播至每一声音采集器的传播时间。Step 1211: Extract the ambient noise, the sound source signal of the person's voice, and the propagation time of the person's voice to each sound collector from the person's sound source information.
步骤1212、根据环境噪声、声源信号和传播时间,确定每个声音采集器的接收信号。Step 1212: Determine the received signal of each sound collector according to the environmental noise, the sound source signal and the propagation time.
步骤1213、利用互相关时延估计算法,对每个声音采集器的接收信号进行处理,得到每两个声音采集器在采集对应的人物声源信息时产生的语音时间差。Step 1213 , using the cross-correlation time delay estimation algorithm to process the received signal of each sound collector to obtain the speech time difference generated when every two sound collectors collect the corresponding character sound source information.
在计算每两个声音采集器的语音时间差时,可利用声音采集器阵列实现声源到达方向估计(direction-of-arrival(DOA)estimation),由DOA估计算法计算声音到达不同声音采集器阵列间的时间差。When calculating the speech time difference between every two sound collectors, the sound source array can be used to estimate the direction-of-arrival (DOA) estimation. time difference.
在声源定位系统中,声音采集器阵列的每个阵元接收到的目标信号都来自于同一个声源。因此,各通道信号之间具有较强的相关性,通过计算每两路信号之间的相关函数,就可以确定每两个声音采集器观测信号之间的时延,即语音时间差。In the sound source localization system, the target signal received by each element of the sound collector array comes from the same sound source. Therefore, there is a strong correlation between the signals of each channel. By calculating the correlation function between each two channels of signals, the time delay between the signals observed by each two sound collectors, that is, the speech time difference, can be determined.
人物在语音时产生的人物声源信息中包括环境噪声和人物语音时的声源信号,还可在人物声源信息中通过识别提取出人物的语音传播至每一声音采集器的传播时间,计算每个声音采集器的接收信号。The character sound source information generated by the character during the speech includes the ambient noise and the sound source signal of the character voice, and the propagation time of the character's voice transmitted to each sound collector can also be extracted from the character sound source information by identifying and extracting, and calculating The received signal of each sound collector.
x i(t)=α is(t-τ i)+n i(t); x i (t)=α i s(t-τ i )+n i (t);
式中,x i(t)为第i个声音采集器的接收信号,s(t)为人物语音时的声源信号,τ i为人物的语音传播至第i个声音采集器的传播时间,n i(t)为环境噪声,α i为修正系数。 In the formula, x i (t) is the received signal of the i-th sound collector, s(t) is the sound source signal when the character's voice is spoken, τ i is the propagation time of the character's voice propagating to the i-th sound collector, n i (t) is the environmental noise, and α i is the correction coefficient.
利用互相关时延估计算法对每个声音采集器的接收信号进行处理,进行时延估计,表示为:
Figure PCTCN2021093589-appb-000001
式中,
Figure PCTCN2021093589-appb-000002
为第i个声音采集器与第i+1个声音采集器之间的时延,即语音时间差。
The cross-correlation delay estimation algorithm is used to process the received signal of each sound collector to estimate the delay, which is expressed as:
Figure PCTCN2021093589-appb-000001
In the formula,
Figure PCTCN2021093589-appb-000002
is the time delay between the i-th sound collector and the i+1-th sound collector, that is, the voice time difference.
带入每个声音采集器的接收信号模型,得到:Bringing in the received signal model of each sound collector, we get:
Figure PCTCN2021093589-appb-000003
Figure PCTCN2021093589-appb-000003
由于s(t)与n i(t)互不相关,因此可简化上式为:
Figure PCTCN2021093589-appb-000004
Since s(t) and n i (t) are independent of each other, the above formula can be simplified as:
Figure PCTCN2021093589-appb-000004
其中,τ ii+1=τ ii+1,n i与n i+1为互不相关的高斯白噪声,则上式进一步简化为: Among them, τ ii+1ii+1 , n i and n i+1 are Gaussian white noises that are not correlated with each other, the above formula is further simplified as:
Figure PCTCN2021093589-appb-000005
Figure PCTCN2021093589-appb-000005
由互相关时延估计算法的性质可知,当τ ii+1=τ ii+1时,
Figure PCTCN2021093589-appb-000006
取最大值,是两个声音采集器的时延,即语音时间差。
From the properties of the cross-correlation delay estimation algorithm, when τ ii+1ii+1 ,
Figure PCTCN2021093589-appb-000006
The maximum value is the time delay of the two sound collectors, that is, the voice time difference.
在声音采集器阵列信号处理实际模型中,由于存在混响和噪声影响,导致
Figure PCTCN2021093589-appb-000007
的峰值不明显,降低了时延估计的精度。为了锐化
Figure PCTCN2021093589-appb-000008
的峰值,可以根据信号和噪声的先验知识,在频域内对互功率谱进行加权,从而能抑制噪声和混响干扰。最后进行傅里叶逆变换,得到广义互相关函数
Figure PCTCN2021093589-appb-000009
In the actual model of sound collector array signal processing, due to the presence of reverberation and noise effects,
Figure PCTCN2021093589-appb-000007
The peak value is not obvious, which reduces the accuracy of delay estimation. for sharpening
Figure PCTCN2021093589-appb-000008
The peak value of , the cross-power spectrum can be weighted in the frequency domain according to the prior knowledge of signal and noise, so as to suppress noise and reverberation interference. Finally, perform inverse Fourier transform to obtain the generalized cross-correlation function
Figure PCTCN2021093589-appb-000009
Figure PCTCN2021093589-appb-000010
其中
Figure PCTCN2021093589-appb-000011
表示频域加权函数。
Figure PCTCN2021093589-appb-000010
in
Figure PCTCN2021093589-appb-000011
represents the frequency domain weighting function.
最后采用PHAT加权,使得信号间的互动率谱更加平滑,得到最终的每两个声音 采集器在采集对应的人物声源信息时产生的语音时间差
Figure PCTCN2021093589-appb-000012
经过PHAT加权的互功率谱近似于单位冲激响应的表达式,突出了时延的峰值,能够有效抑制混响噪声,提高时延(语音时间差)估计的精度和准确度。
Finally, PHAT weighting is used to make the interaction rate spectrum between the signals smoother, and the final speech time difference generated by each two sound collectors when collecting the corresponding character sound source information is obtained.
Figure PCTCN2021093589-appb-000012
The cross-power spectrum weighted by PHAT is similar to the expression of the unit impulse response, which highlights the peak value of the delay, which can effectively suppress the reverberation noise and improve the accuracy and accuracy of the delay (speech time difference) estimation.
在一些实施例中,在步骤S122中,控制器在执行基于语音时间差,计算人物在语音时所处位置的声源角度信息,被进一步配置为执行下述步骤:In some embodiments, in step S122, the controller is further configured to perform the following steps when calculating the sound source angle information of the position of the character when speaking based on the speech time difference:
步骤1221、获取当前环境状态下的声速、每个声音采集器的坐标和声音采集器的设置个数。Step 1221: Acquire the speed of sound in the current environmental state, the coordinates of each sound collector, and the set number of sound collectors.
步骤1222、根据声音采集器的设置个数,确定声音采集器的组合对数量,组合对数量是指声音采集器两两组合得到的组合数。Step 1222: Determine the number of combined pairs of sound collectors according to the set number of sound collectors, where the number of combined pairs refers to the number of combinations obtained by combining two sound collectors.
步骤1223、根据每两个声音采集器对应的语音时间差、声速和每个声音采集器的坐标,建立向量关系方程组,向量关系方程组的数量与组合对数量相同。Step 1223 , according to the speech time difference, the sound speed and the coordinates of each sound collector corresponding to each two sound collectors, establish a vector relational equation set, the number of which is the same as the number of combination pairs.
步骤1224、求解向量关系方程组,得到人物语音时所处位置的声源单位平面波传播向量的向量值。Step 1224: Solve the vector relation equation system to obtain the vector value of the unit plane wave propagation vector of the sound source at the position of the person's speech.
步骤1225、根据向量值,计算人物在语音时所处位置的声源角度信息。Step 1225: Calculate, according to the vector value, the sound source angle information of the position where the character is speaking.
在根据前述实施例提供的方法计算出每两个声音采集器的语音时间差后,可根据每个语音时间差计算人物在语音时所处位置的声源角度信息。After calculating the voice time difference between every two sound collectors according to the method provided in the foregoing embodiment, the sound source angle information of the position of the character when speaking can be calculated according to each voice time difference.
在计算声源角度信息时,需要建立多组向量关系方程组,为保证计算结果的准确性,可设定方程组的数量与声音采集器两两组合得到的组合数相同。为此,获取声音采集器的设置个数N,则所有声音采集器之间两两组合共有N(N-1)/2对组合对。When calculating the sound source angle information, it is necessary to establish multiple sets of vector relationship equations. In order to ensure the accuracy of the calculation results, the number of equations can be set to be the same as the number of combinations obtained by combining the sound collectors in pairs. To this end, the set number N of the sound collectors is obtained, and there are N(N-1)/2 pairs of combinations between all the sound collectors.
在建立向量关系方程组时,获取当前环境状态下的声速c和每个声音采集器的坐标,记第k个声音采集器的坐标为(x k,y k,z k),同时,设定人物语音时所处位置的声源单位平面波传播向量为u=(u,v,w),求解出人物语音时所处位置的声源单位平面波传播向量的向量值即可确定声源角度信息。 When establishing the vector relationship equation system, obtain the sound speed c and the coordinates of each sound collector in the current environmental state, and record the coordinates of the kth sound collector as (x k , y k , z k ), and at the same time, set The sound source unit plane wave propagation vector at the position of the character's speech is u=(u, v, w). The sound source angle information can be determined by solving the vector value of the sound source unit plane wave propagation vector at the character's voice position.
根据每两个声音采集器对应的语音时间差
Figure PCTCN2021093589-appb-000013
声速c、每个声音采集器的坐标(x k,y k,z k)和人物语音时所处位置的声源单位平面波传播向量为(u,v,w),建立N(N-1)/2个向量关系方程组:
Figure PCTCN2021093589-appb-000014
该式代表第i个声音采集器与第j个声音采集器之间建立的向量关系方程组。
According to the voice time difference corresponding to each two sound collectors
Figure PCTCN2021093589-appb-000013
The speed of sound c, the coordinates of each sound collector (x k , y k , z k ) and the unit plane wave propagation vector of the sound source at the location of the character’s speech are (u, v, w), and N(N-1) is established /2 vector relational equations:
Figure PCTCN2021093589-appb-000014
This formula represents the set of vector relationship equations established between the ith sound collector and the jth sound collector.
以N=3为例,可以建立以下方程组:Taking N=3 as an example, the following equations can be established:
Figure PCTCN2021093589-appb-000015
(第1个声音采集器与第2个声音采集器之间建立的向量关系方程组);
Figure PCTCN2021093589-appb-000015
(The set of vector relationship equations established between the first sound collector and the second sound collector);
Figure PCTCN2021093589-appb-000016
(第1个声音采集器与第3个声音采集器之间建立的向量关系方程组);
Figure PCTCN2021093589-appb-000016
(The set of vector relationship equations established between the first sound collector and the third sound collector);
Figure PCTCN2021093589-appb-000017
(第3个声音采集器与第2个声音采集器之间建立的向量关系方程组)。
Figure PCTCN2021093589-appb-000017
(The set of vector relationship equations established between the third sound collector and the second sound collector).
将上述三个向量关系方程组,写成矩阵形式:Write the above three vector relational equations in matrix form:
Figure PCTCN2021093589-appb-000018
Figure PCTCN2021093589-appb-000018
根据上述矩阵求解出u=(u,v,w),再利用正余弦关系,即可得到角度值:Solve u=(u,v,w) according to the above matrix, and then use the sine and cosine relationship to get the angle value:
Figure PCTCN2021093589-appb-000019
即人物在语音时所处位置的方位角度的声源角度信息。
Figure PCTCN2021093589-appb-000019
That is, the sound source angle information of the azimuth angle of the position of the character when speaking.
S13、基于摄像头的当前拍摄角度和声源角度信息,确定摄像头的目标转动方向和目标转动角度。S13. Determine the target rotation direction and target rotation angle of the camera based on the current shooting angle and sound source angle information of the camera.
控制器通过对人物声源信息进行声源识别,以确定出用于表征人物在语音时所处位置的方位角度的声源角度信息。声源角度信息可标识人物当前的所处位置,摄像头的当前拍摄角度可标识摄像头当前的所处位置,根据两个位置之间的相差角度即可确定摄像头在水平方向上需要转动的目标转动角度,以及摄像头在转动时的目标转动方向。The controller determines the sound source angle information used to represent the azimuth angle of the person's position when speaking by performing sound source identification on the sound source information of the person. The sound source angle information can identify the current position of the character, the current shooting angle of the camera can identify the current position of the camera, and the target rotation angle that the camera needs to rotate in the horizontal direction can be determined according to the difference angle between the two positions. , and the target rotation direction when the camera is rotated.
图13中示例性示出了根据一些实施例的确定摄像头的目标转动方向和目标转动角度的方法流程图。具体地,参见图13,控制器在执行基于摄像头的当前拍摄角度和声源角度信息,确定摄像头的目标转动方向和目标转动角度,被进一步配置为执行下述步骤:FIG. 13 exemplarily shows a flowchart of a method for determining a target rotation direction and a target rotation angle of a camera according to some embodiments. Specifically, referring to FIG. 13 , the controller is further configured to perform the following steps when determining the target rotation direction and target rotation angle of the camera based on the current shooting angle and sound source angle information of the camera:
S131、将声源角度信息转换为摄像头的坐标角度。S131. Convert the sound source angle information into the coordinate angle of the camera.
由于声源角度信息表征人物的所处方位角度,因此,为便于准确地根据声源角度信息和摄像头的当前拍摄角度计算出摄像头需要调整的方位角度,可将人物的声源角度信息转换为摄像头的坐标角度,即用摄像头的坐标角度来代替人物的声源角度信息。Since the sound source angle information represents the azimuth angle of the character, in order to accurately calculate the azimuth angle that the camera needs to adjust according to the sound source angle information and the current shooting angle of the camera, the sound source angle information of the character can be converted into the camera. The coordinate angle, that is, the coordinate angle of the camera is used to replace the sound source angle information of the character.
具体地,控制器在执行将声源角度信息转换为摄像头的坐标角度,被进一步配置为执行下述步骤:Specifically, when the controller converts the sound source angle information into the coordinate angle of the camera, the controller is further configured to perform the following steps:
步骤1311、获取人物在语音时的声源角度范围和摄像头转动时的预设角度范围。Step 1311: Acquire the sound source angle range when the character is speaking and the preset angle range when the camera rotates.
步骤1312、计算声源角度范围与预设角度范围之间的角度差值,将角度差值的半值作为转换角度。Step 1312: Calculate the angle difference between the sound source angle range and the preset angle range, and use the half value of the angle difference as the conversion angle.
步骤1313、计算声源角度信息对应的角度与转换角度的角度差,将角度差作为摄像头的坐标角度。Step 1313: Calculate the angle difference between the angle corresponding to the sound source angle information and the conversion angle, and use the angle difference as the coordinate angle of the camera.
由于声源角度范围和摄像头的预设角度范围并不相同,预设角度范围为0°~120°,声源角度范围为0°~180°,无法直接由摄像头的坐标角度代替声源角度信息。因此,先计算声源角度范围与预设角度范围之间的角度差值,再计算角度差值的半值,将半值作为由声源角度信息转换为摄像头的坐标角度时的转换角度。Since the angle range of the sound source is different from the preset angle range of the camera, the preset angle range is 0°~120°, and the sound source angle range is 0°~180°, so the coordinate angle of the camera cannot directly replace the sound source angle information . Therefore, first calculate the angle difference between the sound source angle range and the preset angle range, then calculate the half value of the angle difference, and use the half value as the conversion angle when the sound source angle information is converted into the coordinate angle of the camera.
声源角度范围与预设角度范围之间的角度差值为60°,角度差值的半值为30°,将30°作为转换角度。最后,计算声源角度信息对应的角度与转换角度的角度差,即为将声源角度信息转换成的摄像头的坐标角度。The angle difference between the sound source angle range and the preset angle range is 60°, the half value of the angle difference is 30°, and 30° is used as the conversion angle. Finally, the angle difference between the angle corresponding to the sound source angle information and the conversion angle is calculated, which is the coordinate angle of the camera converted from the sound source angle information.
例如,如果人物位于声音采集器的左侧,控制器通过获取多个声音采集器采集的人物声源信息确定出的声源角度信息对应的角度为50°,而转换角度为30°,因此,计算角度差为20°,即实现将声源角度信息对应的50°替换为摄像头的坐标角度20° 来表示。For example, if the character is located on the left side of the sound collector, the angle corresponding to the sound source angle information determined by the controller by acquiring the character sound source information collected by multiple sound collectors is 50°, and the conversion angle is 30°. Therefore, The calculated angle difference is 20°, that is, the 50° corresponding to the sound source angle information is replaced by the camera's coordinate angle of 20° to represent it.
如果人物位于声音采集器的右侧,控制器通过获取多个声音采集器采集的人物声源信息确定出的声源角度信息对应的角度为130°,而转换角度为30°,因此,计算角度差为100°,即实现将声源角度信息对应的130°替换为摄像头的坐标角度100°来表示。If the character is located on the right side of the sound collector, the angle corresponding to the sound source angle information determined by the controller by acquiring the character sound source information collected by multiple sound collectors is 130°, and the conversion angle is 30°. Therefore, the calculated angle The difference is 100°, that is, the 130° corresponding to the sound source angle information is replaced by the camera's coordinate angle of 100° to represent it.
S132、计算摄像头的坐标角度和摄像头的当前拍摄角度的角度差值,将角度差值作为摄像头的目标转动角度。S132: Calculate the angle difference between the coordinate angle of the camera and the current shooting angle of the camera, and use the angle difference as the target rotation angle of the camera.
摄像头的坐标角度用于标识人物所处位置在摄像头坐标内的角度,因此,根据摄像头的当前拍摄角度与摄像头的坐标角度的角度差值,即可确定出摄像头需要转动的目标转动角度。The coordinate angle of the camera is used to identify the angle of the person's position within the camera coordinates. Therefore, according to the angle difference between the current shooting angle of the camera and the coordinate angle of the camera, the target rotation angle that the camera needs to rotate can be determined.
例如,如果摄像头的当前拍摄角度为100°,摄像头的坐标角度为20°,说明摄像头当前的拍摄区域并未对准人物所处位置,二者相差80°,因此,需将摄像头转动80°后,摄像头的拍摄区域才可对准人物所处位置,即摄像头的目标转动角度为80°。For example, if the current shooting angle of the camera is 100° and the coordinate angle of the camera is 20°, it means that the current shooting area of the camera is not aimed at the position of the person, and the difference between the two is 80°. Therefore, it is necessary to rotate the camera 80° after , the shooting area of the camera can be aimed at the position of the person, that is, the target rotation angle of the camera is 80°.
S133、根据角度差值,确定摄像头的目标转动方向。S133: Determine the target rotation direction of the camera according to the angle difference.
由于以面对显示设备的方向,将左侧作为摄像头0°位置,右侧作为摄像头120°位置,因此,在根据摄像头的坐标角度和摄像头的当前拍摄角度确定出角度差值后,如果当前拍摄角度大于坐标角度,则说明摄像头的拍摄角度位于人物所处位置的右侧,此时角度差值为负值;如果当前拍摄角度小于坐标角度,则说明摄像头的拍摄角度位于人物所处位置的左侧,此时角度差值为正值。Since the direction facing the display device is used, the left side is taken as the 0° position of the camera, and the right side is taken as the 120° position of the camera. Therefore, after the angle difference is determined according to the coordinate angle of the camera and the current shooting angle of the camera, if the If the angle is greater than the coordinate angle, it means that the camera's shooting angle is on the right side of the character's position, and the angle difference is a negative value; if the current shooting angle is less than the coordinate angle, it means that the camera's shooting angle is on the left side of the character's position. side, the angle difference is a positive value at this time.
在一些实施例中,可根据角度差值的正负来确定摄像头的目标转动方向。如果角度差值为正值,说明摄像头的拍摄角度位于人物所处位置的左侧,此时,为使摄像头拍摄到人物的图像,需向右调整摄像头的拍摄角度,则确定摄像头的目标转动方向为向右转动。In some embodiments, the target rotation direction of the camera may be determined according to the positive or negative of the angle difference. If the angle difference is a positive value, it means that the shooting angle of the camera is on the left side of the position of the character. At this time, in order to make the image of the character captured by the camera, the shooting angle of the camera needs to be adjusted to the right, and the target rotation direction of the camera is determined. to turn right.
如果角度差值为负值,说明摄像头的拍摄角度位于人物所处位置的右侧,此时,为使摄像头拍摄到人物的图像,需向左调整摄像头的拍摄角度,则确定摄像头的目标转动方向为向左转动。If the angle difference is a negative value, it means that the shooting angle of the camera is located on the right side of the person's position. At this time, in order to make the camera capture the image of the person, it is necessary to adjust the shooting angle of the camera to the left, and then determine the target rotation direction of the camera. to turn left.
例如,图14中示例性示出了根据一些实施例的调整摄像头拍摄角度的一种场景图。参见图14,如果人物对应的声源角度信息对应的角度为50°,则转换成的摄像头的坐标角度为20°;摄像头的当前拍摄角度为100°,即摄像头的视角中心线位于人物所处位置的右侧,计算得到角度差值为-80°。可见角度差值为负值,此时,需调整摄像头向左转动80°。For example, FIG. 14 exemplarily shows a scene graph for adjusting the shooting angle of the camera according to some embodiments. Referring to Figure 14, if the angle corresponding to the sound source angle information corresponding to the character is 50°, the converted coordinate angle of the camera is 20°; the current shooting angle of the camera is 100°, that is, the center line of the camera's viewing angle is located at the position of the character. To the right of the position, the calculated angle difference is -80°. The visible angle difference is a negative value. At this time, the camera needs to be adjusted to rotate 80° to the left.
图15中示例性示出了根据一些实施例的调整摄像头拍摄角度的另一种场景图。参见图15,如果人物对应的声源角度信息对应的角度为120°,则转换成的摄像头的坐标角度为90°;摄像头的当前拍摄角度为40°,即摄像头的视角中心线位于人物所处位置的左侧,计算得到角度差值为50°。可见角度差值为正值,此时,需调整摄像头向右转动50°。FIG. 15 exemplarily shows another scene diagram for adjusting the shooting angle of the camera according to some embodiments. Referring to Figure 15, if the angle corresponding to the sound source angle information corresponding to the character is 120°, the converted coordinate angle of the camera is 90°; the current shooting angle of the camera is 40°, that is, the center line of the camera's viewing angle is located at the position of the character. To the left of the position, the calculated angle difference is 50°. The visible angle difference is a positive value. At this time, the camera needs to be adjusted to rotate 50° to the right.
S14、按照目标转动方向和目标转动角度,调整摄像头的拍摄角度,以使摄像头的拍摄区域正对人物语音时的所处位置。S14. Adjust the shooting angle of the camera according to the target rotation direction and the target rotation angle, so that the shooting area of the camera faces the position where the person's voice is located.
控制器在确定出摄像头需要调整拍摄角度时所需的目标转动方向和目标转动角度后,即可按照目标转动方向和目标转动角度调整摄像头的拍摄角度,将摄像头的拍摄 区域正对人物所处位置,使得摄像头可拍摄到包括人物的图像,实现根据人物的所处位置调整摄像头的拍摄角度。After the controller determines the target rotation direction and target rotation angle required when the camera needs to adjust the shooting angle, it can adjust the shooting angle of the camera according to the target rotation direction and target rotation angle, so that the shooting area of the camera is facing the position of the character. , so that the camera can capture images including characters, so that the shooting angle of the camera can be adjusted according to the position of the characters.
图16中示例性示出了根据一些实施例的人物语音时所处位置的场景图。由于摄像头的预设角度范围与人物语音时的声源角度范围不同,若体现在角度示意图中,参见图16,预设角度范围的0°位置与声源角度范围的0°位置之间存在30°的角度差值,同样的,预设角度范围的120°位置与声源角度范围的180°位置之间也存在30°的角度差值。FIG. 16 exemplarily shows a scene graph of the position of the character when speaking according to some embodiments. Since the preset angle range of the camera is different from the sound source angle range of the human voice, if it is reflected in the angle diagram, see Figure 16, there is a 30° position between the 0° position of the preset angle range and the 0° position of the sound source angle range ° angle difference, similarly, there is also a 30° angle difference between the 120° position of the preset angle range and the 180° position of the sound source angle range.
那么,如果人物在与显示设备交互时,其所处的位置恰好位于30°的夹角区域范围内,如图16中所示的人物(a)所处位置或人物(b)所处位置。此时,控制器在执行前述步骤S131中将声源角度信息转换为摄像头的坐标角度时,将会出现由人物的声源角度信息转换得到的摄像头的坐标角度为负值的情况,或者大于摄像头的预设角度范围最大值的情况,即转换得到的摄像头的坐标角度并未位于摄像头的预设角度范围内。Then, if the character interacts with the display device, the position of the character is just within the 30° angle range, as shown in Figure 16 where the character (a) is located or the character (b) is located. At this time, when the controller converts the sound source angle information into the coordinate angle of the camera in the aforementioned step S131, the coordinate angle of the camera converted from the sound source angle information of the character will be negative, or larger than the camera. The maximum value of the preset angle range, that is, the coordinate angle of the camera obtained by conversion is not within the preset angle range of the camera.
例如,若人物(a)所处位置对应的声源角度信息为20°,而转换角度为30°,则计算得到的摄像头的坐标角度为-10°。若人物(b)所处位置对应的声源角度信息为170°,而转换角度为30°,则计算得到的摄像头的坐标角度为140°。可见,根据人物(a)所处位置和人物(b)所处位置分别转换得到的摄像头的坐标角度均超出摄像头的预设角度范围。For example, if the sound source angle information corresponding to the position of the person (a) is 20° and the conversion angle is 30°, the calculated coordinate angle of the camera is -10°. If the sound source angle information corresponding to the position of the person (b) is 170°, and the conversion angle is 30°, the calculated coordinate angle of the camera is 140°. It can be seen that the coordinate angles of the camera respectively converted according to the position of the person (a) and the position of the person (b) are beyond the preset angle range of the camera.
如果摄像头的坐标角度均超出摄像头的预设角度范围,说明摄像头无法转动至摄像头的坐标角度(人物语音所处位置)对应的位置。而由于摄像头的可视角度范围位于60°~75°之间,说明在将摄像头转动到0°位置或者120°位置,摄像头的可视角度范围可覆盖预设角度范围的0°位置与声源角度范围的0°位置之间存在30°的角度差,以及,覆盖预设角度范围的120°位置与声源角度范围的180°位置之间存在30°的角度差。If the coordinate angles of the camera are all beyond the preset angle range of the camera, it means that the camera cannot be rotated to the position corresponding to the coordinate angle of the camera (where the voice of the person is located). Since the viewing angle range of the camera is between 60° and 75°, it means that when the camera is rotated to the 0° position or the 120° position, the viewing angle range of the camera can cover the 0° position of the preset angle range and the sound source. There is an angular difference of 30° between the 0° position of the angular range, and a 30° angular difference between the 120° position covering the preset angular range and the 180° position of the sound source angular range.
因此,如果人物的所处位置位于预设角度范围的0°位置与声源角度范围的0°位置之间存在30°的角度差范围内,或者,位于预设角度范围的120°位置与声源角度范围的180°位置之间存在30°的角度差范围内,则为了能够拍摄到包含人物的图像,按照摄像头的预设角度范围的最小值或最大值对应的位置,调整摄像头的拍摄角度。Therefore, if the position of the character is within a 30° angle difference between the 0° position of the preset angle range and the 0° position of the sound source angle range, or, if the character is located at the 120° position of the preset angle range and the sound source If there is a 30° angle difference between the 180° positions of the source angle range, in order to capture images including people, adjust the camera’s shooting angle according to the position corresponding to the minimum or maximum value of the camera’s preset angle range .
在一些实施例中,控制器被进一步配置为执行下述步骤:在人物的声源角度信息转换为摄像头的坐标角度超出摄像头的预设角度范围时,根据摄像头的当前拍摄角度与预设角度范围的最小值或最大值的角度差值,确定摄像头的目标转动方向和目标转动角度。In some embodiments, the controller is further configured to perform the following steps: when the angle information of the sound source of the character is converted into the coordinate angle of the camera beyond the preset angle range of the camera, according to the current shooting angle of the camera and the preset angle range The angle difference between the minimum or maximum value determines the target rotation direction and target rotation angle of the camera.
例如,如果人物(a)位于预设角度范围的0°位置与声源角度范围的0°位置之间存在30°的角度差范围内,即人物(a)的声源角度信息对应的声源角度为20°,摄像头的当前拍摄角度为50°时。根据摄像头的预设角度范围的最小值0°和当前拍摄角度50°计算角度差值,角度差值为-50°,则确定摄像头的目标转动方向为向左转动,目标转动角度为50°。此时,摄像头的视角中心线(a)与摄像头的0°线重合。For example, if the person (a) is located within a 30° angle difference between the 0° position of the preset angle range and the 0° position of the sound source angle range, that is, the sound source corresponding to the sound source angle information of the person (a) When the angle is 20°, the current shooting angle of the camera is 50°. Calculate the angle difference according to the minimum value of the camera's preset angle range of 0° and the current shooting angle of 50°. If the angle difference is -50°, the target rotation direction of the camera is determined to be leftward, and the target rotation angle is 50° . At this time, the center line (a) of the viewing angle of the camera coincides with the 0° line of the camera.
如果人物(b)位于预设角度范围的120°位置与声源角度范围的180°位置之间存在30°的角度差范围内,即人物(b)的声源角度信息对应的声源角度为170°,摄像头的当前拍摄角度为50°时。根据摄像头的预设角度范围的最大值120°和当前拍 摄角度50°计算角度差值,角度差值为70°,则确定摄像头的目标转动方向为向右转动,目标转动角度为70°。此时,摄像头的视角中心线(b)与摄像头的120°线重合。If the person (b) is located within a 30° angle difference between the 120° position of the preset angle range and the 180° position of the sound source angle range, that is, the sound source angle corresponding to the sound source angle information of the person (b) is 170°, when the current shooting angle of the camera is 50°. Calculate the angle difference according to the maximum value of the camera's preset angle range of 120° and the current shooting angle of 50°. If the angle difference is 70°, the target rotation direction of the camera is determined to be rightward, and the target rotation angle is 70°. At this time, the center line (b) of the viewing angle of the camera coincides with the 120° line of the camera.
因此,即使人物所处位置对应的声源角度超出摄像头在转动时的预设角度范围,本申请实施例提供的显示设备,仍可依据人物的所处位置,将摄像头转动至预设角度范围对应的最小值或最大值的位置,依据摄像头的可视角度覆盖范围,拍摄到包含人物的图像。Therefore, even if the sound source angle corresponding to the position of the character exceeds the preset angle range when the camera is rotated, the display device provided by the embodiment of the present application can still rotate the camera to the preset angle range according to the position of the character. The position of the minimum or maximum value of , depending on the viewing angle coverage of the camera, an image containing a person is captured.
可见,本申请实施例提供的一种显示设备,其中的摄像头可在预设角度范围内转动,控制器被配置为获取声音采集器采集的人物声源信息并进行声源识别,确定用于标识人物所在位置的方位角度的声源角度信息;基于摄像头的当前拍摄角度和声源角度信息,确定摄像头的目标转动方向和目标转动角度;按照目标转动方向和目标转动角度,调整摄像头的拍摄角度,以使摄像头的拍摄区域正对人物语音时的所处位置。可见,本申请提供的显示设备,可实现利用人物声源信息触发摄像头的转动,能够自动识别用户的实时所处位置并调整摄像头的拍摄角度,使得摄像头始终能够拍摄到包含人像的图像。It can be seen that, in the display device provided by the embodiment of the present application, the camera can be rotated within a preset angle range, and the controller is configured to obtain the sound source information of the characters collected by the sound collector and identify the sound source, and determine the sound source used for identification. The sound source angle information of the azimuth angle of the character's position; based on the current shooting angle and sound source angle information of the camera, determine the target rotation direction and target rotation angle of the camera; according to the target rotation direction and target rotation angle, adjust the camera's shooting angle, so that the shooting area of the camera is facing the position of the person's voice. It can be seen that the display device provided by the present application can trigger the rotation of the camera by using the sound source information of the person, and can automatically identify the real-time position of the user and adjust the shooting angle of the camera, so that the camera can always capture images containing the portrait.
前述实施例提供的显示设备,在调整摄像头的拍摄角度时,是基于人物与显示设备进行语音交互时的声源信息在水平方向上调整,使得人物的人像能够出现在摄像头的拍摄区域内,进而拍摄到包括人像的图像。The display device provided by the foregoing embodiment, when adjusting the shooting angle of the camera, is adjusted in the horizontal direction based on the sound source information when the character interacts with the display device, so that the portrait of the character can appear in the shooting area of the camera, and further. Images including portraits were captured.
调整拍摄角度后的摄像头,在拍摄人物的人像时,其视角中心线可能并未对准人物,这将使得摄像头拍摄的图像中,人像并未位于图像中心,人像出现偏离,影响视觉效果。因此,显示设备在调整摄像头的拍摄角度以将人像拍摄到之后,还可通过自动聚焦定位人像位置,以将人像显示在图像的中心区域。After adjusting the shooting angle of the camera, when shooting a portrait of a person, the center line of its angle of view may not be aligned with the person, which will make the image captured by the camera not located in the center of the image, and the portrait will deviate, affecting the visual effect. Therefore, after adjusting the shooting angle of the camera to capture the portrait, the display device can also locate the position of the portrait through automatic focusing, so as to display the portrait in the central area of the image.
由于人物在与显示设备交互时,可能以站立姿势还可能以坐立姿势等,使得人物的面部与摄像头存在不同的高度差距。因此,在利用人物的声源信息调整摄像头的拍摄角度后,摄像头的拍摄区域可能位于人物头部的上方或下方,这将导致摄像头无法完整的拍摄到人物的人像。Since the character may be in a standing posture or a sitting posture when interacting with the display device, there is a different height gap between the face of the character and the camera. Therefore, after adjusting the shooting angle of the camera using the sound source information of the person, the shooting area of the camera may be located above or below the head of the person, which will cause the camera to fail to completely capture the person's portrait.
因此,需要在摄像头的拍摄区域位于人物头部上方时,将摄像头沿垂直方向向下调整;或在摄像头的拍摄区域位于人物头部下方时,将摄像头沿垂直方向向上调整;或在摄像头的拍摄区域位于人物头部左方时,将摄像头沿水平方向向右调整;或在摄像头的拍摄区域位于人物头部右方时,将摄像头沿水平方向向左调整。Therefore, it is necessary to adjust the camera vertically downward when the shooting area of the camera is above the character's head; or adjust the camera vertically upward when the shooting area of the camera is below the character's head; or When the area is to the left of the person's head, adjust the camera to the right in the horizontal direction; or when the shooting area of the camera is to the right of the person's head, adjust the camera to the left in the horizontal direction.
图17中示例性示出了根据一些实施例的在预设角度范围内摄像头转动的另一场景图。摄像头可在水平方向上转动,还可在垂直方向上转动。因此,摄像头的预设角度范围包括横向0~120°,纵向0~105°。参见图17示例性示出的摄像头在垂直方向上的转动角度:俯仰0°、俯仰90°、俯仰105°;摄像头在水平方向上的转动角度:水平0°、水平60°、水平120°。FIG. 17 exemplarily shows another scene diagram in which the camera rotates within a preset angle range according to some embodiments. The camera can be rotated horizontally as well as vertically. Therefore, the preset angle range of the camera includes 0 to 120° in the horizontal direction and 0 to 105° in the vertical direction. 17 exemplarily shows the rotation angles of the camera in the vertical direction: pitch 0°, pitch 90°, and pitch 105°; the rotation angles of the camera in the horizontal direction: horizontal 0°, horizontal 60°, horizontal 120°.
为此,本申请实施例提供的一种显示设备,结合前述实施例提供的基于声源信息调整摄像头的拍摄角度以囊括人物的人像后,还通过摄像头图像检测精准识别人物位置信息,以计算出人物的人像与摄像头的图像中心差值,以再次对摄像头的拍摄角度从水平方向和垂直方向进行微调整,使得人物的人像处于摄像头拍摄图像的中心,从而保证显示器图像人物居中。To this end, the display device provided by the embodiment of the present application, after adjusting the shooting angle of the camera based on the sound source information provided by the above-mentioned embodiment to include the portrait of the person, also accurately recognizes the position information of the person through the detection of the camera image, so as to calculate The difference between the portrait of the person and the image center of the camera is used to fine-tune the shooting angle of the camera from the horizontal and vertical directions again, so that the portrait of the person is in the center of the image captured by the camera, so as to ensure that the person in the display image is centered.
图18中示例性示出了根据一些实施例的摄像头的控制方法的流程图;图19中示例性示出了根据一些实施例的摄像头的控制方法的整体数据流图。参见图18和图19,本申请实施例提供的一种显示设备,在对摄像头进行微调时,其控制器被配置为执行下述步骤:FIG. 18 exemplarily shows a flowchart of a camera control method according to some embodiments; FIG. 19 exemplarily shows an overall data flow diagram of a camera control method according to some embodiments. Referring to FIG. 18 and FIG. 19 , in a display device provided by an embodiment of the present application, when fine-tuning the camera, its controller is configured to perform the following steps:
S21、获取摄像头的拍摄参数和采集的人物位于摄像头拍摄区域内的指定图像。S21. Acquire the shooting parameters of the camera and the collected designated image of the person located in the shooting area of the camera.
在对摄像头的拍摄角度进行微调时,在一些实施例中,采用图像检测的方式,通过识别图像中正在拍摄的人物的人像,自动聚焦定位,调整摄像头的拍摄角度,以将人像显示在图像的中心位置。When fine-tuning the shooting angle of the camera, in some embodiments, an image detection method is used, by identifying the portrait of the person being photographed in the image, automatically focusing and positioning, and adjusting the shooting angle of the camera to display the portrait in the image. Central location.
为此,控制器在进行摄像头微调时,实时获取摄像头采集的指定图像,指定图像中包括人物在摄像头拍摄区域内所成的人像。For this reason, when the controller performs fine-tuning of the camera, it acquires the specified image captured by the camera in real time, and the specified image includes the portrait of the person in the shooting area of the camera.
在一些实施例中,如果摄像头执行过基于人物声源信息进行拍摄角度调整的过程后,此时,采集到指定图像的摄像头,是经过调整拍摄角度后的摄像头。那么,控制器需获取调整拍摄角度后的摄像头的拍摄参数和采集的人物位于摄像头拍摄区域内的指定图像。In some embodiments, if the camera has performed the process of adjusting the shooting angle based on the voice source information of the person, at this time, the camera that captures the specified image is the camera after the shooting angle has been adjusted. Then, the controller needs to obtain the shooting parameters of the camera after adjusting the shooting angle and the designated image of the captured person located in the shooting area of the camera.
摄像头的拍摄参数包括摄像头水平视角角度、图像水平宽度、摄像头垂直视角角度和图像垂直高度。摄像头水平视角角度是指摄像头的预设角度范围在横向的角度为0~120°,摄像头垂直视角角度是指摄像头的预设角度范围在纵向的角度为0~105°。图像水平宽度和图像垂直高度与摄像头的分辨率有关,如果摄像头支持1080P图像预览,则图像水平宽度为1920像素,图像垂直高度为1080像素。The shooting parameters of the camera include the horizontal viewing angle of the camera, the horizontal width of the image, the vertical viewing angle of the camera, and the vertical height of the image. The horizontal viewing angle of the camera means that the preset angle range of the camera is 0 to 120° in the horizontal direction, and the vertical viewing angle of the camera means that the preset angle range of the camera is 0 to 105° in the vertical direction. The image horizontal width and image vertical height are related to the resolution of the camera. If the camera supports 1080P image preview, the image horizontal width is 1920 pixels and the image vertical height is 1080 pixels.
S22、对指定图像进行识别处理,得到人物对应的人像区域位置,人像区域位置是指包括人物头部图像的区域。S22 , performing identification processing on the designated image to obtain the position of the portrait region corresponding to the person, where the position of the portrait region refers to the region including the head image of the person.
为基于人像进行定位聚焦显示,控制器对摄像头采集的指定图像进行识别处理,识别出图像中的人像并获取人头部区域位置,作为人像区域位置,以便于能够准确地将人像显示在指定图像的中心区域。In order to perform positioning and focusing display based on the portrait, the controller recognizes the designated image collected by the camera, identifies the portrait in the image, and obtains the position of the head area as the position of the portrait area, so that the portrait can be accurately displayed on the designated image. the central area.
摄像头采集到的指定图像可同步显示在显示器中进行预览,而人像区域位置可采用人脸框的形式显示在指定图像中,进而显示器中显示的指定图像中也显示人脸框。人脸框是将人像的头部和/或少许肢体部位框起来的矩形框或方形框。The designated image captured by the camera can be simultaneously displayed on the monitor for preview, and the position of the portrait area can be displayed in the designated image in the form of a face frame, and then the face frame is also displayed in the designated image displayed on the monitor. A face frame is a rectangular or square frame that encloses the head and/or a few body parts of a portrait.
由于与显示设备进行交互的人物可为多个,那么摄像头在采集指定图像时,指定图像中可能包括多个人物的人像,那么在确定人像区域位置时,需同时考虑多个人物的人像。Since there can be multiple people interacting with the display device, when the camera captures a specified image, the specified image may include portraits of multiple people, so when determining the position of the portrait area, the portraits of multiple people need to be considered at the same time.
具体地,控制器在执行对指定图像进行识别处理,得到人物对应的人像区域位置,被进一步配置为:Specifically, the controller is further configured as:
步骤221、对指定图像进行识别处理,得到至少一个人物对应的头部区域位置信息。Step 221: Perform identification processing on the designated image to obtain the position information of the head region corresponding to at least one person.
步骤222、计算至少一个人物对应的头部区域位置信息的总区域信息,将总区域信息对应的位置作为人物对应的人像区域位置,人像区域位置是指包括至少一个人物头部图像的总区域。Step 222: Calculate the total area information of the head area position information corresponding to at least one character, and use the position corresponding to the total area information as the portrait area position corresponding to the character, and the portrait area position refers to the total area including the head image of at least one character.
对指定图像进行人物数量的识别,如果指定图像中存在多个人像,则将得到多个人物对应的头部区域位置信息。头部区域位置信息是指人脸框所框住区域的位置信息,可以坐标的形式存在,每个人物的人像上均存在对应的人脸框,人像与人脸框为一对 一的关系。Identify the number of persons in the specified image, and if there are multiple portraits in the specified image, the position information of the head region corresponding to the multiple persons will be obtained. The position information of the head area refers to the position information of the area framed by the face frame, which can exist in the form of coordinates. There is a corresponding face frame on the portrait of each person, and the portrait and the face frame are in a one-to-one relationship.
计算至少一个人物对应的头部区域位置信息的总区域信息,即将每个人物对应的人脸框进行结合,得到一个总人脸框,总人脸框是指由多个人脸框所框总区域形成的最小矩形区域。Calculate the total area information of the head area position information corresponding to at least one character, that is, combine the face frames corresponding to each character to obtain a total face frame. The total face frame refers to the total area framed by multiple face frames. The smallest rectangular area formed.
在指定图像中存在多个人像时,总人脸框对应的人像区域位置内囊括多个人物头部图像。人像区域位置也可由位于指定图像中最顶端的那个人物头部位置作为总人脸框的顶边界点,由位于指定图像中最底端的那个人物头部位置作为总人脸框的底边界点,由位于指定图像中最左端的那个人物头部位置作为总人脸框的左边界点,由位于指定图像中最右端的那个人物头部位置作为总人脸框的右边界点。以四个边界点分别做平行于显示器相对应边的平行线,四条平行线两两垂直,相交后即可得到矩形的总人脸框。When there are multiple portraits in the specified image, the position of the portrait region corresponding to the total face frame includes multiple head images of the people. The position of the portrait area can also be used as the top boundary point of the total face frame by the position of the person's head at the top of the specified image, and the bottom boundary point of the total face frame by the position of the head of the person at the bottom of the specified image. The position of the character's head located at the far left in the specified image is used as the left boundary point of the total face frame, and the position of the character's head at the far right in the specified image is used as the right boundary point of the total face frame. Four boundary points are used to make parallel lines parallel to the corresponding sides of the display, and the four parallel lines are perpendicular to each other. After intersecting, a rectangular total face frame can be obtained.
S23、计算人像区域位置的区域中心与指定图像的图像中心的方位距离,方位距离用于标识水平方向距离和垂直方向距离。S23. Calculate the azimuth distance between the area center of the portrait area position and the image center of the specified image, where the azimuth distance is used to identify the horizontal distance and the vertical distance.
为准确确定摄像头需要微调整的拍摄角度,以使人像位于指定图像的图像中心,需先计算人像区域位置的区域中心与指定图像的图像中心的方位距离,方位距离即为控制摄像头调整拍摄角度的依据。In order to accurately determine the shooting angle that the camera needs to fine-tune so that the portrait is located in the image center of the designated image, it is necessary to first calculate the azimuth distance between the area center of the portrait area and the image center of the designated image. in accordance with.
由于人像相对于指定图像来说,可能存在水平方向上的偏移,还可能存在垂直方向上的便宜,因此,为对摄像头进行精准的微调,可在水平方向和垂直方向进行调整,则方位距离包括水平方向距离和垂直方向距离。Compared with the specified image, the portrait may be offset in the horizontal direction and may also be cheap in the vertical direction. Therefore, in order to fine-tune the camera accurately, it can be adjusted in the horizontal and vertical directions, and the azimuth distance Including horizontal distance and vertical distance.
图20中示例性示出了根据一些实施例的计算方位距离的方法流程图;图21中示例性示出了根据一些实施例的计算方位距离的示意图。参见图20和图21,控制器在执行所计算人像区域位置的区域中心与指定图像的图像中心的方位距离,被进一步配置为:FIG. 20 exemplarily shows a flowchart of a method for calculating azimuth distance according to some embodiments; FIG. 21 exemplarily shows a schematic diagram of calculating azimuth distance according to some embodiments. Referring to FIG. 20 and FIG. 21 , the controller is further configured as:
S231、获取人像区域位置的坐标信息和指定图像的图像中心坐标信息,图像中心坐标信息包括图像水平坐标和图像垂直坐标。S231. Acquire the coordinate information of the position of the portrait area and the image center coordinate information of the specified image, where the image center coordinate information includes the image horizontal coordinate and the image vertical coordinate.
在一些实施例中,在计算方位距离时,可根据人像区域位置的坐标位置和指定图像的坐标位置进行计算。控制器在对指定图像进行识别检测时,可以得到人像区域位置每个顶点的坐标信息,即左上顶点、右上顶点、左下顶点、右下顶点的像素坐标值。In some embodiments, when calculating the azimuth distance, the calculation may be performed according to the coordinate position of the position of the portrait area and the coordinate position of the specified image. When the controller recognizes and detects the specified image, it can obtain the coordinate information of each vertex of the position of the portrait area, that is, the pixel coordinate values of the upper left vertex, the upper right vertex, the lower left vertex, and the lower right vertex.
指定图像的图像中心P 0是摄像头采集画面的中心点,也即显示器的中心点。由于指定图像是由摄像头采集得到,因此,指定图像的尺寸与摄像头的分辨率相同,即如果摄像头的分辨率一定的情况下,摄像头所采集到的图像的宽高像素也是一定的,因此,可根据摄像头的分辨率,确定指定图像的图像中心坐标信息。 The image center P 0 of the specified image is the center point of the image captured by the camera, that is, the center point of the display. Since the specified image is captured by the camera, the size of the specified image is the same as the resolution of the camera, that is, if the resolution of the camera is constant, the width and height of the image captured by the camera are also constant. Determine the image center coordinate information of the specified image according to the resolution of the camera.
例如,如果摄像头支持1080P图像预览,则图像水平宽度为1920像素,图像垂直高度为1080像素,此时,以显示器左上角为坐标原点,沿显示器的表面由左向右的方向为X轴正向,由上至下的方向为Y轴正向,则指定图像的图像中心的横向坐标为960像素,纵向坐标为540像素,即指定图像的图像中心P 0坐标信息(x 0,y 0)为(960,540)。 For example, if the camera supports 1080P image preview, the horizontal width of the image is 1920 pixels, and the vertical height of the image is 1080 pixels. At this time, taking the upper left corner of the display as the coordinate origin, the direction from left to right along the surface of the display is the positive X-axis , the direction from top to bottom is the positive Y axis, then the horizontal coordinate of the image center of the specified image is 960 pixels, and the vertical coordinate is 540 pixels, that is, the image center P 0 coordinate information (x 0 , y 0 ) of the specified image is (960,540).
S232、基于人像区域位置的坐标信息,计算人像区域位置的区域中心坐标,区域中心坐标包括区域中心水平坐标和区域中心垂直坐标。S232 , based on the coordinate information of the location of the portrait area, calculate the area center coordinates of the portrait area location, where the area center coordinates include the area center horizontal coordinate and the area center vertical coordinate.
人像区域位置的坐标信息可确定人像区域位置的四个顶点像素坐标值,基于此,可计算得到人像区域位置的区域中心P 1的水平坐标和区域中心垂直坐标。 Coordinate position information area may be determined portrait portrait area location coordinate values of four vertices of the pixel, based on this calculated to obtain the center position of the area portrait region P and the horizontal coordinate of the coordinates 1 is a central vertical region.
例如,如果人像区域位置的坐标信息分别为:左上顶点A(200,100)、右上顶点B(500,100)、左下顶点C(200,400)、右下顶点D(500,400),则计算人像区域位置的区域中心P 1坐标(x 1,y 1)为(350,250)。 For example, if the coordinate information of the position of the portrait area is: upper left vertex A (200, 100), upper right vertex B (500, 100), lower left vertex C (200, 400), and lower right vertex D (500, 400), then calculate the area center P of the position of the portrait area 1 The coordinates (x 1 , y 1 ) are (350, 250).
S233、计算人像区域位置的区域中心水平坐标和指定图像的图像水平坐标的差值,得到人像区域位置的区域中心与指定图像的图像中心的水平方向距离。S233: Calculate the difference between the horizontal coordinate of the area center of the portrait area position and the image horizontal coordinate of the designated image, and obtain the horizontal distance between the area center of the portrait area position and the image center of the designated image.
S234、计算人像区域位置的区域中心垂直坐标和指定图像的图像垂直坐标的差值,得到人像区域位置的区域中心与指定图像的图像中心的垂直方向距离。S234. Calculate the difference between the vertical coordinate of the area center of the portrait area position and the image vertical coordinate of the designated image, and obtain the vertical distance between the area center of the portrait area position and the image center of the designated image.
在确定人像区域位置的区域中心与指定图像的图像中心的方位距离时,需分别计算水平方向距离和垂直方向距离。因此,在计算水平方向距离D时,以人像区域位置的区域中心水平坐标x 1和指定图像的图像水平坐标x 0的差值来确定;在计算垂直方向距离H时,以人像区域位置的区域中心垂直坐标y 1和指定图像的图像垂直坐标y 0的差值来确定。 When determining the azimuth distance between the area center of the portrait area position and the image center of the specified image, the horizontal and vertical distances need to be calculated separately. Therefore, when calculating the distance D in the horizontal direction, it is determined by the difference between the horizontal coordinate x 1 of the area center of the portrait area and the horizontal coordinate x 0 of the image of the specified image; when calculating the distance H in the vertical direction, it is determined by the area of the portrait area position. The difference between the center vertical coordinate y 1 and the image vertical coordinate y 0 of the specified image is determined.
例如,水平方向距离D=x 0-x 1=960-350=610;垂直方向距离H=y 0-y 1=540-250=290。在一些实施例中,水平方向距离和垂直方向距离均以像素坐标值来表示。 For example, the horizontal distance D=x 0 -x 1 =960-350=610; the vertical distance H=y 0 -y 1 =540-250=290. In some embodiments, both the horizontal distance and the vertical distance are expressed in pixel coordinate values.
S24、如果方位距离超过方位设定阈值,则根据方位距离和摄像头的拍摄参数,计算摄像头的目标调整角度。S24. If the azimuth distance exceeds the azimuth setting threshold, calculate the target adjustment angle of the camera according to the azimuth distance and the shooting parameters of the camera.
摄像头在采集人像时,如果人像未居于摄像头拍摄图像的中心区域,那么人像区域位置的区域中心与指定图像的图像中心会存在一定距离。因此,可通过预设方位设定阈值,来判断人像区域位置的区域中心与指定图像的图像中心是否存在距离差,进而判断出人像是否居于指定图像的中心区域。When the camera captures a portrait, if the portrait is not located in the central area of the image captured by the camera, there will be a certain distance between the area center of the portrait area and the image center of the specified image. Therefore, a threshold value can be set by a preset orientation to determine whether there is a distance difference between the area center of the portrait area and the image center of the designated image, thereby determining whether the portrait is located in the center area of the designated image.
将前述实施例确定的人像区域位置的区域中心和指定图像的图像中心的方位距离与方位设定阈值进行对比,如果人像区域位置的区域中心与指定图像的图像中心的方位距离大于或等于方位设定阈值,则说明人像未居于指定图像的中心区域,摄像头未进行聚焦拍摄,人像在指定图像中的显示位置出现偏离。因此,需要控制摄像头的拍摄角度,以调整到人像居于指定图像的中心区域。Compare the azimuth distance between the area center of the portrait area position determined in the previous embodiment and the image center of the specified image with the azimuth setting threshold, if the azimuth distance between the area center of the portrait area position and the image center of the specified image is greater than or equal to the azimuth setting threshold If the threshold is set, it means that the portrait is not located in the center area of the specified image, the camera is not focusing and shooting, and the display position of the portrait in the specified image is deviated. Therefore, it is necessary to control the shooting angle of the camera to adjust the portrait to the central area of the designated image.
为将摄像头采集到的人像居于指定图像的中心区域,需先确定摄像头的目标调整角度。由于摄像头可在水平方向转动,还可在垂直方向上转动,那么摄像头的目标调整角度包括目标水平调整角度和目标垂直调整角度。In order to place the portrait captured by the camera in the center area of the specified image, the target adjustment angle of the camera needs to be determined first. Since the camera can be rotated in the horizontal direction and also in the vertical direction, the target adjustment angle of the camera includes the target horizontal adjustment angle and the target vertical adjustment angle.
而由于指定图像的图像水平坐标可能与人像区域位置的区域中心水平坐标相同,指定图像的图像垂直坐标可能与人像区域位置的区域中心垂直坐标不相同,即摄像头的拍摄角度可能在水平方向上正对人物,在垂直方向上存在偏差,此时,无需控制摄像头在水平方向上调整拍摄角度,只需调整摄像头在垂直方向上的拍摄角度即可。同样的,还会存在指定图像的图像垂直坐标可能与人像区域位置的区域中心垂直坐标相同,指定图像的图像水平坐标可能与人像区域位置的区域中心水平坐标不相同,即摄像头的拍摄角度可能在垂直方向上正对人物,在水平方向上存在偏差,此时,无需控制摄像头在垂直方向上调整拍摄角度,只需调整摄像头在水平方向上的拍摄角度即可。Since the image horizontal coordinate of the specified image may be the same as the horizontal coordinate of the area center of the portrait area, the image vertical coordinate of the specified image may not be the same as the vertical coordinate of the area center of the portrait area, that is, the shooting angle of the camera may be positive in the horizontal direction. For characters, there is a deviation in the vertical direction. At this time, there is no need to control the camera to adjust the shooting angle in the horizontal direction, but only need to adjust the shooting angle of the camera in the vertical direction. Similarly, there is also a possibility that the image vertical coordinate of the specified image may be the same as the vertical coordinate of the area center of the portrait area, and the image horizontal coordinate of the specified image may be different from the area center horizontal coordinate of the portrait area, that is, the shooting angle of the camera may be in The vertical direction is facing the person, and there is a deviation in the horizontal direction. At this time, there is no need to control the camera to adjust the shooting angle in the vertical direction, just adjust the shooting angle of the camera in the horizontal direction.
因此,为准确判断摄像头是需要在水平方向上调整拍摄角度,还是在垂直方向上调整拍摄角度,还是同时在水平方向上和垂直方向上调整拍摄角度,用于判定的方位设定阈值包括水平设定阈值和垂直设定阈值。Therefore, in order to accurately determine whether the camera needs to adjust the shooting angle in the horizontal direction, the shooting angle in the vertical direction, or adjust the shooting angle in the horizontal direction and the vertical direction at the same time, the orientation setting threshold used for determination includes the horizontal setting threshold. Threshold and Vertical Threshold.
在计算摄像头的目标调整角度时,根据方位距离和摄像头的拍摄参数来确定。摄 像头的拍摄参数包括摄像头视角角度和图像宽度,具体地,摄像头的拍摄参数包括摄像头水平视角角度、图像水平宽度、摄像头垂直视角角度和图像垂直高度。在一些实施例中,摄像头水平视角角度范围为0~120°,摄像头垂直视角角度范围为0~105°。如果摄像头支持1080P图像预览,则图像水平宽度为1920像素,图像垂直高度为1080像素。When calculating the target adjustment angle of the camera, it is determined according to the azimuth distance and the shooting parameters of the camera. The shooting parameters of the camera include the viewing angle of the camera and the width of the image. Specifically, the shooting parameters of the camera include the horizontal viewing angle of the camera, the horizontal width of the image, the vertical viewing angle of the camera, and the vertical height of the image. In some embodiments, the horizontal viewing angle of the camera ranges from 0 to 120°, and the vertical viewing angle of the camera ranges from 0 to 105°. If the camera supports 1080P image preview, the horizontal width of the image is 1920 pixels, and the vertical height of the image is 1080 pixels.
在一些实施例中,在确定摄像头在水平方向上需调整的拍摄角度时,方位设定阈值为水平设定阈值,人像区域位置的区域中心与指定图像的图像中心的方位距离为水平方向距离,摄像头的拍摄参数包括摄像头水平视角角度和图像水平宽度。In some embodiments, when determining the shooting angle that the camera needs to adjust in the horizontal direction, the azimuth setting threshold is the horizontal setting threshold, and the azimuth distance between the area center of the portrait area position and the image center of the designated image is the horizontal direction distance, The shooting parameters of the camera include the horizontal viewing angle of the camera and the horizontal width of the image.
此时,控制器在执行如果方位距离超过方位设定阈值,则根据方位距离和摄像头的拍摄参数,计算摄像头的目标调整角度,被进一步配置为:如果水平方向距离大于水平设定阈值,则根据水平方向距离、摄像头水平视角角度和图像水平宽度,计算摄像头的目标水平调整角度。At this time, if the azimuth distance exceeds the azimuth setting threshold, the controller calculates the target adjustment angle of the camera according to the azimuth distance and the shooting parameters of the camera, and is further configured as: if the horizontal distance is greater than the horizontal setting threshold, according to The distance in the horizontal direction, the horizontal viewing angle of the camera, and the horizontal width of the image are used to calculate the target horizontal adjustment angle of the camera.
如果人像区域位置的区域中心与指定图像的图像中心的水平方向距离大于或等于水平设定阈值,说明摄像头在水平方向上人像的位置与指定图像的中心位置存在偏差,使得人像区域位置的区域中心与指定图像的图像中心存在一定距离。因此,为使人像能够居于指定图像的中心位置,需控制摄像头的拍摄角度进行调整,根据人像区域位置的区域中心与指定图像的图像中心的水平方向距离D、摄像头水平视角角度α和图像水平宽度IW,计算摄像头的目标水平调整角度θ 1If the horizontal distance between the area center of the portrait area position and the image center of the specified image is greater than or equal to the horizontal set threshold, it means that there is a deviation between the portrait position of the camera in the horizontal direction and the center position of the specified image, so that the area center of the portrait area position There is a certain distance from the image center of the specified image. Therefore, in order to make the portrait in the center of the specified image, it is necessary to control the shooting angle of the camera to adjust, according to the horizontal distance D between the area center of the portrait area and the image center of the specified image, the horizontal viewing angle of the camera α and the horizontal width of the image. IW, calculate the target horizontal adjustment angle θ 1 of the camera.
图22中示例性示出了根据一些实施例的摄像头水平视角角度的示意图;图23中示例性示出了根据一些实施例的计算目标水平调整角度的示意图。在计算摄像头的目标水平调整角度时,参见图22和图23,图像水平宽度:IW=1920,指定图像的图像水平坐标x 0=960;人像区域位置的区域中心水平坐标为x 1,摄像头水平视角角度为α。 FIG. 22 exemplarily shows a schematic diagram of a horizontal viewing angle of a camera according to some embodiments; FIG. 23 exemplarily shows a schematic diagram of calculating a target horizontal adjustment angle according to some embodiments. When calculating the target horizontal adjustment angle of the camera, see Figure 22 and Figure 23, the horizontal width of the image: IW=1920, the image horizontal coordinate of the specified image x 0 =960; the horizontal coordinate of the area center of the portrait area position is x 1 , the camera horizontal The viewing angle is α.
计算人像区域位置的区域中心与指定图像的图像中心的水平方向距离D=x 0-x 1 Calculate the horizontal distance D=x 0 -x 1 between the area center of the portrait area position and the image center of the designated image.
那么计算摄像头的目标水平调整角度:θ 1=atan(2*Math.abs(x 0-x 1)*tan(α/2)/IW)。 Then calculate the target horizontal adjustment angle of the camera: θ 1 =atan(2*Math.abs(x 0 -x 1 )*tan(α/2)/IW).
由于如果人物位于面对显示器的左侧位置,那么人像区域位置的区域中心位于指定图像的图像中心的左侧,即x 0>x 1(图21所示状态)。而如果人物位于面对显示器的右侧位置,那么人像区域位置的区域中心位于指定图像的图像中心的右侧,即x 0<x 1(图23所示状态)。可见,在计算人像区域位置的区域中心与指定图像的图像中心的水平方向距离D时,可能会出现负值的情况,因此,为准确得到摄像头的目标水平调整角度,在计算水平方向距离D时,对x 0-x 1的差值进行取绝对值计算。 Since if the person is located at the left position facing the display, the area center of the portrait area position is located to the left of the image center of the specified image, ie x 0 >x 1 (the state shown in Figure 21). And if the person is located on the right side facing the display, then the area center of the portrait area location is located to the right of the image center of the specified image, that is, x 0 <x 1 (the state shown in Figure 23). It can be seen that when calculating the horizontal distance D between the area center of the portrait area and the image center of the specified image, a negative value may occur. Therefore, in order to accurately obtain the target horizontal adjustment angle of the camera, when calculating the horizontal distance D , and calculate the absolute value of the difference between x 0 -x 1.
在一些实施例中,在确定摄像头在垂直方向上需调整的拍摄角度时,方位设定阈值为垂直设定阈值,人像区域位置的区域中心与指定图像的图像中心的方位距离为垂直方向距离,摄像头的拍摄参数包括摄像头垂直视角角度和图像垂直宽度。In some embodiments, when determining the shooting angle that the camera needs to adjust in the vertical direction, the azimuth setting threshold is the vertical setting threshold, and the azimuth distance between the area center of the portrait area position and the image center of the designated image is the vertical direction distance, The shooting parameters of the camera include the vertical viewing angle of the camera and the vertical width of the image.
此时,控制器在执行所述方位距离超过方位设定阈值,则根据所述距离和摄像头的拍摄参数,计算摄像头的目标调整角度,被进一步配置为:如果垂直方向距离大于垂直设定阈值,则根据垂直方向距离、摄像头垂直视角角度和图像垂直高度,计算摄像头的目标垂直调整角度。At this time, when the controller executes that the azimuth distance exceeds the azimuth setting threshold, it calculates the target adjustment angle of the camera according to the distance and the shooting parameters of the camera, and is further configured as: if the vertical distance is greater than the vertical setting threshold, Then, according to the vertical distance, the vertical viewing angle of the camera and the vertical height of the image, the vertical adjustment angle of the target of the camera is calculated.
如果人像区域位置的区域中心与指定图像的图像中心的垂直方向距离大于或等于垂直设定阈值,说明摄像头在垂直方向上人像的位置与指定图像的中心位置存在偏差,使得人像区域位置的区域中心与指定图像的图像中心存在一定距离。因此,为使人像 能够居于指定图像的中心位置,需控制摄像头的拍摄角度进行调整,根据人像区域位置的区域中心与指定图像的图像中心的垂直方向距离H、摄像头垂直视角角度β和图像垂直宽度IH,计算摄像头的目标垂直调整角度θ 2If the vertical distance between the area center of the portrait area position and the image center of the specified image is greater than or equal to the vertical set threshold, it means that there is a deviation between the portrait position of the camera in the vertical direction and the center position of the specified image, so that the area center of the portrait area position There is a certain distance from the image center of the specified image. Therefore, in order to make the portrait in the center of the specified image, it is necessary to control the shooting angle of the camera to adjust, according to the vertical distance H between the area center of the portrait area and the image center of the specified image, the vertical viewing angle of the camera β and the vertical width of the image. IH, calculate the target vertical adjustment angle θ 2 of the camera.
图24中示例性示出了根据一些实施例的摄像头垂直视角角度的示意图;图25中示例性示出了根据一些实施例的计算目标垂直调整角度的示意图。在计算摄像头的目标垂直调整角度时,参见图24和图25,图像垂直宽度:IH=1080,指定图像的图像垂直坐标y 0=540;人像区域位置的区域中心垂直坐标为y 1,摄像头垂直视角角度为β。 FIG. 24 exemplarily shows a schematic diagram of a vertical viewing angle of a camera according to some embodiments; FIG. 25 exemplarily shows a schematic diagram of calculating a vertical adjustment angle of a target according to some embodiments. When calculating the target vertical adjustment angle of the camera, see Figure 24 and Figure 25, the vertical width of the image: IH=1080, the image vertical coordinate y 0 =540 of the specified image; the vertical coordinate of the area center of the portrait area position is y 1 , the camera is vertical The viewing angle is β.
计算人像区域位置的区域中心与指定图像的图像中心的垂直方向距离H=y 0-y 1 Calculate the vertical distance H=y 0 -y 1 between the area center of the portrait area position and the image center of the designated image.
那么计算摄像头的目标垂直调整角度:θ 2=atan(2*Math.abs(y 0-y 1)*tan(β/2)/IH)。 Then calculate the target vertical adjustment angle of the camera: θ 2 =atan(2*Math.abs(y 0 -y 1 )*tan(β/2)/IH).
由于如果人物位于面对显示器的上侧位置,那么人像区域位置的区域中心位于指定图像的图像中心的上侧,即y 0>y 1(图21所示状态)。而如果人物位于面对显示器的下侧位置,那么人像区域位置的区域中心位于指定图像的图像中心的下侧,即y 0<y 1(图25所示状态)。可见,在计算人像区域位置的区域中心与指定图像的图像中心的垂直方向距离H时,可能会出现负值的情况,因此,为准确得到摄像头的目标垂直调整角度,在计算垂直方向距离H时,对y 0-y 1的差值进行取绝对值计算。 Since if the person is located at the upper position facing the display, the area center of the portrait area position is located at the upper side of the image center of the designated image, that is, y 0 >y 1 (the state shown in FIG. 21 ). On the other hand, if the person is located on the lower side facing the display, the area center of the portrait area position is located on the lower side of the image center of the specified image, that is, y 0 <y 1 (the state shown in FIG. 25 ). It can be seen that when calculating the vertical distance H between the area center of the portrait area and the image center of the specified image, a negative value may occur. Therefore, in order to accurately obtain the target vertical adjustment angle of the camera, when calculating the vertical distance H , and calculate the absolute value of the difference between y 0 -y 1.
可见,在人像区域位置的区域中心与指定图像的图像中心在水平方向上的方位距离超过水平设定阈值,和/或,在垂直方向上的方位预置超过垂直设定阈值时,则根据方位距离和摄像头的拍摄参数,计算摄像头的目标水平调整角度和/或目标垂直调整角度。控制器按照目标水平调整角度和/或目标垂直调整角度控制摄像头的拍摄角度进行调整,可以保证摄像头采集到的人像位于指定图像的中心区域。It can be seen that the azimuth distance in the horizontal direction between the area center of the portrait area position and the image center of the designated image exceeds the horizontal set threshold, and/or, when the azimuth preset in the vertical direction exceeds the vertical set threshold, then according to the azimuth The distance and the shooting parameters of the camera are used to calculate the target horizontal adjustment angle and/or the target vertical adjustment angle of the camera. The controller controls the shooting angle of the camera to adjust according to the target horizontal adjustment angle and/or the target vertical adjustment angle, which can ensure that the portrait captured by the camera is located in the center area of the designated image.
S25、基于摄像头的目标调整角度,调整摄像头的拍摄角度,以使人物的人像位于摄像头采集的指定图像的中心区域。S25. Adjust the angle based on the target of the camera, and adjust the shooting angle of the camera, so that the portrait of the person is located in the center area of the designated image collected by the camera.
在确定出摄像头需要进行调整的目标调整角度后,控制器即可发送控制指令至马达控制服务,由马达控制服务响应控制指令控制摄像头进行拍摄角度调整,调整拍摄角度后的摄像头在采集人像时,可使人像位于摄像头采集指定图像的中心区域。After determining the target adjustment angle that the camera needs to adjust, the controller can send a control command to the motor control service, and the motor control service responds to the control command to control the camera to adjust the shooting angle. The portrait can be placed in the center area of the specified image captured by the camera.
以面对显示器的方向,如果人像区域位置的区域中心在水平方向上位于指定图像的图像中心左侧,则按照目标水平调整角度向右转动摄像头;反之向左转动。如果人像区域位置的区域中心在水平方向上位于指定图像的图像中心上侧,则按照目标垂直调整角度向下转动摄像头;反之向上转动。In the direction facing the monitor, if the area center of the portrait area position is horizontally to the left of the image center of the specified image, the camera will be rotated to the right according to the target horizontal adjustment angle; otherwise, it will be rotated to the left. If the area center of the portrait area position is horizontally above the image center of the specified image, rotate the camera downward according to the target vertical adjustment angle; otherwise, rotate it upward.
控制器在控制摄像头调整拍摄角度时,如果转速过快,出现图像抖动,转动到指定角度时停止不稳。因此,为了能够获得稳定的图像,需要准确的确定摄像头在调整拍摄角度时的转动方向和转动速度。具体地,控制器在执行基于摄像头的目标调整角度,调整摄像头的拍摄角度,被进一步配置为:When the controller controls the camera to adjust the shooting angle, if the rotation speed is too fast, the image will shake, and it will stop unstable when it rotates to the specified angle. Therefore, in order to obtain a stable image, it is necessary to accurately determine the rotation direction and rotation speed of the camera when adjusting the shooting angle. Specifically, the controller is further configured to:
步骤251、根据摄像头的目标调整角度,确定摄像头的目标转动速度和目标调整方向。Step 251: Determine the target rotation speed and target adjustment direction of the camera according to the target adjustment angle of the camera.
由于摄像头的默认转速为90°/s,如果用默认最大转速转动摄像头,会造成摄像头转速过快,出现图像抖动,转动到指定角度时停止不稳。因此,在一些实施例中,将摄像头的转动速度与目标调整角度进行关联。Since the default rotation speed of the camera is 90°/s, if the camera is rotated at the default maximum rotation speed, the rotation speed of the camera will be too fast, the image will shake, and the camera will stop unstable when it rotates to the specified angle. Therefore, in some embodiments, the rotational speed of the camera is associated with the target adjustment angle.
设定最大转速逻辑值和最小转速逻辑值,使得摄像头在最大转速逻辑值和最小转速逻辑值对应的转速范围内进行转动,实现根据不同的目标调整角度控制摄像头以对 应的目标转动速度调整摄像头。例如,默认最大转速逻辑值为100,即100°/s,最小转速逻辑值为10,即100°/s。Set the logic value of the maximum speed and the logic value of the minimum speed, so that the camera rotates within the speed range corresponding to the logic value of the maximum speed and the logic value of the minimum speed. For example, the default maximum speed logic value is 100, which is 100°/s, and the minimum speed logic value is 10, which is 100°/s.
在一些实施例中,如果摄像头的目标调整角度大于或等于最大转速逻辑值,则将最大转速逻辑值作为摄像头的目标转动速度。如果摄像头的目标调整角度大于或等于最大转速逻辑值100,则设定摄像头的目标转动速度为100°/s。In some embodiments, if the target adjustment angle of the camera is greater than or equal to the maximum rotational speed logic value, the maximum rotational speed logic value is used as the target rotational speed of the camera. If the target adjustment angle of the camera is greater than or equal to the maximum rotational speed logic value of 100, the target rotation speed of the camera is set to 100°/s.
在一些实施例中,如果摄像头的目标调整角度小于或等于最小转速逻辑值,则将最小转速逻辑值作为摄像头的目标转动速度。如果摄像头的目标调整角度小于或等于最小转速逻辑值10,则设定摄像头的目标转动速度为10°/s。In some embodiments, if the target adjustment angle of the camera is less than or equal to the minimum rotational speed logic value, the minimum rotational speed logic value is used as the target rotational speed of the camera. If the target adjustment angle of the camera is less than or equal to the minimum rotation speed logic value of 10, the target rotation speed of the camera is set to 10°/s.
在一些实施例中,如果摄像头的目标调整角度位于最大转速逻辑值和最小转速逻辑值之间,则将目标调整角度的数值作为摄像头的目标转动速度。如果摄像头的目标调整角度位于100~10之间,则以实际的目标调整角度设定为摄像头的目标转动速度。例如,如果摄像头的目标调整角度为30,则设定摄像头的目标转动速度为30°/s。In some embodiments, if the target adjustment angle of the camera is between the maximum rotation speed logic value and the minimum rotation speed logic value, the value of the target adjustment angle is used as the target rotation speed of the camera. If the target adjustment angle of the camera is between 100 and 10, the actual target adjustment angle is set as the target rotation speed of the camera. For example, if the target adjustment angle of the camera is 30, set the target rotation speed of the camera to 30°/s.
可见,通过计算得到的摄像头目标调整角度,在转动前设置对应的摄像头转动速度,再执行摄像头旋转。从而实现在调整角度较小,则转动速度较为平缓,如果调整角度较大,则以较快速度实现旋转,以使得摄像头能够及时并稳定地实现拍摄角度的调整,使人像位于指定图像的中心区域。It can be seen that, by calculating the adjustment angle of the camera target, the corresponding camera rotation speed is set before the rotation, and then the camera rotation is performed. Therefore, when the adjustment angle is small, the rotation speed is relatively gentle. If the adjustment angle is large, the rotation is realized at a faster speed, so that the camera can adjust the shooting angle in a timely and stable manner, so that the portrait is located in the center area of the specified image. .
在确定摄像头的目标调整方向时,可根据人像区域位置的区域中心与指定图像的图像中心的方位距离的正负值进行确定。在水平方向上,如果水平方向距离(D=x 0-x 1)为负值,则说明摄像头采集的指定图像的图像中心位于人像区域位置的区域中心的左侧,此时,为使人像位于摄像头拍摄到指定图像的中心区域,需向右调整摄像头的拍摄角度,则确定摄像头的目标调整方向为向右转动。反之,如果水平方向距离(D=x 0-x 1)为正值,则确定摄像头的目标调整方向为向左转动。 When determining the target adjustment direction of the camera, it can be determined according to the positive or negative value of the azimuth distance between the area center of the portrait area position and the image center of the specified image. In the horizontal direction, if the horizontal distance (D=x 0 -x 1 ) is a negative value, it means that the image center of the specified image captured by the camera is located to the left of the center of the area where the portrait area is located. When the camera captures the center area of the specified image, it is necessary to adjust the shooting angle of the camera to the right, then determine that the target adjustment direction of the camera is to rotate to the right. On the contrary, if the distance in the horizontal direction (D=x 0 -x 1 ) is a positive value, it is determined that the target adjustment direction of the camera is to rotate to the left.
在垂直方向上,如果垂直方向距离(H=y 0-y 1)为负值,则说明摄像头采集的指定图像的图像中心位于人像区域位置的区域中心的上侧,此时,为使人像位于摄像头拍摄到指定图像的中心区域,需向下调整摄像头的拍摄角度,则确定摄像头的目标调整方向为向下转动。反之,如果垂直方向距离(H=y 0-y 1)为正值,则确定摄像头的目标调整方向为向上转动。 In the vertical direction, if the vertical distance (H=y 0 -y 1 ) is a negative value, it means that the image center of the specified image captured by the camera is located on the upper side of the area center of the portrait area position. When the camera captures the center area of the specified image, and the shooting angle of the camera needs to be adjusted downward, the target adjustment direction of the camera is determined to be downward rotation. Conversely, if the vertical distance (H=y 0 -y 1 ) is a positive value, it is determined that the target adjustment direction of the camera is upward rotation.
步骤252、按照目标调整角度、目标调整方向和目标转动速度,调整摄像头的拍摄角度。Step 252: Adjust the shooting angle of the camera according to the target adjustment angle, the target adjustment direction and the target rotation speed.
在确定出摄像头的目标调整角度、目标调整方向和目标转动速度,即可控制摄像头执行相应的转动,以调整拍摄角度,实现人物位置的聚焦定位,使得摄像头采集到的人像位于指定图像的中心,显示在显示器的中心区域。After the target adjustment angle, target adjustment direction and target rotation speed of the camera are determined, the camera can be controlled to perform the corresponding rotation to adjust the shooting angle and realize the focus positioning of the character position, so that the portrait captured by the camera is located in the center of the specified image. displayed in the center area of the display.
可见,本申请实施例提供的显示设备,在对摄像头进行控制时,基于前述实施例提供的显示设备中通过人物声源信息粗略调整摄像头的拍摄角度的方案的基础上,可再次对摄像头拍摄的图像进行识别检测,以更加精准的调整摄像头的拍摄角度,有效定位人物具体位置,以及摄像头拍摄图像人像检测精准度较高。本实施例提供的显示设备,通过对声源定位和摄像头图像分析进行综合利用,利用声源定位空间感知能力较强的优势,首先确认人物的大致位置,驱动摄像头朝向声源方向,同时利用摄像头图像分析精准度高的优点,对拍摄图像进行人物检测确定具体位置,驱动摄像头进行微调,以此达到精准定位,使摄像头拍摄人物能够在指定图像的中心区域显示,实现在显示器中聚焦显示。本实施例提供的显示设备适用于视频通话、健身等场景,人物 站立位置不在默认摄像头拍摄区域内对于快速准确定位聚焦人物十分有效。It can be seen that in the display device provided by the embodiments of the present application, when controlling the camera, based on the solution of roughly adjusting the shooting angle of the camera through the voice source information of the characters in the display device provided in the foregoing embodiment, the camera can be used again. The image is recognized and detected to more accurately adjust the shooting angle of the camera, effectively locate the specific position of the person, and the camera captures images with high portrait detection accuracy. The display device provided by this embodiment comprehensively utilizes sound source localization and camera image analysis, and takes advantage of the sound source localization's strong spatial perception ability. First, the approximate position of the character is confirmed, the camera is driven toward the sound source, and the camera is used at the same time. The advantage of high accuracy of image analysis is to perform person detection on the captured image to determine the specific position, and drive the camera to perform fine-tuning, so as to achieve precise positioning, so that the person captured by the camera can be displayed in the center area of the designated image, and the focused display is realized on the display. The display device provided in this embodiment is suitable for scenes such as video calls and fitness, and it is very effective for quickly and accurately locating the focused person if the person's standing position is not within the shooting area of the default camera.
前述实施例提供的显示设备,基于人像区域位置的区域中心与指定图像的图像中心在水平方向上的方位距离超过方位设定阈值,通过再次微调整摄像头的拍摄角度实现人像显示在显示器的中心区域。而在其他实施例中,如果人像区域位置的区域中心与指定图像的图像中心在水平方向上的方位距离未超过方位设定阈值,说明人像在指定图像中的显示并未出现偏离,体现在显示器中时,人像可显示在显示器的中心。此时,则无需微调整摄像头的拍摄角度。In the display device provided by the foregoing embodiment, based on the azimuth distance between the area center of the portrait area position and the image center of the designated image in the horizontal direction exceeding the azimuth setting threshold, the portrait is displayed in the center area of the display by fine-adjusting the shooting angle of the camera again. . In other embodiments, if the azimuth distance between the area center of the portrait area and the image center of the specified image in the horizontal direction does not exceed the azimuth setting threshold, it means that the display of the portrait in the specified image does not deviate, which is reflected in the display. When centered, the portrait can be displayed in the center of the display. In this case, there is no need to fine-tune the shooting angle of the camera.
但是,如果人物站立在距离显示设备上的摄像头较远的位置,那么摄像头采集到的指定图像中,人像显示的区域较小,导致人物无法在较远距离观看到显示器中显示的自己的人像。因此,为使人物在距离较远的情况下,仍然能够看清自己的人像,本申请实施例提供的显示设备,可对人像区域位置进行人像聚焦放大显示。However, if the person is standing far away from the camera on the display device, in the designated image captured by the camera, the area displayed for the person's portrait is small, so that the person cannot view the person's portrait displayed on the display from a long distance. Therefore, in order for the person to still be able to clearly see his own portrait even when the distance is relatively far, the display device provided by the embodiment of the present application can perform a portrait focus and magnification display on the position of the portrait area.
图26中示例性示出了根据一些实施例的人像聚焦放大显示的方法流程图。具体地,参见图26,基于前述实施例提供的显示设备的基础上,控制器被进一步配置为:FIG. 26 exemplarily shows a flowchart of a method for focusing and zooming in on a portrait display according to some embodiments. Specifically, referring to FIG. 26 , based on the display device provided by the foregoing embodiment, the controller is further configured as:
S26、如果方位距离未超过方位设定阈值,则获取预设数量帧的指定图像。S26. If the azimuth distance does not exceed the azimuth setting threshold, acquire a specified image of a preset number of frames.
S27、如果预设数量帧的指定图像中人像区域位置不变,则识别指定图像中的人像区域位置的尺寸。S27. If the position of the portrait area in the designated image of the preset number of frames does not change, identify the size of the position of the portrait area in the designated image.
S28、如果人像区域位置的尺寸小于或等于指定图像的预设比例,则将指定图像中的人像区域位置在显示器中进行人像聚焦放大显示。S28. If the size of the portrait area position is smaller than or equal to the preset ratio of the specified image, perform a portrait focus and magnification display on the display at the portrait area position in the specified image.
如果人像区域位置的区域中心与指定图像的图像中心在水平方向上的方位距离未超过方位设定阈值,说明摄像头当前采集到的指定图像中人像正处于中心位置,在此场景下,控制器获取预设数量帧的指定图像。在一些实施例中,预设数量帧可为20帧。If the azimuth distance in the horizontal direction between the area center of the portrait area position and the image center of the specified image does not exceed the azimuth setting threshold, it means that the portrait in the specified image currently captured by the camera is in the center position. In this scenario, the controller obtains The specified image for a preset number of frames. In some embodiments, the preset number of frames may be 20 frames.
若预设数量帧的指定图像中人像区域位置不变,说明人物当前保持相对区域不动。此时,由控制器对预设数量帧的图像识别,以判定人像区域位置相比整个指定图像区域占比较小时,会自动将人头部所在区域进行聚焦放大,以适应人物与显示设备远距离需求。If the position of the portrait area in the designated image of the preset number of frames does not change, it means that the person currently keeps the relative area unchanged. At this time, the controller will recognize the image of the preset number of frames to determine that the position of the portrait area is smaller than the entire designated image area, and will automatically focus and enlarge the area where the person's head is located to adapt to the distance between the person and the display device. need.
在一些实施例中,预设比例可设为三分之一,如果人像区域位置的尺寸小于或等于指定图像的三分之一,则说明人像区域位置显示过小,需进行聚焦放大显示。人像区域位置的比例计算方式可以像素面积(像素点数量)来计算。In some embodiments, the preset ratio can be set to one third. If the size of the portrait area position is less than or equal to one third of the specified image, it means that the portrait area position is displayed too small and needs to be focused and enlarged. The proportional calculation method of the position of the portrait area can be calculated by the pixel area (number of pixels).
在对指定图像中人像区域位置进行聚焦放大时,采用人像区域位置与显示器宽高比值的对比方式进行放大。具体地,控制器在执行如果人像区域位置的尺寸小于或等于指定图像的预设比例,则将指定图像中的人像区域位置在显示器中进行人像聚焦放大显示,被进一步配置为:When focusing and zooming in on the position of the portrait area in the specified image, the zoom is performed by comparing the position of the portrait area with the aspect ratio of the monitor. Specifically, if the size of the position of the portrait area is smaller than or equal to the preset ratio of the specified image, the controller is further configured to:
步骤281、如果人像区域位置的尺寸小于或等于指定图像的预设比例,则计算显示器的宽高比值和人像区域位置的宽高比值。Step 281: If the size of the portrait area position is smaller than or equal to the preset ratio of the specified image, calculate the aspect ratio value of the display and the aspect ratio value of the portrait area position.
图27中示例性示出了根据一些实施例的人像聚焦放大显示的示意图。参见图27,在一些实施例中,对指定图像中人像区域位置进行聚焦放大时,可依据人像区域位置与显示器的宽高比值来确定。因此,需分别计算显示器的宽高比值和人像区域位置的宽高比值。FIG. 27 exemplarily shows a schematic diagram of zoomed-in portrait display according to some embodiments. Referring to FIG. 27 , in some embodiments, when focusing and zooming in on the position of the portrait region in a specified image, it may be determined according to the ratio of the position of the portrait region to the aspect ratio of the display. Therefore, the aspect ratio of the display and the aspect ratio of the portrait area position need to be calculated separately.
宽高比值可根据像素坐标值来计算,显示器的宽高比值是显示器的宽度值与高度值的比值,而显示器的宽度值和高度值与摄像头的分辨率相同,即如果摄像头支持1080P图像预览,则图像水平宽度为1920像素,图像垂直高度为1080像素,那么显示器的宽度值为1920像素,高度值为1080像素值。The aspect ratio can be calculated according to the pixel coordinate value. The aspect ratio of the display is the ratio of the width and height of the display, and the width and height of the display are the same as the resolution of the camera, that is, if the camera supports 1080P image preview, Then the horizontal width of the image is 1920 pixels, and the vertical height of the image is 1080 pixels, then the width value of the display is 1920 pixels, and the height value is 1080 pixels.
人像区域位置的宽高比值是指人像区域位置的宽度值和高度值的比值。人像区域位置可仅包括人物头部位置,或者,包括人物头部位置和少许肢体部分。人像区域位置的宽度值与高度值可以想象坐标值进行确定,具体的方法可参照前述实施例中确定 人像区域位置的坐标信息的方法,此处不再赘述。The aspect ratio value of the portrait area position refers to the ratio of the width value and the height value of the portrait area position. The portrait area position may include only the position of the head of the character, or, the position of the head of the character and a few body parts. The width value and height value of the position of the portrait area can be determined by imagining the coordinate value, and the specific method can refer to the method for determining the coordinate information of the position of the portrait area in the foregoing embodiment, which will not be repeated here.
步骤282、如果显示器的宽高比值和人像区域位置的宽高比值不一致时,则调整人像区域位置的宽高比值,人像区域位置的调整后的宽高比值与显示器的宽高比值相同。Step 282: If the aspect ratio value of the display is inconsistent with the aspect ratio value of the portrait area position, adjust the aspect ratio value of the portrait area position, and the adjusted aspect ratio value of the portrait area position is the same as the aspect ratio value of the display.
由于同时与显示设备进行交互的人物可为多个,那么摄像头采集到的指定图像中会包括多个人物的人像,那么多个人物的人像所围成的人像区域位置可为长方形或矩形。为了对人像区域位置进行放大时,不会造成人像的变形,需要将人像区域位置的宽高比值与显示器的宽高比值相同。Since there can be multiple characters interacting with the display device at the same time, the designated image captured by the camera will include portraits of multiple characters, and the position of the portrait area enclosed by the portraits of multiple characters may be a rectangle or a rectangle. In order to enlarge the position of the portrait area without causing deformation of the portrait, the aspect ratio of the portrait area needs to be the same as the aspect ratio of the display.
如果显示器的宽高比值和人像区域位置的宽高比值不一致时,如图27中(a)所示的状态,则调整人像区域位置的宽高比值,以使得调整宽高比值后的人像区域位置的宽高比值与显示器的宽高比值相同,如图27中(b)所示的状态。显示器的宽高比值和人像区域位置的宽高比值不一致的情况包括两种情形,一是人像区域位置的宽高比值大于显示器的宽高比值的情形,二是人像区域位置的宽高比值小于显示器的宽高比值的情形。If the aspect ratio value of the display is inconsistent with the aspect ratio value of the portrait area position, as shown in (a) in Figure 27, adjust the aspect ratio value of the portrait area position so that the portrait area position after adjusting the aspect ratio value The aspect ratio value is the same as that of the display, as shown in Figure 27(b). There are two situations in which the aspect ratio of the display and the aspect ratio of the portrait area are inconsistent. One is that the aspect ratio of the portrait area is greater than the aspect ratio of the monitor, and the second is that the aspect ratio of the portrait area is smaller than that of the monitor. of the aspect ratio.
在一些实施例中,如果人像区域位置的宽高比值大于显示器的宽高比值,则调整人像区域位置的高度值,人像区域位置的原宽度值与调整后的高度值的宽高比值与显示器的宽高比值相同。In some embodiments, if the aspect ratio value of the portrait area position is greater than the aspect ratio value of the display, the height value of the portrait area position is adjusted, and the aspect ratio value of the original width value of the portrait area position and the adjusted height value is the same as that of the display. The aspect ratio values are the same.
如果人像区域位置的宽高比值大于显示器的宽高比值,则为了保持人像区域位置的尺寸同显示器的比例,则应以人像区域位置的区域中心点,上下两边进行扩充,将人像区域位置的高度值增大。If the aspect ratio of the portrait area position is greater than the aspect ratio of the monitor, in order to keep the size of the portrait area position with the ratio of the monitor, you should use the area center point of the portrait area position to expand the upper and lower sides, and the height of the portrait area position should be expanded. value increases.
那么,为了避免改变人像区域位置的区域中心点位置,需同时调整人像区域位置的高度值对应的上下两边,则上下两边各扩充大小为(IH*pW/IW-pH)/2,式中,IW为显示器的宽度值,IH为显示器的高度值,pW为人像区域位置的宽度值,pH为人像区域位置的高度值。Then, in order to avoid changing the position of the area center point of the portrait area position, it is necessary to adjust the upper and lower sides corresponding to the height value of the portrait area position at the same time, then the expansion size of the upper and lower sides is (IH*pW/IW-pH)/2, where, IW is the width value of the display, IH is the height value of the display, pW is the width value of the portrait area position, and pH is the height value of the portrait area position.
在一些实施例中,如果人像区域位置的宽高比值小于显示器的宽高比值,则调整人像区域位置的宽度值,人像区域位置的调整后的宽度值与原高度值的宽高比值与显示器的宽高比值相同。In some embodiments, if the aspect ratio value of the portrait area position is smaller than the aspect ratio value of the display, the width value of the portrait area position is adjusted, and the aspect ratio value of the adjusted width value of the portrait area position and the original height value is the same as that of the display. The aspect ratio values are the same.
如果人像区域位置的宽高比值小于显示器的宽高比值,则为了保持人像区域位置的尺寸同显示器的比例,则应以人像区域位置的区域中心点,左右两边进行扩充,将人像区域位置的宽度值增大。If the aspect ratio of the portrait area position is smaller than the aspect ratio of the monitor, in order to keep the size of the portrait area position with the ratio of the monitor, the center point of the portrait area position should be used to expand the left and right sides, and the width of the portrait area position should be expanded. value increases.
那么,为了避免改变人像区域位置的区域中心点位置,需同时调整人像区域位置的宽度值对应的左右两边,则左右两边各扩充大小为(PH*IW/IH-pW)/2,式中,IW为显示器的宽度值,IH为显示器的高度值,pW为人像区域位置的宽度值,pH为人像区域位置的高度值。Then, in order to avoid changing the position of the area center point of the portrait area position, it is necessary to adjust the left and right sides corresponding to the width value of the portrait area position at the same time, then the expansion size of the left and right sides is (PH*IW/IH-pW)/2, where, IW is the width value of the display, IH is the height value of the display, pW is the width value of the portrait area position, and pH is the height value of the portrait area position.
步骤283、按照宽高比值调整后的人像区域位置,确定人像区域位置的目标放大区域。Step 283: Determine the target enlarged area for the position of the portrait area according to the position of the portrait area adjusted by the aspect ratio.
由于人像区域位置中仅包括人物头部位置,或者,包括人物头部位置和少许肢体部分,若将人像区域位置直接放大显示在显示器中,则会出现失真现象。因此,为了防止较小的人像区域位置放大到全屏显示时,因放大比例过大造成图像失真严重,需确定目标放大区域。Since the position of the portrait area only includes the position of the head of the character, or includes the position of the head of the character and a few limbs, if the position of the portrait area is directly enlarged and displayed on the display, distortion will occur. Therefore, in order to prevent the image distortion from being seriously distorted due to the excessive enlargement ratio when the position of the small portrait area is enlarged to full-screen display, it is necessary to determine the target enlargement area.
目标放大区域是显示器中将要显示的区域,目标放大区域包括人像区域位置以及周围区域位置,在一些实施例中,目标放大区域为人像区域位置的1.5倍左右。基于宽高比值调整后的人像区域位置按照1.5倍进行区域扩大,即可得到目标放大区域,如图27中(c)所示的虚线矩形区域。将目标放大区域对应的图像放大到全屏显示,不会造成图像失真现象。The target enlargement area is the area to be displayed in the display, and the target enlargement area includes the position of the portrait area and the surrounding area. In some embodiments, the target enlargement area is about 1.5 times the position of the portrait area. The position of the portrait area adjusted based on the aspect ratio is enlarged by 1.5 times, and the target enlarged area can be obtained, such as the dotted rectangle area shown in (c) in Figure 27. Enlarge the image corresponding to the target zoom area to full screen display without causing image distortion.
步骤284、将目标放大区域对应的人像进行聚焦放大,全屏显示在显示器中。Step 284 , focus and enlarge the portrait corresponding to the target enlargement area, and display it on the display in full screen.
将目标放大区域对应的图像放大到全屏显示,在实现人像聚焦放大显示的同时,不会造成图像失真现象。Enlarging the image corresponding to the target zoom area to a full-screen display will not cause image distortion while realizing the focus zoom display of the portrait.
由于目标放大区域是以显示器同比例调整后的人像区域位置以区域中心为中心点同比例放大得到的区域,可能会因为靠近图像边缘而有部分区域超出边界,因此,为避免放大显示时部分区域的图像无法显示在显示器中,则将目标放大区域根据超出边界部分进行调整,使超出边界部分与指定图像对应的边缘重合。Since the target magnification area is the area of the portrait area adjusted in the same proportion of the monitor, and the center of the area is the center point and the area is enlarged in the same proportion, some areas may exceed the boundary due to being close to the edge of the image. If the image cannot be displayed on the monitor, adjust the target zoom area according to the part beyond the boundary, so that the part beyond the boundary coincides with the edge corresponding to the specified image.
具体地,控制器在执行将目标放大区域对应的人像进行聚焦放大,全屏显示在显示器中,被进一步配置为执行下述步骤:Specifically, the controller is further configured to perform the following steps when performing focusing and zooming in on the portrait corresponding to the target zoom area and displaying it on the display in full screen:
步骤2841、获取目标放大区域的中心点坐标。Step 2841: Obtain the coordinates of the center point of the target zoom-in area.
由于目标放大区域是以人像区域位置的区域中心为中心点,按照放大比例倍数进行放大后得到的区域,因此,目标放大区域的中心与人像区域位置的区域中心相同,目标放大区域的中心点坐标即为人像区域位置的区域中心坐标。Since the target magnified area is the center point of the area center of the portrait area, the area obtained by enlarging according to the magnification ratio, therefore, the center of the target magnified area is the same as the area center of the portrait area, and the coordinates of the center point of the target magnified area It is the area center coordinate of the portrait area position.
步骤2842、计算中心点坐标与目标放大区域的任一条边界的第一距离,以及,中心点坐标与显示器的任一条边界的第二距离,目标放大区域的任一条边界与显示器的任一条边界位置相对应。Step 2842, calculate the first distance between the center point coordinates and any border of the target zoom area, and the second distance between the center point coordinates and any border of the display, and the position of any border of the target zoom area and any border of the display Corresponding.
人像区域位置被放大后,得到的目标放大区域的某一条边有可能超出显示器的某一条边界,例如,如图27中(c)所示,目标放大区域的左边界超出显示器的左边界。After the position of the portrait area is enlarged, a certain side of the obtained target enlarged area may exceed a certain boundary of the display. For example, as shown in (c) in Figure 27, the left boundary of the target enlarged area exceeds the left boundary of the display.
为了在目标放大区域的某一条边超过显示器的边界时对目标放大区域的位置进行调整,计算目标放大区域的中心点坐标与目标放大区域的任一条边界的第一距离L 1,计算目标放大区域的中心点坐标与显示器的任一条边界的第二距离L 2In order to adjust the position of the target enlargement area when a certain side of the target enlargement area exceeds the boundary of the display, calculate the first distance L 1 between the center point coordinates of the target enlargement area and any boundary of the target enlargement area, and calculate the target enlargement area The second distance L 2 between the coordinates of the center point of the display and any border of the display.
例如,计算目标放大区域的中心点坐标与目标放大区域的左边界的第一距离L 11,计算目标放大区域的中心点坐标与目标放大区域的上边界的第一距离L 12,计算目标放大区域的中心点坐标与目标放大区域的右边界的第一距离L 13,计算目标放大区域的中心点坐标与目标放大区域的下边界的第一距离L 14 For example, calculate the first distance L 11 between the coordinates of the center point of the target enlargement area and the left border of the target enlargement area, calculate the first distance L 12 between the center point coordinates of the target enlargement area and the upper boundary of the target enlargement area, and calculate the target enlargement area The first distance L 13 between the coordinates of the center point of the target enlargement area and the right border of the target enlargement area is calculated, and the first distance L 14 between the center point coordinates of the target enlargement area and the lower boundary of the target enlargement area is calculated.
计算目标放大区域的中心点坐标与显示器的左边界的第二距离L 21,计算目标放大区域的中心点坐标与显示器的上边界的第二距离L 22,计算目标放大区域的中心点坐标与显示器的右边界的第二距离L 23,计算目标放大区域的中心点坐标与显示器的下边界的第二距离L 24 Calculate the second distance L 21 between the coordinates of the center point of the target enlargement area and the left border of the display , calculate the second distance L 22 between the coordinates of the center point of the target enlargement area and the upper border of the display, and calculate the coordinates of the center point of the target enlargement area and the display. The second distance L 23 of the right border is calculated, and the second distance L 24 between the coordinates of the center point of the target enlargement area and the lower border of the display is calculated.
步骤2843、如果第二距离与第一距离的距离差小于零,则按照距离差调整目标放大区域的位置。Step 2843: If the distance difference between the second distance and the first distance is less than zero, adjust the position of the target enlarged area according to the distance difference.
在判断目标放大区域是否超出显示器的某一条边界时,以位于同一侧的目标放大区域的那条边对应的第一距离与显示器的那条边对应的第二距离做差来判断。When judging whether the target enlargement area exceeds a certain boundary of the display, the judgment is based on the difference between the first distance corresponding to the side of the target enlargement area located on the same side and the second distance corresponding to the side of the display.
计算第二距离L 2与第一距离L 1的距离差,如果小于零,说明第一距离对应的那条边超出第二距离对应的那条边。例如,如图27中(c)所示,显示器的左边界对应的第二距离L 21与目标放大区域的左边界对应的第一距离L 11的距离差小于零,说明目标放大区域的左边界超出显示器的左边界。 Calculate the distance difference between the second distance L 2 and the first distance L 1 , if it is less than zero, it means that the edge corresponding to the first distance exceeds the edge corresponding to the second distance. For example, as shown in (c) of FIG. 27 , the distance difference between the second distance L 21 corresponding to the left border of the display and the first distance L 11 corresponding to the left border of the target enlargement area is less than zero, indicating that the left border of the target enlargement area is less than zero. beyond the left border of the display.
在距离差小于零时,将目标放大区域整体沿目标放大区域超出显示器的那条边的相反方向移动位置,使得目标放大区域中超出显示器的那条边与显示器的那条边重合。例如,如图27中(d)所示,目标放大区域的左边界超出显示器的左边界,则将目标放大区域整体向右移动,使得目标放大区域的左边界与显示器的左边界重合。When the distance difference is less than zero, the entire target enlarged area is moved in the opposite direction of the side of the target enlarged area beyond the display, so that the side of the target enlarged area beyond the display coincides with the side of the display. For example, as shown in FIG. 27(d), if the left border of the target enlargement area exceeds the left border of the display, the entire target enlargement area is moved to the right so that the left border of the target enlargement area coincides with the left border of the display.
如果目标放大区域的右边界超出,则目标放大区域整体向左平移,使右边界与显示器右边界重合;如果目标放大区域的上边界超出,则目标放大区域整体向下平移, 使上边界与显示器上边界重合;如果目标放大区域的下边界超出,则目标放大区域整体向上平移,使下边界与显示器下边界重合。If the right border of the target enlargement area exceeds, the entire target enlargement area is shifted to the left, so that the right border coincides with the right border of the display; The upper boundary coincides; if the lower boundary of the target enlargement area exceeds, the entire target enlargement area is shifted upward, so that the lower boundary coincides with the lower boundary of the display.
目标放大区域移动位置时的移动程度以距离差来决定,即按照距离差的值来调整目标放大区域的位置。例如,如果距离差为L 0=|L 21-L 11|,则将目标放大区域整体向右移动L 0的距离,以使目标放大区域的左边界与显示器的左边界重合,使得目标放大区域内的所有图像都显示在显示器中。 The degree of movement of the target enlarged area when moving the position is determined by the distance difference, that is, the position of the target enlarged area is adjusted according to the value of the distance difference. For example, if the distance difference is L 0 =|L 21 -L 11 |, move the entire target enlargement area to the right by the distance L 0 so that the left border of the target enlargement area coincides with the left border of the display, so that the target enlargement area is All images within are displayed on the monitor.
步骤2844、将位置调整后的目标放大区域对应的人像进行聚焦放大,全屏显示在显示器中。Step 2844 , focus and zoom in on the portrait corresponding to the target zoom-in area whose position has been adjusted, and display it on the display in full screen.
位置调整后的目标放大区域内的所有图像都显示在显示器中,进而可对目标放大区域对应的人像进行聚焦放大,即目标放大区域整体所包含的图像全屏显示在显示器中,聚焦放大显示效果如图27中(e)所示。All images in the target zoom area after the position adjustment are displayed on the monitor, and then the portrait corresponding to the target zoom area can be focused and enlarged, that is, the images contained in the entire target zoom area are displayed on the monitor in full screen, and the focus zoom display effect is as follows This is shown in (e) of FIG. 27 .
在一些实施例中,如果摄像头采集人物的人像时,人像位于显示器的中心区域,且人物始终未改变自身位置,此时,显示设备无需控制摄像头调整拍摄角度,以当前拍摄角度持续拍摄人物的人像。累积预设数量帧的指定图像,对人物均未发生位置变化,则在指定图像中人像区域位置占指定图像的比例较小时,将人像区域位置进行人像聚焦放大显示,以将人像区域位置对应的图像全屏显示在显示器中。In some embodiments, when the camera captures the portrait of the person, if the portrait is located in the central area of the display, and the person never changes his position, at this time, the display device does not need to control the camera to adjust the shooting angle, and continues to shoot the portrait of the person at the current shooting angle . Accumulate the designated images of the preset number of frames, and there is no change in the position of the person. When the proportion of the portrait area in the designated image in the designated image is small, the portrait area will be displayed in a zoomed-in portrait focus, so as to display the position corresponding to the portrait area. The image is displayed full screen in the monitor.
但是,如果人像区域位置以聚焦放大形式显示在显示器中时,人物出现位置改变,显示设备需重新确定人像区域位置的区域中心,若人像区域位置的区域中心与指定图像的图像中心存在方位距离,则需控制摄像头调整拍摄角度,以保证人像始终位于指定图像的中心,并显示在显示器的中心区域。However, if the position of the portrait area is displayed on the display in the form of focus and magnification, and the position of the person appears changes, the display device needs to re-determine the area center of the portrait area position. Then you need to control the camera to adjust the shooting angle to ensure that the portrait is always in the center of the specified image and displayed in the center area of the monitor.
由于人物由未改变位置的状态变为改变位置的状态时,显示器中显示的是聚焦放大显示的人像区域位置,因此,为保证基于图像检测识别方法判断摄像头调整拍摄角度的准确性,需将显示器中正处于聚焦放大显示的人像区域位置恢复至原始状态,再进行后续计算摄像头的目标调整角度的步骤。Because when the person changes from the state of unchanged position to the state of changing position, the display shows the position of the portrait area displayed in focus and magnification. Therefore, in order to ensure the accuracy of judging the camera to adjust the shooting angle based on the image detection and recognition method, the monitor needs to be adjusted. The position of the portrait area that is currently in focus and magnified display is restored to the original state, and then the subsequent steps of calculating the target adjustment angle of the camera are performed.
具体地,控制器在执行计算摄像头的目标调整角度之前,被进一步配置为执行下述步骤:Specifically, before the controller performs the calculation of the target adjustment angle of the camera, it is further configured to perform the following steps:
步骤0241、判断指定图像是否进行人像聚焦放大显示操作。Step 0241: Determine whether the specified image is subjected to a portrait focus zoom display operation.
步骤0242、如果指定图像未进行人像聚焦放大显示操作,则执行计算摄像头的目标调整角度的步骤。Step 0242: If the specified image has not been subjected to the portrait focus and zoom-in display operation, execute the step of calculating the target adjustment angle of the camera.
步骤0243、如果指定图像已进行人像聚焦放大显示操作,则恢复指定图像的显示,以及,执行计算摄像头的目标调整角度的步骤。Step 0243: If the specified image has been displayed by focusing and zooming in on the portrait, restore the display of the specified image, and perform the step of calculating the target adjustment angle of the camera.
如果控制器对指定图像中的人像区域位置进行人像聚焦放大显示操作,会在当前指定图像上生成放大标记。控制器若检测到当前指定图像上存在放大标记,则可判定指定图像进行人像聚焦放大显示操作;若未检测到放大标记,则判定指定图像未进行人像聚焦放大显示操作。If the controller performs the portrait focus magnification display operation on the portrait area position in the specified image, a magnification mark will be generated on the current specified image. If the controller detects that there is a magnifying mark on the currently designated image, it can determine that the designated image is subjected to the portrait focus zoom display operation; if no zoom mark is detected, it is determined that the designated image has not been subjected to the portrait focus zoom display operation.
在控制器判断指定图像未进行人像聚焦放大显示操作时,可直接对指定图像进行图像检测分析,继续执行后续的计算摄像头的目标调整角度的步骤。When the controller determines that the specified image has not been subjected to the portrait focus and magnification display operation, it can directly perform image detection and analysis on the specified image, and continue to perform the subsequent steps of calculating the target adjustment angle of the camera.
在控制器判断指定图像已进行人像聚焦放大显示操作时,由于放大显示的指定图像会对图像检测分析造成准确率影响,因此,需先恢复指定图像至原始状态,取消人像聚焦放大显示操作,此时,显示器中显示的是处于原始状态的指定图像,而后继续 执行后续的计算摄像头的目标调整角度的步骤。When the controller judges that the specified image has been zoomed in portrait display operation, the zoomed-in specified image will affect the accuracy of image detection and analysis. Therefore, it is necessary to restore the specified image to its original state and cancel the zoomed in portrait display operation. At the time, the specified image in the original state is displayed on the display, and then the subsequent steps of calculating the target adjustment angle of the camera are continued.
由以上技术方案可知,本申请实施例提供的一种显示设备,控制器对摄像头采集的指定图像进行识别处理,得到人像区域位置,计算人像区域位置的区域中心与指定图像的图像中心的方位距离;如果方位距离超过方位设定阈值,则根据方位距离和摄像头的拍摄参数,计算摄像头的目标调整角度;基于摄像头的目标调整角度,调整摄像头的拍摄角度,以使人物的人像位于摄像头采集的指定图像的中心区域。可见,本申请实施例提供的显示设备,通过摄像头图像检测精准识别人物位置信息,自动聚焦定位人像位置,以对摄像头的拍摄角度从水平方向和垂直方向进行微调整,使得人物的人像处于摄像头拍摄图像的中心,从而保证显示器图像人物居中。It can be seen from the above technical solutions that in the display device provided by the embodiment of the present application, the controller performs recognition processing on the designated image collected by the camera, obtains the position of the portrait area, and calculates the azimuth distance between the area center of the portrait area position and the image center of the designated image. ;If the azimuth distance exceeds the azimuth setting threshold, calculate the target adjustment angle of the camera according to the azimuth distance and the shooting parameters of the camera; adjust the shooting angle of the camera based on the target adjustment angle of the camera, so that the portrait of the person is located at the specified location captured by the camera. the central area of the image. It can be seen that the display device provided by the embodiment of the present application accurately recognizes the position information of the person through the image detection of the camera, and automatically focuses and locates the position of the portrait, so as to finely adjust the shooting angle of the camera from the horizontal direction and the vertical direction, so that the portrait of the person is captured by the camera. The center of the image, thus ensuring that the display image character is centered.
图18中示例性示出了根据一些实施例的摄像头的控制方法的流程图。参见图18,本申请还提供了一种摄像头的控制方法,所述方法包括:FIG. 18 exemplarily shows a flowchart of a control method of a camera according to some embodiments. Referring to FIG. 18 , the present application also provides a method for controlling a camera, the method comprising:
S21、获取所述摄像头的拍摄参数和采集的人物位于摄像头拍摄区域内的指定图像;S21, acquiring the shooting parameters of the camera and a designated image of the collected character located in the camera shooting area;
S22、对所述指定图像进行识别处理,得到所述人物对应的人像区域位置,所述人像区域位置是指包括人物头部图像的区域;S22, performing identification processing on the designated image to obtain a portrait area position corresponding to the person, where the portrait area position refers to an area including a head image of a person;
S23、计算所述人像区域位置的区域中心与所述指定图像的图像中心的方位距离,所述方位距离用于标识水平方向距离和垂直方向距离;S23, calculate the azimuth distance between the area center of the portrait area position and the image center of the designated image, and the azimuth distance is used to identify the horizontal direction distance and the vertical direction distance;
S24、如果所述方位距离超过方位设定阈值,则根据所述方位距离和摄像头的拍摄参数,计算摄像头的目标调整角度;S24, if the azimuth distance exceeds the azimuth setting threshold, calculate the target adjustment angle of the camera according to the azimuth distance and the shooting parameters of the camera;
S25、基于所述摄像头的目标调整角度,调整所述摄像头的拍摄角度,以使所述人物的人像位于摄像头采集的指定图像的中心区域。S25. Adjust the shooting angle of the camera based on the target adjustment angle of the camera, so that the portrait of the person is located in the center area of the designated image captured by the camera.
具体实现中,本申请还提供一种计算机存储介质,其中,该计算机存储介质可存储有程序,该程序执行时可包括本申请提供的摄像头的控制方法的各实施例中的部分或全部步骤。所述的存储介质可为磁碟、光盘、只读存储记忆体(英文:read-only memory,简称:ROM)或随机存储记忆体(英文:random access memory,简称:RAM)等。In a specific implementation, the present application further provides a computer storage medium, wherein the computer storage medium can store a program, and when the program is executed, it can include some or all of the steps in each embodiment of the camera control method provided by the present application. The storage medium may be a magnetic disk, an optical disk, a read-only memory (English: read-only memory, abbreviated as: ROM) or a random access memory (English: random access memory, abbreviated as: RAM) and the like.
本领域的技术人员可以清楚地了解到本申请实施例中的技术可借助软件加必需的通用硬件平台的方式来实现。基于这样的理解,本申请实施例中的技术方案本质上或者说对相关技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品可以存储在存储介质中,如ROM/RAM、磁碟、光盘等,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本申请各个实施例或者实施例的某些部分所述的方法。Those skilled in the art can clearly understand that the technology in the embodiments of the present application can be implemented by means of software plus a necessary general hardware platform. Based on this understanding, the technical solutions in the embodiments of the present application can be embodied in the form of software products in essence or in the parts that make contributions to related technologies, and the computer software products can be stored in storage media, such as ROM/RAM, A magnetic disk, an optical disk, etc., includes several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to execute the methods described in various embodiments or some parts of the embodiments of the present application.
最后应说明的是:以上各实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述各实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分或者全部技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的范围。Finally, it should be noted that the above embodiments are only used to illustrate the technical solutions of the present application, but not to limit them; although the present application has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that: The technical solutions described in the foregoing embodiments can still be modified, or some or all of the technical features thereof can be equivalently replaced; and these modifications or replacements do not make the essence of the corresponding technical solutions deviate from the technical solutions of the embodiments of the present application. Scope.
为了方便解释,已经结合具体的实施方式进行了上述说明。但是,上述示例性的讨论不是意图穷尽或者将实施方式限定到上述公开的具体形式。根据上述的教导,可以得到多种修改和变形。上述实施方式的选择和描述是为了更好的解释原理以及实际的应用,从而使得本领域技术人员更好的使用所述实施方式以及适于具体使用考虑的各种不同的变形的实施方式。For the convenience of explanation, the above description has been made in conjunction with specific embodiments. However, the above exemplary discussions are not intended to be exhaustive or to limit implementations to the specific forms disclosed above. Numerous modifications and variations are possible in light of the above teachings. The above embodiments are chosen and described to better explain the principles and practical applications, so as to enable those skilled in the art to better utilize the described embodiments and various modified embodiments suitable for specific use considerations.

Claims (17)

  1. 一种显示设备,其特征在于,包括:A display device, comprising:
    摄像头,所述摄像头被配置为采集人像以及实现在预设角度范围内的转动;a camera, the camera is configured to capture a portrait and realize rotation within a preset angle range;
    与所述摄像头连接的控制器,所述控制器被配置为:A controller connected to the camera, the controller being configured to:
    获取所述摄像头的拍摄参数和采集的人物位于摄像头拍摄区域内的指定图像;Acquiring the shooting parameters of the camera and the designated images of the collected characters located in the shooting area of the camera;
    对所述指定图像进行识别处理,得到所述人物对应的人像区域位置,所述人像区域位置是指包括人物头部图像的区域;Performing identification processing on the designated image to obtain a portrait area position corresponding to the person, where the portrait area position refers to an area including the head image of the person;
    计算所述人像区域位置的区域中心与所述指定图像的图像中心的方位距离,所述方位距离用于标识水平方向距离和垂直方向距离;Calculate the azimuth distance between the area center of the portrait area position and the image center of the designated image, and the azimuth distance is used to identify the horizontal direction distance and the vertical direction distance;
    如果所述方位距离超过方位设定阈值,则根据所述方位距离和摄像头的拍摄参数,计算摄像头的目标调整角度;If the azimuth distance exceeds the azimuth setting threshold, calculate the target adjustment angle of the camera according to the azimuth distance and the shooting parameters of the camera;
    基于所述摄像头的目标调整角度,调整所述摄像头的拍摄角度,以使所述人物的人像位于摄像头采集的指定图像的中心区域。Based on the target adjustment angle of the camera, the shooting angle of the camera is adjusted so that the portrait of the person is located in the center area of the designated image captured by the camera.
  2. 根据权利要求1所述的显示设备,其特征在于,所述控制器在执行所述对指定图像进行识别处理,得到所述人物对应的人像区域位置,被进一步配置为:The display device according to claim 1, wherein the controller is further configured to perform the recognizing process on the designated image to obtain the position of the portrait region corresponding to the person:
    对所述指定图像进行识别处理,得到至少一个人物对应的头部区域位置信息;Performing identification processing on the designated image to obtain the position information of the head region corresponding to at least one character;
    计算至少一个人物对应的头部区域位置信息的总区域信息,将所述总区域信息对应的位置作为人物对应的人像区域位置,所述人像区域位置是指包括至少一个人物头部图像的总区域。Calculate the total area information of the head area position information corresponding to at least one character, take the position corresponding to the total area information as the portrait area position corresponding to the character, and the portrait area position refers to the total area including the head image of at least one character .
  3. 根据权利要求1所述的显示设备,其特征在于,所述方位距离包括水平方向距离和垂直方向距离;以及,所述控制器在执行所述计算人像区域位置的区域中心与所述指定图像的图像中心的方位距离,被进一步配置为:The display device according to claim 1, wherein the azimuth distance includes a horizontal direction distance and a vertical direction distance; Azimuth distance from the center of the image, further configured as:
    获取所述人像区域位置的坐标信息和所述指定图像的图像中心坐标信息,所述图像中心坐标信息包括图像水平坐标和图像垂直坐标;Obtain the coordinate information of the position of the portrait area and the image center coordinate information of the designated image, and the image center coordinate information includes the image horizontal coordinate and the image vertical coordinate;
    基于所述人像区域位置的坐标信息,计算所述人像区域位置的区域中心坐标,所述区域中心坐标包括区域中心水平坐标和区域中心垂直坐标;Based on the coordinate information of the position of the portrait area, calculate the area center coordinates of the portrait area position, and the area center coordinates include the horizontal coordinates of the area center and the vertical coordinates of the area center;
    计算所述人像区域位置的区域中心水平坐标和指定图像的图像水平坐标的差值,得到所述人像区域位置的区域中心与所述指定图像的图像中心的水平方向距离;Calculate the difference between the horizontal coordinate of the area center of the portrait area position and the image horizontal coordinate of the designated image, and obtain the horizontal distance between the area center of the portrait area position and the image center of the designated image;
    计算所述人像区域位置的区域中心垂直坐标和指定图像的图像垂直坐标的差值,得到所述人像区域位置的区域中心与所述指定图像的图像中心的垂直方向距离。Calculate the difference between the vertical coordinate of the area center of the portrait area position and the image vertical coordinate of the designated image, and obtain the vertical distance between the area center of the portrait area position and the image center of the designated image.
  4. 根据权利要求1所述的显示设备,其特征在于,所述方位设定阈值包括水平设定阈值,所述方位距离包括水平方向距离,所述摄像头的拍摄参数包括摄像头水平视角角度和图像水平宽度;The display device according to claim 1, wherein the azimuth setting threshold comprises a horizontal setting threshold, the azimuth distance comprises a horizontal distance, and the shooting parameters of the camera comprise a horizontal viewing angle of the camera and a horizontal width of the image ;
    所述控制器在执行所述如果方位距离超过方位设定阈值,则根据所述方位距离和摄像头的拍摄参数,计算摄像头的目标调整角度,被进一步配置为:The controller calculates the target adjustment angle of the camera according to the azimuth distance and the shooting parameters of the camera, and is further configured as:
    如果所述水平方向距离大于所述水平设定阈值,则根据所述水平方向距离、摄像头水平视角角度和图像水平宽度,计算摄像头的目标水平调整角度。If the horizontal distance is greater than the horizontal set threshold, calculate the target horizontal adjustment angle of the camera according to the horizontal distance, the horizontal viewing angle of the camera, and the horizontal width of the image.
  5. 根据权利要求1所述的显示设备,其特征在于,所述方位设定阈值包括垂直设定阈值,所述方位距离包括垂直方向距离,所述摄像头的拍摄参数包括摄像头垂直视 角角度和图像垂直高度;The display device according to claim 1, wherein the azimuth setting threshold comprises a vertical setting threshold, the azimuth distance comprises a vertical distance, and the shooting parameters of the camera comprise a vertical viewing angle of the camera and a vertical height of the image ;
    所述控制器在执行所述如果方位距离超过方位设定阈值,则根据所述方位距离和摄像头的拍摄参数,计算摄像头的目标调整角度,被进一步配置为:The controller calculates the target adjustment angle of the camera according to the azimuth distance and the shooting parameters of the camera, and is further configured as:
    如果所述垂直方向距离大于所述垂直设定阈值,则根据所述垂直方向距离、摄像头垂直视角角度和图像垂直高度,计算摄像头的目标垂直调整角度。If the vertical distance is greater than the vertical set threshold, calculate the target vertical adjustment angle of the camera according to the vertical distance, the vertical viewing angle of the camera, and the vertical height of the image.
  6. 根据权利要求1所述的显示设备,其特征在于,所述控制器被进一步配置为:The display device of claim 1, wherein the controller is further configured to:
    如果所述方位距离未超过方位设定阈值,则获取预设数量帧的指定图像;If the azimuth distance does not exceed the azimuth setting threshold, obtain a specified image of a preset number of frames;
    如果所述预设数量帧的指定图像中人像区域位置不变,则识别所述指定图像中的人像区域位置的尺寸;If the position of the portrait area in the designated image of the preset number of frames does not change, then identifying the size of the position of the portrait area in the designated image;
    如果所述人像区域位置的尺寸小于或等于所述指定图像的预设比例,则将所述指定图像中的人像区域位置在显示器中进行人像聚焦放大显示。If the size of the position of the portrait area is smaller than or equal to the preset ratio of the designated image, the position of the portrait area in the designated image is displayed on the display in a focus on portrait zoom.
  7. 根据权利要求6所述的显示设备,其特征在于,所述控制器在执行所述如果人像区域位置的尺寸小于或等于所述指定图像的预设比例,则将所述指定图像中的人像区域位置在显示器中进行人像聚焦放大显示,被进一步配置为:The display device according to claim 6, wherein the controller is performing the step of converting the portrait area in the designated image if the size of the position of the portrait area is smaller than or equal to a preset ratio of the designated image. The position is displayed in the display for portrait focus magnification, which is further configured to:
    如果所述人像区域位置的尺寸小于或等于所述指定图像的预设比例,则计算所述显示器的宽高比值和人像区域位置的宽高比值;If the size of the portrait area position is less than or equal to the preset ratio of the specified image, calculate the aspect ratio value of the display and the aspect ratio value of the portrait area position;
    如果所述显示器的宽高比值和人像区域位置的宽高比值不一致时,则调整所述人像区域位置的宽高比值,所述人像区域位置的调整后的宽高比值与所述显示器的宽高比值相同;If the aspect ratio value of the display is inconsistent with the aspect ratio value of the portrait area position, the aspect ratio value of the portrait area position is adjusted, and the adjusted aspect ratio value of the portrait area position is the same as that of the display. the same ratio;
    按照所述宽高比值调整后的人像区域位置,确定人像区域位置的目标放大区域;Determine the target enlarged area of the position of the portrait area according to the position of the portrait area adjusted by the aspect ratio;
    将所述目标放大区域对应的人像进行聚焦放大,全屏显示在所述显示器中。Focusing and enlarging the portrait corresponding to the target enlargement area is performed on the display in full screen.
  8. 根据权利要求7所述的显示设备,其特征在于,所述控制器在执行所述如果显示器的宽高比值和人像区域位置的宽高比值不一致时,则调整所述人像区域位置的宽高比值,被进一步配置为:The display device according to claim 7, wherein the controller adjusts the aspect ratio value of the portrait area position if the aspect ratio value of the display is inconsistent with the aspect ratio value of the portrait area position. , is further configured as:
    如果所述人像区域位置的宽高比值大于所述显示器的宽高比值,则调整所述人像区域位置的高度值,所述人像区域位置的原宽度值与调整后的高度值的宽高比值与所述显示器的宽高比值相同;If the aspect ratio value of the portrait area position is greater than the aspect ratio value of the display, the height value of the portrait area position is adjusted, and the aspect ratio value of the original width value of the portrait area position and the adjusted height value is the same as the height value. The displays have the same aspect ratio;
    如果所述人像区域位置的宽高比值小于所述显示器的宽高比值,则调整所述人像区域位置的宽度值,所述人像区域位置的调整后的宽度值与原高度值的宽高比值与所述显示器的宽高比值相同。If the aspect ratio value of the portrait area position is smaller than the aspect ratio value of the display, adjust the width value of the portrait area position, and the aspect ratio value of the adjusted width value of the portrait area position and the original height value is the same as The displays have the same aspect ratio.
  9. 根据权利要求7所述的显示设备,其特征在于,所述控制器在执行所述将目标放大区域对应的人像进行聚焦放大,全屏显示在所述显示器中,被进一步配置为:The display device according to claim 7, wherein the controller is further configured to:
    获取所述目标放大区域的中心点坐标;obtaining the coordinates of the center point of the target zoom area;
    计算所述中心点坐标与所述目标放大区域的任一条边界的第一距离,以及,所述中心点坐标与所述显示器的任一条边界的第二距离,所述目标放大区域的任一条边界与所述显示器的任一条边界位置相对应;Calculate the first distance between the coordinates of the center point and any border of the target zoom area, and the second distance between the center point coordinates and any border of the display, and any border of the target zoom area corresponding to any border position of the display;
    如果所述第二距离与所述第一距离的距离差小于零,则按照所述距离差调整所述目标放大区域的位置;If the distance difference between the second distance and the first distance is less than zero, adjusting the position of the target enlarged area according to the distance difference;
    将位置调整后的目标放大区域对应的人像进行聚焦放大,全屏显示在所述显示器中。Focusing and enlarging the portrait corresponding to the target enlargement area after the position adjustment is performed, and displaying it on the display in full screen.
  10. 根据权利要求1所述的显示设备,其特征在于,所述控制器在执行所述计算摄像头的目标调整角度之前,被进一步配置为:The display device according to claim 1, wherein before the controller performs the calculation of the target adjustment angle of the camera, it is further configured to:
    判断所述指定图像是否进行人像聚焦放大显示操作;judging whether the specified image is subjected to a portrait focus zoom display operation;
    如果所述指定图像未进行人像聚焦放大显示操作,则执行所述计算摄像头的目标调整角度的步骤;If the designated image has not been subjected to a portrait focus zoom-in display operation, the step of calculating the target adjustment angle of the camera is performed;
    如果所述指定图像已进行人像聚焦放大显示操作,则恢复所述指定图像的显示,以及,执行所述计算摄像头的目标调整角度的步骤。If the designated image has been subjected to an operation of focusing and zooming in on the portrait, the display of the designated image is resumed, and the step of calculating the target adjustment angle of the camera is performed.
  11. 根据权利要求1所述的显示设备,其特征在于,所述控制器在执行所述基于摄像头的目标调整角度,调整所述摄像头的拍摄角度,被进一步配置为:The display device according to claim 1, wherein, when the controller performs the camera-based target angle adjustment to adjust the shooting angle of the camera, the controller is further configured to:
    根据所述摄像头的目标调整角度,确定摄像头的目标转动速度和目标调整方向;Determine the target rotation speed and target adjustment direction of the camera according to the target adjustment angle of the camera;
    按照所述目标调整角度、目标调整方向和目标转动速度,调整所述摄像头的拍摄角度。The shooting angle of the camera is adjusted according to the target adjustment angle, the target adjustment direction and the target rotation speed.
  12. 根据权利要求11所述的显示设备,其特征在于,所述控制器在执行所述根据摄像头的目标调整角度,确定摄像头的目标转动速度,被进一步配置为:The display device according to claim 11, wherein the controller is further configured to:
    如果所述摄像头的目标调整角度大于或等于最大转速逻辑值,则将所述最大转速逻辑值作为摄像头的目标转动速度;If the target adjustment angle of the camera is greater than or equal to the maximum rotational speed logic value, the maximum rotational speed logic value is used as the target rotational speed of the camera;
    如果所述摄像头的目标调整角度小于或等于最小转速逻辑值,则将所述最小转速逻辑值作为摄像头的目标转动速度;If the target adjustment angle of the camera is less than or equal to the minimum rotational speed logic value, the minimum rotational speed logic value is used as the target rotational speed of the camera;
    如果所述摄像头的目标调整角度位于所述最大转速逻辑值和最小转速逻辑值之间,则将所述目标调整角度的数值作为摄像头的目标转动速度。If the target adjustment angle of the camera is located between the maximum rotation speed logic value and the minimum rotation speed logic value, the value of the target adjustment angle is used as the target rotation speed of the camera.
  13. 根据权利要求1所述的显示设备,其特征在于,还包括声音采集器,所述声音采集器被配置为采集人物声源信息,所述人物声源信息是指人物通过语音与显示设备交互时产生的声音信息;The display device according to claim 1, further comprising a sound collector, wherein the sound collector is configured to collect sound source information of a person, and the sound source information of the person refers to when the person interacts with the display device through voice generated sound information;
    所述控制器在执行所述获取摄像头的拍摄参数和采集的人物位于摄像头拍摄区域内的指定图像,被进一步配置为:The controller is further configured to:
    获取所述声音采集器采集的人物声源信息和所述摄像头的当前拍摄角度;Obtain the character sound source information collected by the sound collector and the current shooting angle of the camera;
    对所述人物声源信息进行声源识别,确定声源角度信息,所述声源角度信息用于表征人物在语音时所处位置的方位角度;performing sound source identification on the character sound source information, and determining sound source angle information, where the sound source angle information is used to represent the azimuth angle of the character's position when speaking;
    基于所述摄像头的当前拍摄角度和声源角度信息,确定摄像头的目标转动方向和目标转动角度;Determine the target rotation direction and target rotation angle of the camera based on the current shooting angle and sound source angle information of the camera;
    按照所述目标转动方向和目标转动角度,调整所述摄像头的拍摄角度,所述调整拍摄角度后的摄像头的拍摄区域正对人物语音时的所处位置;According to the target rotation direction and the target rotation angle, adjust the shooting angle of the camera, and the position where the shooting area of the camera after the adjustment of the shooting angle is facing the voice of the character;
    获取调整拍摄角度后的摄像头的拍摄参数和采集的人物位于摄像头拍摄区域内的指定图像。Obtain the shooting parameters of the camera after adjusting the shooting angle and the designated image of the captured person located in the shooting area of the camera.
  14. 根据权利要求13所述的显示设备,其特征在于,所述控制器在执行所述基于摄像头的当前拍摄角度和声源角度信息,确定摄像头的目标转动方向和目标转动角度,被进一步配置为:The display device according to claim 13, wherein the controller is further configured to: determine the target rotation direction and target rotation angle of the camera based on the current shooting angle and sound source angle information of the camera.
    将所述声源角度信息转换为摄像头的坐标角度;Convert the sound source angle information into the coordinate angle of the camera;
    计算所述摄像头的坐标角度和摄像头的当前拍摄角度的角度差值,将所述角度差值作为所述摄像头的目标转动角度;Calculate the angle difference between the coordinate angle of the camera and the current shooting angle of the camera, and use the angle difference as the target rotation angle of the camera;
    根据所述角度差值,确定摄像头的目标转动方向。According to the angle difference, the target rotation direction of the camera is determined.
  15. 根据权利要求14所述的显示设备,其特征在于,所述控制器在执行所述将声源角度信息转换为摄像头的坐标角度,被进一步配置为:The display device according to claim 14, wherein the controller is further configured to:
    获取所述人物在语音时的声源角度范围和摄像头转动时的预设角度范围;Obtain the sound source angle range of the character when speaking and the preset angle range when the camera is rotated;
    计算所述声源角度范围与所述预设角度范围之间的角度差值,将所述角度差值的半值作为转换角度;Calculate the angle difference between the sound source angle range and the preset angle range, and use the half value of the angle difference as the conversion angle;
    计算所述声源角度信息对应的角度与所述转换角度的角度差,将所述角度差作为摄像头的坐标角度。Calculate the angle difference between the angle corresponding to the sound source angle information and the conversion angle, and use the angle difference as the coordinate angle of the camera.
  16. 根据权利要求14所述的显示设备,其特征在于,所述控制器在执行所述根据角度差值,确定摄像头的目标转动方向,被进一步配置为:The display device according to claim 14, wherein the controller is further configured to: determine the target rotation direction of the camera according to the angle difference.
    如果所述角度差值为正值,则确定摄像头的目标转动方向为向右转动;If the angle difference is a positive value, it is determined that the target rotation direction of the camera is to rotate to the right;
    如果所述角度差值为负值,则确定摄像头的目标转动方向为向左转动。If the angle difference is a negative value, it is determined that the target rotation direction of the camera is to rotate to the left.
  17. 一种摄像头的控制方法,其特征在于,所述方法包括:A method for controlling a camera, wherein the method comprises:
    获取所述摄像头的拍摄参数和采集的人物位于摄像头拍摄区域内的指定图像;Acquiring the shooting parameters of the camera and the designated images of the collected characters located in the shooting area of the camera;
    对所述指定图像进行识别处理,得到所述人物对应的人像区域位置,所述人像区域位置是指包括人物头部图像的区域;Performing identification processing on the designated image to obtain the position of the portrait region corresponding to the person, where the position of the portrait region refers to the region including the head image of the person;
    计算所述人像区域位置的区域中心与所述指定图像的图像中心的方位距离,所述方位距离用于标识水平方向距离和垂直方向距离;Calculate the azimuth distance between the area center of the portrait area position and the image center of the designated image, and the azimuth distance is used to identify the horizontal direction distance and the vertical direction distance;
    如果所述方位距离超过方位设定阈值,则根据所述方位距离和摄像头的拍摄参数,计算摄像头的目标调整角度;If the azimuth distance exceeds the azimuth setting threshold, calculate the target adjustment angle of the camera according to the azimuth distance and the shooting parameters of the camera;
    基于所述摄像头的目标调整角度,调整所述摄像头的拍摄角度,以使所述人物的人像位于摄像头采集的指定图像的中心区域。Based on the target adjustment angle of the camera, the shooting angle of the camera is adjusted so that the portrait of the person is located in the center area of the designated image captured by the camera.
PCT/CN2021/093589 2020-07-01 2021-05-13 Camera control method and display device WO2022001407A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010628749.6 2020-07-01
CN202010628749.6A CN111669508A (en) 2020-07-01 2020-07-01 Camera control method and display device

Publications (1)

Publication Number Publication Date
WO2022001407A1 true WO2022001407A1 (en) 2022-01-06

Family

ID=72391139

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/093589 WO2022001407A1 (en) 2020-07-01 2021-05-13 Camera control method and display device

Country Status (2)

Country Link
CN (1) CN111669508A (en)
WO (1) WO2022001407A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115022661A (en) * 2022-06-02 2022-09-06 壹加艺术(武汉)文化有限公司 Video live broadcast environment monitoring, analyzing, regulating and controlling method and device and computer storage medium
CN116866720A (en) * 2023-09-04 2023-10-10 国网山东省电力公司东营供电公司 Camera angle self-adaptive regulation and control method, system and terminal based on sound source localization

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111669508A (en) * 2020-07-01 2020-09-15 海信视像科技股份有限公司 Camera control method and display device
CN112333391A (en) * 2020-11-03 2021-02-05 深圳创维-Rgb电子有限公司 Method and device for automatically tracking portrait based on sound, intelligent terminal and medium
CN112700568B (en) * 2020-12-28 2023-04-18 科大讯飞股份有限公司 Identity authentication method, equipment and computer readable storage medium
CN114845037B (en) * 2021-02-01 2024-02-13 浙江宇视科技有限公司 PTZ camera calibration method, device, electronic equipment and storage medium
CN113301367A (en) * 2021-03-23 2021-08-24 阿里巴巴新加坡控股有限公司 Audio and video processing method, device and system and storage medium
CN113099308B (en) * 2021-03-31 2023-10-27 聚好看科技股份有限公司 Content display method, display equipment and image collector
CN113141518B (en) * 2021-04-20 2022-09-06 北京安博盛赢教育科技有限责任公司 Control method and control device for video frame images in live classroom
CN113382222B (en) * 2021-05-27 2023-03-31 深圳市瑞立视多媒体科技有限公司 Display method based on holographic sand table in user moving process
CN113573021A (en) * 2021-07-26 2021-10-29 嘉应学院 Method for monitoring surrounding conditions of orchard transport vehicle
CN114040109A (en) * 2021-11-23 2022-02-11 慧之安信息技术股份有限公司 Intelligent instrument transformation equipment based on image recognition
CN114815257A (en) * 2022-04-25 2022-07-29 歌尔股份有限公司 XR glasses and camera adjusting method, system, equipment and medium
CN115883970B (en) * 2022-12-02 2024-04-05 浙江省广播电视工程公司 Unmanned management system of broadcast television shooting and recording equipment
CN116980744B (en) * 2023-09-25 2024-01-30 深圳市美高电子设备有限公司 Feature-based camera tracking method and device, electronic equipment and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040100563A1 (en) * 2002-11-27 2004-05-27 Sezai Sablak Video tracking system and method
US20110019066A1 (en) * 2009-07-22 2011-01-27 Yoshijiro Takano Af frame auto-tracking system
CN105049709A (en) * 2015-06-30 2015-11-11 广东欧珀移动通信有限公司 Large-view angle camera control method and user terminal
CN109977770A (en) * 2019-02-21 2019-07-05 安克创新科技股份有限公司 A kind of auto-tracking shooting method, apparatus, system and storage medium
CN111669508A (en) * 2020-07-01 2020-09-15 海信视像科技股份有限公司 Camera control method and display device

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101877764B (en) * 2009-04-29 2012-05-30 鸿富锦精密工业(深圳)有限公司 Camera system and method for carrying out assisted composition by utilizing same
CN110086992A (en) * 2019-04-29 2019-08-02 努比亚技术有限公司 Filming control method, mobile terminal and the computer storage medium of mobile terminal

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040100563A1 (en) * 2002-11-27 2004-05-27 Sezai Sablak Video tracking system and method
US20110019066A1 (en) * 2009-07-22 2011-01-27 Yoshijiro Takano Af frame auto-tracking system
CN105049709A (en) * 2015-06-30 2015-11-11 广东欧珀移动通信有限公司 Large-view angle camera control method and user terminal
CN109977770A (en) * 2019-02-21 2019-07-05 安克创新科技股份有限公司 A kind of auto-tracking shooting method, apparatus, system and storage medium
CN111669508A (en) * 2020-07-01 2020-09-15 海信视像科技股份有限公司 Camera control method and display device

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115022661A (en) * 2022-06-02 2022-09-06 壹加艺术(武汉)文化有限公司 Video live broadcast environment monitoring, analyzing, regulating and controlling method and device and computer storage medium
CN116866720A (en) * 2023-09-04 2023-10-10 国网山东省电力公司东营供电公司 Camera angle self-adaptive regulation and control method, system and terminal based on sound source localization
CN116866720B (en) * 2023-09-04 2023-11-28 国网山东省电力公司东营供电公司 Camera angle self-adaptive regulation and control method, system and terminal based on sound source localization

Also Published As

Publication number Publication date
CN111669508A (en) 2020-09-15

Similar Documents

Publication Publication Date Title
WO2022001407A1 (en) Camera control method and display device
WO2022001406A1 (en) Display method and display device
AU2013276984B2 (en) Display apparatus and method for video calling thereof
US7990421B2 (en) Arrangement and method relating to an image recording device
US11301051B2 (en) Using natural movements of a hand-held device to manipulate digital content
WO2014034556A1 (en) Image processing apparatus and image display apparatus
CN112866772B (en) Display device and sound image character positioning and tracking method
JP2017525024A (en) Architecture for managing input data
WO2022037535A1 (en) Display device and camera tracking method
CN112672062B (en) Display device and portrait positioning method
CN111708383A (en) Method for adjusting shooting angle of camera and display device
WO2022100262A1 (en) Display device, human body posture detection method, and application
KR20220005087A (en) Filming method and terminal
CN117918057A (en) Display device and device control method
CN113473024A (en) Display device, holder camera and camera control method
US10764535B1 (en) Facial tracking during video calls using remote control input
US11087435B1 (en) Adaptive dewarping of wide angle video frames
WO2022037215A1 (en) Camera, display device and camera control method
WO2021218473A1 (en) Display method and display device
US11232796B2 (en) Voice activity detection using audio and visual analysis
WO2020238913A1 (en) Video recording method and terminal
WO2022037229A1 (en) Human image positioning methods and display devices
US11720245B1 (en) User interface information enhancement based on user distance
CN113587812B (en) Display equipment, measuring method and device
CN114647983A (en) Display device and distance detection method based on portrait

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21834311

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21834311

Country of ref document: EP

Kind code of ref document: A1