WO2020001015A1 - 场景操控的方法、装置及电子设备 - Google Patents

场景操控的方法、装置及电子设备 Download PDF

Info

Publication number
WO2020001015A1
WO2020001015A1 PCT/CN2019/073076 CN2019073076W WO2020001015A1 WO 2020001015 A1 WO2020001015 A1 WO 2020001015A1 CN 2019073076 W CN2019073076 W CN 2019073076W WO 2020001015 A1 WO2020001015 A1 WO 2020001015A1
Authority
WO
WIPO (PCT)
Prior art keywords
scene
action
trigger object
area
present disclosure
Prior art date
Application number
PCT/CN2019/073076
Other languages
English (en)
French (fr)
Inventor
李旭刚
冯宇飞
柳杨光
Original Assignee
北京微播视界科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京微播视界科技有限公司 filed Critical 北京微播视界科技有限公司
Priority to GB2100223.3A priority Critical patent/GB2590207B/en
Priority to JP2020571800A priority patent/JP7372945B2/ja
Priority to US16/769,368 priority patent/US11755119B2/en
Publication of WO2020001015A1 publication Critical patent/WO2020001015A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72448User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions
    • H04M1/72454User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions according to context-related or environment-related conditions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/28Recognition of hand or arm movements, e.g. recognition of deaf sign language

Definitions

  • the present disclosure relates to the field of image processing, and in particular, to a method, device, and electronic device for scene manipulation.
  • gestures As a natural and intuitive way of communication, are an important part of human-computer interaction.
  • science and technology and the increasing popularity of computer vision, people have higher and higher requirements for the naturalness of human-computer interaction.
  • the traditional mouse-and-keyboard-based interaction methods have shown their limitations and new human-computer interaction methods. Become a research hotspot.
  • gestures are an efficient way for human-computer interaction and device control. Vision-based gesture recognition is a challenging research topic in the fields of human-computer interaction and pattern recognition.
  • 3D depth-based camera 3D scanning equipment is large in size, high in hardware cost, required higher computing power, and difficult to integrate and apply to popular smart terminals.
  • controlling a scene needs to be based on a specific complex device.
  • these complex devices are often not installed in the smart phone.
  • the control of the scene often involves the movement of the scene perspective. , Interaction with objects in the scene, but can not control the shape of the scene or change the shape of the scene's objects.
  • embodiments of the present disclosure provide a method, an apparatus, and an electronic device for scene manipulation, which at least partially solve the problems in the prior art.
  • an embodiment of the present disclosure provides a scene manipulation method, including:
  • the identification trigger object includes:
  • the acquiring characteristic information of the trigger object is specifically:
  • the determining the action of the triggering object includes:
  • the action of the triggering object is determined based on the first feature information and the second feature information.
  • the determining the action of the triggering object includes:
  • the action of the triggering object is determined based on the area.
  • the determining the action of the triggering object includes:
  • a motion of the triggering object is determined.
  • obtaining the area of the trigger object includes: setting a minimum rule box so that the trigger object is completely contained in the minimum rule box, calculating the area of the minimum rule box, and obtaining the trigger object. area.
  • the trigger object is a human hand.
  • the actions include pinching, rotating, moving near and far, or changing gestures.
  • an embodiment of the present disclosure further provides a scene manipulation device, including:
  • Display module display the first form of the scene
  • Identification module used to identify the trigger object
  • a judging module for judging the action of the triggering object
  • a control module switching the first form of the scene to the second form of the scene based on the action, and the first form of the scene is associated with the second form of the scene.
  • the identification module includes:
  • a feature information acquisition module used to obtain feature information of the trigger object
  • a comparison module used to compare the feature information with standard feature information
  • Trigger object judgment module used to identify whether it is a trigger object according to the comparison result.
  • the acquiring characteristic information of the trigger object is specifically:
  • the judgment module includes:
  • a first characteristic information acquisition module configured to acquire first characteristic information of a triggering object
  • a second characteristic information acquisition module configured to acquire second characteristic information of a triggering object
  • the first action judgment module is configured to judge an action of a triggering object based on the first feature information and the second feature information.
  • the judgment module includes:
  • Area acquisition module used to obtain the area of the trigger object.
  • the judgment module includes:
  • a first area acquisition module for acquiring a first area of a trigger object;
  • a second area acquisition module for acquiring a second area of a trigger object;
  • the second action judgment module is configured to judge the action of the triggering object based on a comparison result between the first area and the second area.
  • the area acquisition module includes:
  • Rule box setting module set the minimum rule box so that the trigger object is completely contained in the minimum rule box
  • Area calculation module Calculate the area of the smallest regular box to get the area of the trigger object.
  • the trigger object is a human hand.
  • the actions include pinching, rotating, moving near and far, or changing gestures.
  • an embodiment of the present disclosure further provides an electronic device including: at least one processor; and,
  • a memory connected in communication with the at least one processor; wherein,
  • the memory stores instructions that can be executed by the at least one processor, and the instructions are executed by the at least one processor, so that the at least one processor can perform any of the scene manipulations described in the first aspect. method.
  • an embodiment of the present disclosure further provides a non-transitory computer-readable storage medium, where the non-transitory computer-readable storage medium stores computer instructions, and the computer instructions are used to cause a computer to execute any one of the claims of the first aspect.
  • the method for scene control is not limited to:
  • the method, device, electronic device and non-transitory computer-readable storage medium for scene manipulation provided by the embodiments of the present disclosure, wherein the method for scene manipulation: calling corresponding set scene information according to the action of a triggering object, thereby achieving Change the scene or control an element in the scene, and associate the action of the triggering object with the scene. Without the need to upgrade the existing electronic equipment, you can control the scene in the electronic equipment, thereby achieving Cost reduction purposes.
  • FIG. 1 is a flowchart of a scene manipulation method according to an embodiment of the present disclosure
  • FIG. 2 is a flowchart of identifying a triggering object according to an embodiment of the present disclosure
  • FIG. 3 is a flowchart of determining an action of a triggering object according to an embodiment of the present disclosure
  • FIG. 4 is a schematic diagram of an action of triggering an object based on area judgment according to an embodiment of the present disclosure
  • FIG. 5 is a schematic diagram of setting a minimum rule box according to an embodiment of the present disclosure.
  • FIG. 6 is a schematic block diagram of a scene manipulation apparatus according to an embodiment of the present disclosure.
  • FIG. 7 is a schematic block diagram of an electronic device according to an embodiment of the present disclosure.
  • FIG. 8 is a schematic diagram of a computer-readable storage medium according to an embodiment of the present disclosure.
  • FIG. 9 is a schematic block diagram of a terminal according to an embodiment of the present disclosure.
  • an embodiment of the present disclosure provides a method for scene manipulation.
  • the scene manipulation method includes the following steps:
  • S101 Display the first form of the scene.
  • the first form of the scene that is, the form before the scene is switched, for example, in a specific application scenario
  • the first form of the scene can be a 2D scene, which can be displayed by a mobile phone.
  • the first form of the scene can also be a 3D scene, that is, the form before the scene is switched, and the trigger object and the recognition of the first form of the scene are completed. Determine the action of the triggering object.
  • step S102 After the trigger object is identified in step S102, it is necessary to determine whether the trigger object has performed a corresponding action, determine whether the action of the trigger object is compared with the saved action data, and determine that the action is specifically the action in the action data.
  • S104 Switch the first form of the scene to the second form of the scene based on the action, and the first form of the scene is associated with the second form of the scene.
  • step S103 After determining the corresponding action in step S103, it is necessary to call the scene information corresponding to the action and display the called scene information, or use the called scene information to combine with the existing scene information to form new scene information. Or use the recalled scene information to replace an element in the existing scene information, or directly replace the existing scene information with the recalled scene information.
  • a scene is a picture composed of various elements. You can manipulate the scene, you can switch between different scenes, or you can manipulate an element in the same scene, such as controlling the distance of an object (such as a ball) in the scene. motion.
  • the trigger object is a hand
  • the actions of the trigger object are mainly various gestures, such as Zhanghe, rotation, distance movement, or gesture change.
  • the scene change in the mobile phone is controlled through gestures.
  • the mobile phone displays the real-time picture taken by the camera and is equipped with background music.
  • the real-time picture with background music is the first form of the scene, such as a scene with raining 2D maps.
  • the human hand is first identified, and then it is judged whether the human hand has made an open movement, that is, from a fist to an open palm. If it is determined that the human hand has made an opening action, then the corresponding scene information is called, such as the rain information in the 3D particles, and the rain information is combined with the currently displayed shooting real-time picture, so as to be on the mobile phone Show the scene of 3D particles of raindrops, and the background music can be selected according to the original settings.
  • the scene in which 3D particles of raindrops are displayed on the mobile phone is the second form of the scene.
  • the association between the second form of the scene and the first form of the scene is preset.
  • the action is to call the corresponding information to transform the scene from the first form to the second form.
  • the mobile phone displays a blue sky and white clouds.
  • the mobile phone detects that the image captured by the camera contains human hands, that is, after the mobile phone recognizes the human hand, it is next necessary to determine whether the human hand moves forward or backward relative to the mobile phone camera. , Such as judging that the human hand has moved back and forth relative to the mobile phone camera, then call the corresponding scene information to make the cloud move according to the relative movement of the hand, that is, when the hand is close to the camera, the cloud is close to the screen, and when the hand is away When you look at the camera, the clouds are far away from the shield.
  • the scene before the white cloud moves is the first form of the scene
  • the scene after the white cloud moves is the second form of the scene.
  • the mobile phone is only used as an example for easy understanding.
  • the disclosure is not limited to mobile phones, and electronic devices with information processing functions, such as tablets and portable computers, can use the technical solutions of the disclosure.
  • the use of the hand as a trigger object is merely an exemplary description, and the trigger object is not limited to the hand.
  • the trigger object may also be a head or a limb, and the action may be a movement such as shaking a head or blinking.
  • identifying the trigger object in step S102 specifically includes:
  • the feature information of an object is information that can characterize the object, such as the contour information and keypoint information of the object image.
  • the techniques for extracting the contour information and keypoint information of an image in the prior art are relatively mature, and are not repeated here.
  • the obtained feature information needs to be compared with the pre-stored feature information.
  • the contour or keypoint information of the image needs to be obtained, and then the acquired contour or keypoint information is compared with the pre-stored contour or keypoint information of the human hand. If the comparison results match, the acquired image is considered to be a human hand, that is, the trigger object is identified.
  • the contour or keypoint information of the acquired image needs to be compared with the pre-stored contour or keypoint information of the human hand one by one. As long as the comparison result is consistent, the acquired image is identified as a human hand image.
  • the acquiring characteristic information of the trigger object is specifically: acquiring a key point on the trigger object.
  • the step S103 of determining the action of the triggering object specifically includes:
  • Step S301 Acquire first feature information of a trigger object.
  • Step S302 Acquire the second characteristic information of the trigger object.
  • Step S303 Determine the action of the triggering object based on the first feature information and the second feature information.
  • This exemplary description uses key points as characteristic information.
  • a human hand for example, in judging a movement of a human hand from a fist to an open motion, it is necessary to first obtain a key point when the human hand is fisting. Then get the key points after the hands are open. Then the key points of the human hand fist are compared with the pre-stored key points of the hand, respectively, so as to determine that the hand is in a state of fist, and then the key points of the opened hand are compared with the pre-stored key points of the hand, so that the hand is in the open position. On state, to determine that the human hand has made an open action.
  • determining the action of the triggering object in step S103 further includes: acquiring an area of the triggering object, and determining the action of the triggering object based on the area.
  • S403 Determine the action of the triggering object based on a comparison result between the first area and the second area.
  • the farther the human hand is from the mobile phone the smaller the area of the human hand when it is phased, and the closer the human hand is to the mobile phone, the more its area is when it is phased. Big. Therefore, according to the calculation of the area of the human hand, the movement of the human hand relative to the mobile phone can be easily determined.
  • the obtaining the area of the triggering object specifically includes the following steps:
  • the rectangle can just wrap the hand, and the area of the hand is replaced by calculating the area of the rectangle, thereby simplifying the calculation amount.
  • a regular rectangle is set.
  • the movement of the elements of the scene can also be associated with the length of the side of the rectangle.
  • the distance traveled is related to control the trajectory of the cloud according to the side length of the rectangle. Since the change in side length is linear, no jump occurs. Therefore, the movement of the clouds will be relatively smooth, and there will be no jumping movement.
  • an embodiment of the present disclosure provides a scene manipulation device, including:
  • Display module 600 displaying a first form of a scene
  • Identification module 601 used to identify a trigger object
  • a judging module 602 configured to judge an action of the triggering object
  • the control module 603 is configured to switch the first form of the scene to the second form of the scene based on the action, and the first form of the scene is associated with the second form of the scene.
  • the identification module 601 includes:
  • Feature information acquisition module 6011 configured to obtain feature information of the trigger object
  • Comparison module 6012 configured to compare the feature information with standard feature information
  • Trigger object judgment module 6013 used to identify whether it is a trigger object according to the comparison result.
  • the acquiring characteristic information of the trigger object is specifically:
  • the determining module 602 includes:
  • a first characteristic information acquisition module 6021 configured to acquire first characteristic information of a trigger object
  • a second characteristic information acquisition module 6022 configured to acquire second characteristic information of the trigger object
  • the first action judgment module 6023 is configured to judge an action of a triggering object based on the first feature information and the second feature information.
  • the determining module 602 includes:
  • Area acquisition module used to obtain the area of the trigger object.
  • the determining module 602 includes:
  • a first area acquisition module configured to acquire a first area of a triggering object
  • a second area acquisition module for acquiring a second area of the triggering object
  • a second action judging module based on a comparison result between the first area and the second area,
  • the area acquisition module includes:
  • Rule box setting module set the minimum rule box so that the trigger object is completely contained in the minimum rule box
  • Area calculation module Calculate the area of the smallest regular box to get the area of the trigger object.
  • the trigger object is a human hand.
  • the actions include pinching, rotating, moving near and far, or changing gestures.
  • FIG. 7 is a hardware block diagram of an electronic device according to an embodiment of the present disclosure. As shown in FIG. 7, the electronic device 70 according to an embodiment of the present disclosure includes a memory 71 and a processor 72.
  • the memory 71 is configured to store non-transitory computer-readable instructions.
  • the memory 71 may include one or more computer program products, and the computer program product may include various forms of computer-readable storage media, such as volatile memory and / or non-volatile memory.
  • the volatile memory may include, for example, a random access memory (RAM) and / or a cache memory.
  • the non-volatile memory may include, for example, a read-only memory (ROM), a hard disk, a flash memory, and the like.
  • the processor 72 may be a central processing unit (CPU) or other form of processing unit having data processing capabilities and / or instruction execution capabilities, and may control other components in the electronic device 70 to perform desired functions.
  • the processor 72 is configured to run the computer-readable instructions stored in the memory 71, so that the electronic device 70 performs all or part of the steps of scene manipulation of the foregoing embodiments of the present disclosure.
  • this embodiment may also include well-known structures such as a communication bus and an interface. These well-known structures should also be included in the protection scope of the present disclosure. within.
  • FIG. 8 is a schematic diagram illustrating a computer-readable storage medium according to an embodiment of the present disclosure.
  • a computer-readable storage medium 80 according to an embodiment of the present disclosure has non-transitory computer-readable instructions 81 stored thereon.
  • the non-transitory computer-readable instructions 81 are executed by a processor, all or part of the steps of scene manipulation of the foregoing embodiments of the present disclosure are performed.
  • the computer-readable storage medium 80 includes, but is not limited to, optical storage media (for example, CD-ROM and DVD), magneto-optical storage media (for example, MO), magnetic storage media (for example, magnetic tape or mobile hard disk), Non-volatile memory rewritable media (for example: memory card) and media with built-in ROM (for example: ROM box).
  • optical storage media for example, CD-ROM and DVD
  • magneto-optical storage media for example, MO
  • magnetic storage media for example, magnetic tape or mobile hard disk
  • Non-volatile memory rewritable media for example: memory card
  • media with built-in ROM for example: ROM box
  • FIG. 9 is a schematic diagram illustrating a hardware structure of a terminal device according to an embodiment of the present disclosure. As shown in FIG. 9, the terminal 90 includes the foregoing embodiment of a scene manipulation device.
  • the terminal device may be implemented in various forms, and the terminal device in the present disclosure may include, but is not limited to, such as a mobile phone, a smart phone, a notebook computer, a digital broadcast receiver, a PDA (personal digital assistant), a PAD (tablet computer), Mobile terminal devices such as PMPs (portable multimedia players), navigation devices, vehicle-mounted terminal devices, vehicle-mounted display terminals, vehicle-mounted electronic rear-view mirrors, and the like, and fixed terminal devices such as digital TVs, desktop computers, and the like.
  • PMPs portable multimedia players
  • navigation devices such as PMPs (portable multimedia players), navigation devices, vehicle-mounted terminal devices, vehicle-mounted display terminals, vehicle-mounted electronic rear-view mirrors, and the like
  • fixed terminal devices such as digital TVs, desktop computers, and the like.
  • the terminal 90 may further include other components.
  • the terminal 90 may include a power supply unit 91, a wireless communication unit 92, an A / V (audio / video) input unit 93, a user input unit 94, a sensing unit 95, an interface unit 96, a controller 97, The output unit 98 and the storage unit 99 and so on.
  • FIG. 9 shows a terminal with various components, but it should be understood that it is not required to implement all the illustrated components, and more or fewer components may be implemented instead.
  • the wireless communication unit 92 allows radio communication between the terminal 90 and a wireless communication system or network.
  • the A / V input unit 93 is used to receive audio or video signals.
  • the user input unit 94 may generate key input data according to a command input by the user to control various operations of the terminal device.
  • the sensing unit 95 detects the current state of the terminal 90, the position of the terminal 90, the presence or absence of a user's touch input to the terminal 90, the orientation of the terminal 90, the acceleration or deceleration movement and direction of the terminal 90, and the like, and generates a signal for controlling the terminal 90's operation command or signal.
  • the interface unit 96 functions as an interface through which at least one external device can be connected to the terminal 90.
  • the output unit 98 is configured to provide an output signal in a visual, audio, and / or tactile manner.
  • the storage unit 99 may store software programs and the like for processing and control operations performed by the controller 97, or may temporarily store data that has been output or is to be output.
  • the storage unit 99 may include at least one type of storage medium.
  • the terminal 90 may cooperate with a network storage device that performs a storage function of the storage unit 99 through a network connection.
  • the controller 97 generally controls the overall operation of the terminal device.
  • the controller 97 may include a multimedia module for reproducing or playing back multimedia data.
  • the controller 97 may perform a pattern recognition process to recognize a handwriting input or a picture drawing input performed on the touch screen as characters or images.
  • the power supply unit 91 receives external power or internal power under the control of the controller 97 and provides appropriate power required to operate each element and component.
  • scene manipulation proposed by the present disclosure may be implemented using computer-readable media, such as computer software, hardware, or any combination thereof.
  • various embodiments of scene manipulation proposed by the present disclosure can be implemented by using application-specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), At least one of a field programmable gate array (FPGA), processor, controller, microcontroller, microprocessor, electronic unit designed to perform the functions described herein is implemented, and in some cases, the present disclosure proposes Various embodiments of scene manipulation can be implemented in the controller 97.
  • ASICs application-specific integrated circuits
  • DSPs digital signal processors
  • DSPDs digital signal processing devices
  • PLDs programmable logic devices
  • FPGA field programmable gate array
  • processor controller
  • microcontroller microcontroller
  • microprocessor electronic unit designed to perform the functions described herein
  • the various embodiments of scene manipulation proposed by the present disclosure may be implemented with a separate software module allowing at least one function or operation to be performed.
  • the software codes may be implemented by a software application (or program) written in any suitable programming language, and the software codes may be stored in the storage unit 99 and executed by the controller 97.
  • relational terms such as first and second are used only to distinguish one entity or operation from another entity or operation, and do not necessarily require or imply that any such relationship exists between these entities or operations.
  • the block diagrams of the devices, devices, devices, and systems involved in this disclosure are only illustrative examples and are not intended to require or imply that they must be connected, arranged, and configured in the manner shown in the block diagrams. As those skilled in the art would realize, these devices, devices, equipment, and systems may be connected, arranged, and configured in any manner. Words such as “including,” “including,” “having,” and the like are open words, meaning “including, but not limited to,” and can be used interchangeably with them.
  • the words “or” and “and” refer to the words “and / or” and are used interchangeably with each other, unless the context clearly indicates otherwise.
  • the term “such as” refers to the phrase “such as, but not limited to,” and is used interchangeably with it.
  • an "or” used in an enumeration of items beginning with “at least one” indicates a separate enumeration such that, for example, an "at least one of A, B, or C” enumeration means A or B or C, or AB or AC or BC, or ABC (ie A and B and C).
  • the word "exemplary” does not mean that the described example is preferred or better than other examples.
  • each component or each step can be disassembled and / or recombined.
  • These decompositions and / or recombinations should be considered as equivalent solutions of the present disclosure.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Psychiatry (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Social Psychology (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Environmental & Geological Engineering (AREA)
  • User Interface Of Digital Computer (AREA)
  • Telephone Function (AREA)

Abstract

本公开实施例公开了一种场景操控的方法、装置及电子设备,涉及场景操控领域。其中,场景操控方法,包括:显示场景的第一形态;识别触发对象;判断所述触发对象的动作;基于所述动作将所述场景的第一形态切换为场景的第二形态,所述场景的第一形态与所述场景的第二形态相关联。依据触发对象的动作,调用相应的设定好的场景信息,从而实现对场景的切换或对场景中某个元素的控制,并将触发对象的动作与场景相关联,不需要对现有的电子设备进行硬件升级,即可实现对电子设备中场景的控制,从而达到降低成本目的。

Description

场景操控的方法、装置及电子设备
本申请要求于2018年6月29日提交中国专利局、申请号为201810699063.9,发明名称为“场景操控的方法、装置及电子设备”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本公开涉及图像处理领域,尤其涉及一种场景操控的方法、装置及电子设备。
背景技术
随着人机交互技术的快速发展,手势作为一种自然而直观的交流方式,是人机交互中的一个重要组成部分。随着科学技术的迅猛发展与计算机视觉的日益普及,人们对人机交互自然性的要求越来越高,传统的基于鼠标、键盘的交互方式显示出自身的局限性,新的人机交互方式成为研究的热点。其次,手势是一种高效的人机交互和设备控制的方式,基于视觉的手势识别是人机交互、模式识别等领域的一个富有挑战性的研究课题。
目前的手势识别系统多采用以下两种:
(1)数据手套或佩戴物:此方式可降低检测和识别算法的复杂性,但佩戴式的操作方式显然难以满足自然的人机交互的需要;
(2)基于3D深度相机:3D扫描设备体积较大,硬件成本较高,所需运算能力更高,难以集成并应用到大众化的智能终端上。
发明内容
现有技术中对场景进行控制,需要基于特定的复杂器件。而在实际应用中,如在使用智能手机进行直播时,为了智能手机的便携性,智能手机中往往没有安装这些复杂的器件,另外在现有技术中,对场景的控制往往涉及场景视角的移动,与场景中物体的交互,但是并不能控制场景的形态或者无法改变场景的对象的形态。
有鉴于此,本公开实施例提供了一种场景操控的方法、装置及电子设备,至少部分的解决现有技术中存在的问题。
第一方面,本公开实施例提供了一种场景操控的方法,包括:
显示场景的第一形态;
识别触发对象;
判断所述触发对象的动作;
基于所述动作将所述场景的第一形态切换为场景的第二形态,所述场景的第一形态与所述场景的第二形态相关联。
作为本公开实施例的一种具体实现方式,所述识别触发对象,包括:
获取所述触发对象的特征信息;
将所述特征信息与标准特征信息相比对;
根据比对结果识别是否为触发对象。
作为本公开实施例的一种具体实现方式,所述获取所述触发对象的特征信息,具体为:
获取所述触发对象上的关键点。
作为本公开实施例的一种具体实现方式,所述判断所述触发对象的动作,包括:
获取触发对象的第一特征信息;
获取触发对象的第二特征信息;
基于所述第一特征信息和第二特征信息判断触发对象的动作。
作为本公开实施例的一种具体实现方式,所述判断所述触发对象的动作,包括:
获取触发对象的面积;
基于所述面积判断所述触发对象的动作。
作为本公开实施例的一种具体实现方式,所述判断所述触发对象的动作,包括:
获取触发对象的第一面积;
获取触发对象的第二面积;
基于所述第一面积与所述第二面积的比对结果,判断触发对象的动作。
作为本公开实施例的一种具体实现方式,所述获取触发对象的面积,包括:设定最小规则框,使得触发对象完全包含在最小规则框内,计算最小规则框的面积,得到触发对象的面积。
作为本公开实施例的一种具体实现方式,所述触发对象为人手。
作为本公开实施例的一种具体实现方式,所述动作,包括张合、旋转、远近移动或手势变化。
第二方面,本公开实施例还提供了一种场景操控的装置,包括:
显示模块:显示场景的第一形态;
识别模块:用于识别触发对象;
判断模块:用于判断所述触发对象的动作;
控制模块:基于所述动作将所述场景的第一形态切换为场景的第二形态,所述场景的第一形态与所述场景的第二形态相关联。
作为本公开实施例的一种具体实现方式,所述识别模块,包括:
特征信息获取模块:用于获取所述触发对象的特征信息;比对模块:用于将所述特征信息与标准特征信息相比对;
触发对象判断模块:用于根据比对结果识别是否为触发对象。
作为本公开实施例的一种具体实现方式,所述获取所述触发对象的特征信息,具体为:
获取所述触发对象上的关键点。
作为本公开实施例的一种具体实现方式,所述判断模块,包括:
第一特征信息获取模块:用于获取触发对象的第一特征信息;
第二特征信息获取模块:用于获取触发对象的第二特征信息;
第一动作判断模块:用于基于所述第一特征信息和第二特征信息判断触发对象的动作。
作为本公开实施例的一种具体实现方式,所述判断模块,包括:
面积获取模块:用于获取触发对象的面积。
作为本公开实施例的一种具体实现方式,所述判断模块,包括:
第一面积获取模块:用于获取触发对象的第一面积;第二面积获取模块:用于获取触发对象的第二面积;
第二动作判断模块:用于基于所述第一面积与所述第二面积的比对结果,判断触发对象的动作。
作为本公开实施例的一种具体实现方式,所述面积获取模块,包括:
规则框设定模块:设定最小规则框,使得触发对象完全包含在最小规则框内;
面积计算模块:计算最小规则框的面积,得到触发对象的面积。
作为本公开实施例的一种具体实现方式,所述触发对象为人手。
作为本公开实施例的一种具体实现方式,所述动作,包括张合、旋转、远近移动或手势变化。
第三方面,本公开实施例还提供了一种电子设备,该电子设备包括:至少一个处理器;以及,
与所述至少一个处理器通信连接的存储器;其中,
所述存储器存储有能被所述至少一个处理器执行的指令,所述指令被所述至少一个处理器执行,以使所述至少一个处理器能够执行第一方面任一所述的场景操控的方法。
第四方面,本公开实施例还提供了一种非暂态计算机可读存储介质,该非暂态计算机可读存储介质存储计算机指令,该计算机指令用于使计算机执行权利要求第一方面任一所述的场景操控的方法。
本公开实施例提供的场景操控的方法、装置、电子设备及非暂态计算机可读存储介质,其中该场景操控的方法:依据触发对象的动作,调用相应的设定好的场景信息,从而实现对场景的切换或对场景中某个元素的控制,并将触发对象的动作与场景相关联,不需要对现有的电子设备进行硬件升级,即可实现对电子设备中场景的控制,从而达到降低成本目的。
上述说明仅是本公开技术方案的概述,为了能更清楚了解本公开的技术手段,而可依照说明书的内容予以实施,并且为让本公开的上述和其他目的、特征和优点能够更明显易懂,以下特举较佳实施例,并配合附图,详细说明如下。
附图说明
为了更清楚地说明本公开实施例的技术方案,下面将对实施例中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本公开的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其它的附图。
图1为本公开实施例提供的一种场景操控的方法的流程图;
图2为本公开实施例提供的识别触发对象的流程图;
图3为本公开实施例提供的判断触发对象的动作的流程图;
图4为本公开实施例提供的基于面积判断触发对象的动作的示意图;
图5为本公开实施例提供的设置最小规则框的示意图;
图6为本公开实施例提供的一种场景操控的装置的原理框图;
图7为本公开实施例提供的一种电子设备的原理框图;
图8为本公开实施例提供的一种计算机可读存储介质的示意图;
图9为本公开实施例提供的一种终端的原理框图。
具体实施方式
下面结合附图对本公开实施例进行详细描述。
应当明确,以下通过特定的具体实例说明本公开的实施方式,本领域技术人员可由本说明书所揭露的内容轻易地了解本公开的其他优点与功效。显然,所描述的实施例仅仅是本公开一部分实施例,而不是全部的实施例。本公开还可以通过另外不同的具体实施方式加以实施或应用,本说明书中的各项细节也可以基于不同观点与应用,在没有背离本公开的精神下进行各种修饰或改变。
需说明的是,在不冲突的情况下,以下实施例及实施例中的特征可以相互组合。
基于本公开中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本公开保护的范围。
需要说明的是,下文描述在所附权利要求书的范围内的实施例的各种方面。
应显而易见,本文中所描述的方面可体现于广泛多种形式中,且本文中所描述的任何特定结构及/或功能仅为说明性的。基于本公开,所属领域的技术人员应了解,本文中所描述的一个方面可与任何其它方面独立地实施,且可以各种方式组合这些方面中的两者或两者以上。举例来说,可使用本文中所阐述的任何数目个方面来实施设备及/或实践方法。另外,可使用除了本文中所阐述的方面中的一或多者之外的其它结构及/或功能性实施此设备及/或实践此方法。
还需要说明的是,以下实施例中所提供的图示仅以示意方式说明本公开的基本构相,图式中仅显示与本公开中有关的组件而非按照实际实施时的组件数目、形状及尺寸绘制,其实际实施时各组件的型态、数量及比 例可为一种随意的改变,且其组件布局型态也可能更为复杂。
另外,在以下描述中,提供具体细节是为了便于透彻理解实例。然而,所属领域的技术人员将理解,可在没有这些特定细节的情况下实践所述方面。
如图1所示,本公开实施例提供一种场景操控的方法。该场景操控的方法,包括如下步骤:
S101:显示场景的第一形态。
场景的第一形态,即场景切换前的形态,比如在一个具体的应用场景中,
场景的第一形态可以是2D场景,通过手机显示出来,场景的第一形态也可以是3D场景,即在场景切换前的形态,并在显示场景的第一形态的状态下完成识别触发对象和判断所述触发对象的动作。
S102:识别触发对象。
对场景进行操控,首先要判断对象是否为触发对象,如不是触发对象则不响应,只有识别出是触发对象时,才需要进一步对触发对象是否动作,以及做出了什么样的动作进行判断,然后根据不同的动作调取不同的场景信息。
S103:判断所述触发对象的动作。
在步骤S102识别出触发对象后,需要判断该触发对象是否做出了相应的动作,判断触发对象的动作是否与保存的动作数据相比对,并判断出具体是动作数据中的那种动作。
S104:基于所述动作将所述场景的第一形态切换为场景的第二形态,所述场景的第一形态与所述场景的第二形态相关联。
在步骤S103判断出相应动作后,就需要调用与动作相对应的场景信息,并将调用的场景信息显示出来,或使用调用的场景信息与现有的场景信息进行组合从而组成新的场景信息,或使用调用的场景信息替代现有的场景信息中的某个元素,或直接用调用的场景信息替换掉现有的场景信息。
场景即有各种元素构成的画面,对场景进行操控,可以在不同场景间的切换,也可以是对同一场景中某个元素的操控,如控制场景中的某个物 体(如球)远近的运动。
在一个具体的应用场景中,触发对象为手,而触发对象的动作主要为各种手势,如张合、旋转、远近移动或手势变化。具体为通过手势来来控制手机中场景的变化。
开始时,手机显示的是摄像头拍摄的实时画面,且配有背景音乐,配有背景音乐的实时画面即为场景的第一形态,如具有下雨2D贴图的场景,当摄像头拍摄到张开的人手时,首先识别人手,然后再判断人手是否做出了张开的动作,即从握拳到张开手掌。如判断人手做出了张开动作,那么则调用相应的场景信息,如3D粒子中的下雨的信息,并将下雨的信息与当前显示的拍摄的实时画面结合到一起,从而在手机上显示雨滴的3D粒子的场景,而背景音乐可以根据原有的设定进行选择,可以选择继续播放原来的背景音乐,也可以将背景音乐关闭,或者换一首与下雨场景相适配的音乐,手机上显示雨滴的3D粒子的场景即为场景的第二形态,场景的第二形态与场景的第一形态其关联是预设好的,当在场景的第一形态下检测到人手张开的动作,即调用相应的信息,将场景从第一形态转换为第二形态。
另一个具体的应用场景中,手机显示的场景为蓝天白云,当手机检测到摄像头拍摄的图像包含人手后,即当手机识别出人手后,接下来就要判断人手是否做出相对手机摄像头前后移动的动作,如判断人手做出了相对手机摄像头前后移动的动作,则调用相应的场景信息,使云彩根据手的相对移动而移动,即可以是当手靠近相机时,云彩靠近屏幕,当手远离相机时,云彩则远离屏蔽。也可以是当手靠近相机时,云彩远离屏幕,当手远离相机时,云彩则靠近屏蔽,白云移动前的场景即为场景的第一形态,白云移动后的场景即为场景的第二形态,当在场景的第一形态下检测到人手移动的动作,即调用相应的信息,将场景从第一形态转换为第二形态。
在上两个具体的应用中,手机只是为了方便理解而进行示例性说明使用,
本公开本不仅仅限定于手机,如平板、便携式电脑等具有信息处理功能的电子设备均可以使用本公开的技术方案。且该应用场景中以手为触发对象只是进行示例性的描述,并不将触发对象仅仅限定为手,触发对象也 可以是头或肢体等,而动作也可以是摇头或眨眼等动作。
作为本公开实施例的一种具体实现方式,如图2所示,步骤S102中识别触发对象,具体包括:
S201:获取所述触发对象的特征信息。
对象的特征信息,是能够表征对象的信息,如对象图像的轮廓信息、关键点信息等。现有技术中关于图像的轮廓信息和关键点信息提取的技术比较成熟,在此不再累述。
S202:将所述特征信息与标准特征信息相比对。
在步骤201获取到特征信息后,需要将获取的特征信息与预存的特征信息相对比。
S203:根据比对结果识别是否为触发对象。
如在以人手为例的场景中,当获取到图像后,需要获取图像的轮廓或关键点信息,然后将获取的轮廓或关键点信息,与预存的人手的轮廓或关键点信息相比对,如比对的结果相符,则认为获取到的图像为人手,即识别出触发对象。
而在实际应用中,因用户较多,而人手的轮廓或关键点信息并不相同,因此预存的人手的轮廓或关键点信息有多种,其中有不同大小的人手轮廓或关键点信息。在对比时,需要将获取的图像的轮廓或关键点信息逐一与预存的人手的轮廓或关键点信息相比对,只要有一次对比结果为相符,即将该获取的图像识别为人手图像。
作为本公开实施例的一种具体实现方式,所述获取所述触发对象的特征信息,具体为:获取所述触发对象上的关键点。
作为本公开实施例的一种具体实现方式,如图3所示,步骤S103所述判断所述触发对象的动作,具体包括:
步骤S301:获取触发对象的第一特征信息。
步骤S302:获取触发对象的第二特征信息。
步骤S303:基于所述第一特征信息和第二特征信息判断触发对象的动作。
本示例性说明采用关键点为特征信息,在人手的应用场景中,如在判断人手由握拳到张开的动作中,首先需要获取人手在握拳时的关键点。然 后再获取人手张开后的关键点。然后分别将人手握拳的关键点与预存的人手关键点相比对,从而判断人手是握拳状态,再将人手张开后的关键点与预存的人手关键点相比对,从而判断人手时处于张开状态,从而判断出人手做出了张开的动作。
作为本公开实施例的一种具体实现方式,步骤S103所述判断所述触发对象的动作,还包括:获取触发对象的面积,基于所述面积判断所述触发对象的动作。
而基于所述面积判断所述触发对象的动作,如图4所示,具体为:
S401:获取触发对象的第一面积;
S402:获取触发对象的第二面积;
S403:基于所述第一面积与所述第二面积的比对结果,判断触发对象的动作。
还以上述的手有握拳到张开的动作为例,在获取到人手的图像后,首先获取人手在握拳时的面积,然后再获取人手张开后的面积,然后分别将人手握拳的面积与预存的人手面积相比对,从而判断人手是握拳状态,再将人手张开后的面积与预存的人手面积相比对,从而判断人手时处于张开状态,从而判断出人手做出了张开的动作。而在判断人手相对于手机远近运动时,在摄像头参数设置不变的前提下,人手离手机越远,其在成相时面积越小,人手离手机越近时,其在成相时面积越大。从而根据计算人手的面积大小可以很容易的就判断出人手相对于手机的动作。
因为人手的形状不太规则,如果直接计算人手的面积,会较复杂点,因此作为本公开实施例的一种具体实现方式,所述获取触发对象的面积,具体包括以下步骤:
设定最小规则框,使得触发对象完全包含在最小规则框内,计算最小规则框的面积,得到触发对象的面积。
如图5所示,通过设置一个规则的长方形,该长方形正好能将手包裹住,从而通过计算该长方形的面积来代替手的面积,从而简化的了计算量。
本公开设置规则的长方形,除了用长方形的面积代替人手的面积,还可以将场景的元素运动与长方形的边长相关联,如在上述人手相对于手机 前后移动时,可以通过长方形的边长与人手移动的距离相关联,从而根据长方形的边长控制云彩的运动轨迹。由于边长的变化是线性的,不会发生跃变。因此云彩的运动也会比较平滑,不会出现跳跃式的运动。
如图6所示,本公开实施例提供一种场景操控的装置,包括:
显示模块600:显示场景的第一形态;
识别模块601:用于识别触发对象;
判断模块602:用于判断所述触发对象的动作;
控制模块603:用于基于所述动作将所述场景的第一形态切换为场景的第二形态,所述场景的第一形态与所述场景的第二形态相关联。
作为本公开实施例的一种具体实现方式,所述识别模块601,包括:
特征信息获取模块6011:用于获取所述触发对象的特征信息;
比对模块6012:用于将所述特征信息与标准特征信息相比对;
触发对象判断模块6013:用于根据比对结果识别是否为触发对象。
作为本公开实施例的一种具体实现方式,所述获取所述触发对象的特征信息,具体为:
获取所述触发对象上的关键点。
作为本公开实施例的一种具体实现方式,所述判断模块602,包括:
第一特征信息获取模块6021:用于获取触发对象的第一特征信息;
第二特征信息获取模块6022:用于获取触发对象的第二特征信息;
第一动作判断模块6023:用于基于所述第一特征信息和第二特征信息判断触发对象的动作。
作为本公开实施例的一种具体实现方式,所述判断模块602,包括:
面积获取模块:用于获取触发对象的面积。
作为本公开实施例的一种具体实现方式,所述判断模块602,包括:
第一面积获取模块:用于获取触发对象的第一面积;
第二面积获取模块:用于获取触发对象的第二面积;
第二动作判断模块:用于基于所述第一面积与所述第二面积的比对结果,
判断触发对象的动作。
作为本公开实施例的一种具体实现方式,所述面积获取模块,包括:
规则框设定模块:设定最小规则框,使得触发对象完全包含在最小规则框内;
面积计算模块:计算最小规则框的面积,得到触发对象的面积。
作为本公开实施例的一种具体实现方式,所述触发对象为人手。
作为本公开实施例的一种具体实现方式,所述动作,包括张合、旋转、远近移动或手势变化。
图7是根据本公开的实施例的电子设备的硬件框图。如图7所示,根据本公开实施例的电子设备70包括存储器71和处理器72。
该存储器71用于存储非暂时性计算机可读指令。具体地,存储器71可以包括一个或多个计算机程序产品,该计算机程序产品可以包括各种形式的计算机可读存储介质,例如易失性存储器和/或非易失性存储器。该易失性存储器例如可以包括随机存取存储器(RAM)和/或高速缓冲存储器(cache)等。该非易失性存储器例如可以包括只读存储器(ROM)、硬盘、闪存等。
该处理器72可以是中央处理单元(CPU)或者具有数据处理能力和/或指令执行能力的其它形式的处理单元,并且可以控制电子设备70中的其它组件以执行期望的功能。在本公开的一个实施例中,该处理器72用于运行该存储器71中存储的该计算机可读指令,使得该电子设备70执行前述的本公开各实施例的场景操控的全部或部分步骤。
本领域技术人员应能理解,为了解决如何获得良好用户体验效果的技术问题,本实施例中也可以包括诸如通信总线、接口等公知的结构,这些公知的结构也应包含在本公开的保护范围之内。
有关本实施例的详细说明可以参考前述各实施例中的相应说明,在此不再赘述。
图8是图示根据本公开的实施例的计算机可读存储介质的示意图。如图8所示,根据本公开实施例的计算机可读存储介质80,其上存储有非暂时性计算机可读指令81。当该非暂时性计算机可读指令81由处理器运行时,执行前述的本公开各实施例的场景操控的全部或部分步骤。
上述计算机可读存储介质80包括但不限于:光存储介质(例如:CD-ROM和DVD)、磁光存储介质(例如:MO)、磁存储介质(例如: 磁带或移动硬盘)、具有内置的可重写非易失性存储器的媒体(例如:存储卡)和具有内置ROM的媒体(例如:ROM盒)。
有关本实施例的详细说明可以参考前述各实施例中的相应说明,在此不再赘述。
图9是图示根据本公开实施例的终端设备的硬件结构示意图。如图9所示,该终端90包括上述场景操控装置实施例。
该终端设备可以以各种形式来实施,本公开中的终端设备可以包括但不限于诸如移动电话、智能电话、笔记本电脑、数字广播接收器、PDA(个人数字助理)、PAD(平板电脑)、PMP(便携式多媒体播放器)、导航装置、车载终端设备、车载显示终端、车载电子后视镜等等的移动终端设备以及诸如数字TV、台式计算机等等的固定终端设备。
作为等同替换的实施方式,该终端90还可以包括其他组件。如图9所示,该终端90可以包括电源单元91、无线通信单元92、A/V(音频/视频)输入单元93、用户输入单元94、感测单元95、接口单元96、控制器97、输出单元98和存储单元99等等。图9示出了具有各种组件的终端,但是应理解的是,并不要求实施所有示出的组件,也可以替代地实施更多或更少的组件。
其中,无线通信单元92允许终端90与无线通信系统或网络之间的无线电通信。A/V输入单元93用于接收音频或视频信号。用户输入单元94可以根据用户输入的命令生成键输入数据以控制终端设备的各种操作。感测单元95检测终端90的当前状态、终端90的位置、用户对于终端90的触摸输入的有无、终端90的取向、终端90的加速或减速移动和方向等等,并且生成用于控制终端90的操作的命令或信号。接口单元96用作至少一个外部装置与终端90连接可以通过的接口。输出单元98被构造为以视觉、音频和/或触觉方式提供输出信号。存储单元99可以存储由控制器97执行的处理和控制操作的软件程序等等,或者可以暂时地存储己经输出或将要输出的数据。存储单元99可以包括至少一种类型的存储介质。而且,终端90可以与通过网络连接执行存储单元99的存储功能的网络存储装置协作。控制器97通常控制终端设备的总体操作。另外,控制器97可以包括用于再现或回放多媒体数据的多媒体模块。 控制器97可以执行模式识别处理,以将在触摸屏上执行的手写输入或者图片绘制输入识别为字符或图像。电源单元91在控制器97的控制下接收外部电力或内部电力并且提供操作各元件和组件所需的适当的电力。
本公开提出的场景操控的各种实施方式可以使用例如计算机软件、硬件或其任何组合的计算机可读介质来实施。对于硬件实施,本公开提出的场景操控的各种实施方式可以通过使用特定用途集成电路(ASIC)、数字信号处理器(DSP)、数字信号处理装置(DSPD)、可编程逻辑装置(PLD)、现场可编程门阵列(FPGA)、处理器、控制器、微控制器、微处理器、被设计为执行这里描述的功能的电子单元中的至少一种来实施,在一些情况下,本公开提出的场景操控的各种实施方式可以在控制器97中实施。对于软件实施,本公开提出的场景操控的各种实施方式可以与允许执行至少一种功能或操作的单独的软件模块来实施。软件代码可以由以任何适当的编程语言编写的软件应用程序(或程序)来实施,软件代码可以存储在存储单元99中并且由控制器97执行。
有关本实施例的详细说明可以参考前述各实施例中的相应说明,在此不再赘述。
以上结合具体实施例描述了本公开的基本原理,但是,需要指出的是,在本公开中提及的优点、优势、效果等仅是示例而非限制,不能认为这些优点、优势、效果等是本公开的各个实施例必须具备的。另外,上述公开的具体细节仅是为了示例的作用和便于理解的作用,而非限制,上述细节并不限制本公开为必须采用上述具体的细节来实现。
在本公开中,诸如第一和第二等之类的关系术语仅仅用来将一个实体或者操作与另一个实体或操作区分开来,而不一定要求或者暗示这些实体或操作之间存在任何这种实际的关系或者顺序,本公开中涉及的器件、装置、设备、系统的方框图仅作为例示性的例子并且不意图要求或暗示必须按照方框图示出的方式进行连接、布置、配置。如本领域技术人员将认识到的,可以按任意方式连接、布置、配置这些器件、装置、设备、系统。诸如“包括”、“包含”、“具有”等等的词语是开放性词汇,指“包括但不限于”,且可与其互换使用。这里所使用的词汇“或”和“和”指词汇“和/或”,且可与其互换使用,除非上下文明确指示不是如此。这里所 使用的词汇“诸如”指词组“诸如但不限于”,且可与其互换使用。
另外,如在此使用的,在以“至少一个”开始的项的列举中使用的“或”指示分离的列举,以便例如“A、B或C的至少一个”的列举意味着A或B或C,或AB或AC或BC,或ABC(即A和B和C)。此外,措辞“示例的”不意味着描述的例子是优选的或者比其他例子更好。
还需要指出的是,在本公开的系统和方法中,各部件或各步骤是可以分解和/或重新组合的。这些分解和/或重新组合应视为本公开的等效方案。
可以不脱离由所附权利要求定义的教导的技术而进行对在此所述的技术的各种改变、替换和更改。此外,本公开的权利要求的范围不限于以上所述的处理、机器、制造、事件的组成、手段、方法和动作的具体方面。可以利用与在此所述的相应方面进行基本相同的功能或者实现基本相同的结果的当前存在的或者稍后要开发的处理、机器、制造、事件的组成、手段、方法或动作。因而,所附权利要求包括在其范围内的这样的处理、机器、制造、事件的组成、手段、方法或动作。
提供所公开的方面的以上描述以使本领域的任何技术人员能够做出或者使用本公开。对这些方面的各种修改对于本领域技术人员而言是非常显而易见的,并且在此定义的一般原理可以应用于其他方面而不脱离本公开的范围。因此,本公开不意图被限制到在此示出的方面,而是按照与在此公开的原理和新颖的特征一致的最宽范围。
为了例示和描述的目的已经给出了以上描述。此外,此描述不意图将本公开的实施例限制到在此公开的形式。尽管以上已经讨论了多个示例方面和实施例,但是本领域技术人员将认识到其某些变型、修改、改变、添加和子组合。

Claims (12)

  1. 一种场景操控方法,其特征在于,包括:
    显示场景的第一形态;
    识别触发对象;
    判断所述触发对象的动作;
    基于所述动作将所述场景的第一形态切换为场景的第二形态,所述场景的第一形态与所述场景的第二形态相关联。
  2. 根据权利要求1所述的场景操控方法,其特征在于,所述识别触发对象,包括:
    获取所述触发对象的特征信息;
    将所述特征信息与标准特征信息相比对;
    根据比对结果识别是否为触发对象。
  3. 根据权利要求2所述的场景操控方法,其特征在于,所述获取所述触发对象的特征信息,具体为:
    获取所述触发对象上的关键点。
  4. 根据权利要求1所述的场景操控方法,其特征在于,所述判断所述触发对象的动作,包括:
    获取触发对象的第一特征信息;
    获取触发对象的第二特征信息;
    基于所述第一特征信息和第二特征信息判断触发对象的动作。
  5. 根据权利要求4所述的场景操控方法,其特征在于,所述判断所述触发对象的动作,包括:
    获取触发对象的面积;
    基于所述面积判断所述触发对象的动作。
  6. 根据权利要求5所述的场景操控方法,其特征在于,所述判断所述触发对象的动作,包括:
    获取触发对象的第一面积;
    获取触发对象的第二面积;
    基于所述第一面积与所述第二面积的比对结果,判断触发对象的动作。
  7. 根据权利要求5所述的场景操控方法,其特征在于,所述获取触发对象的面积,包括:
    设定最小规则框,使得触发对象完全包含在最小规则框内,
    计算最小规则框的面积,得到触发对象的面积。
  8. 根据权利要求1所述的场景操控方法,其特征在于:
    所述触发对象为人手。
  9. 根据权利要求8所述的场景操控方法,其特征在于:
    所述动作,包括张合、旋转、远近移动或手势变化。
  10. 一种场景操控装置,其特征在于,包括:
    显示模块:显示场景的第一形态;
    识别模块:用于识别触发对象;
    判断模块:用于判断所述触发对象的动作;
    控制模块:基于所述动作将所述场景的第一形态切换为场景的第二形态,所述场景的第一形态与所述场景的第二形态相关联。
  11. 一种电子设备,其特征在于,所述电子设备包括:
    至少一个处理器;以及,
    与所述至少一个处理器通信连接的存储器;其中,
    所述存储器存储有可被所述至少一个处理器执行的指令,所述指令被所述至少一个处理器执行,以使所述至少一个处理器能够执行权利要求1-9任一所述的场景操控方法。
  12. 一种非暂态计算机可读存储介质,其特征在于,该非暂态计算机可读存储介质存储计算机指令,该计算机指令用于使计算机执行权利要求1-9任一所述的场景操控方法。
PCT/CN2019/073076 2018-06-29 2019-01-25 场景操控的方法、装置及电子设备 WO2020001015A1 (zh)

Priority Applications (3)

Application Number Priority Date Filing Date Title
GB2100223.3A GB2590207B (en) 2018-06-29 2019-01-25 Scene controlling method, device and electronic equipment
JP2020571800A JP7372945B2 (ja) 2018-06-29 2019-01-25 シナリオ制御方法、装置および電子装置
US16/769,368 US11755119B2 (en) 2018-06-29 2019-01-25 Scene controlling method, device and electronic equipment

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201810699063.9A CN108989553A (zh) 2018-06-29 2018-06-29 场景操控的方法、装置及电子设备
CN201810699063.9 2018-06-29

Publications (1)

Publication Number Publication Date
WO2020001015A1 true WO2020001015A1 (zh) 2020-01-02

Family

ID=64539579

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/073076 WO2020001015A1 (zh) 2018-06-29 2019-01-25 场景操控的方法、装置及电子设备

Country Status (5)

Country Link
US (1) US11755119B2 (zh)
JP (1) JP7372945B2 (zh)
CN (1) CN108989553A (zh)
GB (1) GB2590207B (zh)
WO (1) WO2020001015A1 (zh)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108989553A (zh) * 2018-06-29 2018-12-11 北京微播视界科技有限公司 场景操控的方法、装置及电子设备
CN112445324A (zh) * 2019-08-30 2021-03-05 北京小米移动软件有限公司 人机交互方法及装置
CN111931762B (zh) * 2020-09-25 2021-07-30 广州佰锐网络科技有限公司 基于ai的图像识别解决方法、装置及可读存储介质

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103019378A (zh) * 2012-12-07 2013-04-03 无锡清华信息科学与技术国家实验室物联网技术中心 一种移动电子设备手势控制交互方法、装置及移动终端
US20130271360A1 (en) * 2012-04-16 2013-10-17 Qualcomm Incorporated Interacting with a device using gestures
CN103383598A (zh) * 2012-05-04 2013-11-06 三星电子株式会社 终端和基于空间交互控制所述终端的方法
CN205304923U (zh) * 2015-12-23 2016-06-08 武汉哒呤科技有限公司 一种通过手势操作实现交互的手机
CN108989553A (zh) * 2018-06-29 2018-12-11 北京微播视界科技有限公司 场景操控的方法、装置及电子设备

Family Cites Families (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8872899B2 (en) 2004-07-30 2014-10-28 Extreme Reality Ltd. Method circuit and system for human to machine interfacing by hand gestures
KR101364571B1 (ko) 2010-10-06 2014-02-26 한국전자통신연구원 영상 기반의 손 검출 장치 및 그 방법
CN102226880A (zh) * 2011-06-03 2011-10-26 北京新岸线网络技术有限公司 一种基于虚拟现实的体感操作方法及系统
JP5701714B2 (ja) 2011-08-05 2015-04-15 株式会社東芝 ジェスチャ認識装置、ジェスチャ認識方法およびジェスチャ認識プログラム
US9734393B2 (en) * 2012-03-20 2017-08-15 Facebook, Inc. Gesture-based control system
US9477303B2 (en) * 2012-04-09 2016-10-25 Intel Corporation System and method for combining three-dimensional tracking with a three-dimensional display for a user interface
JP6207240B2 (ja) 2013-06-05 2017-10-04 キヤノン株式会社 情報処理装置及びその制御方法
CN103530613B (zh) * 2013-10-15 2017-02-01 易视腾科技股份有限公司 一种基于单目视频序列的目标人手势交互方法
US10156908B2 (en) * 2015-04-15 2018-12-18 Sony Interactive Entertainment Inc. Pinch and hold gesture navigation on a head-mounted display
JP6398870B2 (ja) 2015-05-25 2018-10-03 コニカミノルタ株式会社 ウェアラブル電子機器およびウェアラブル電子機器のジェスチャー検知方法
US10643390B2 (en) 2016-03-30 2020-05-05 Seiko Epson Corporation Head mounted display, method for controlling head mounted display, and computer program
JP2018084886A (ja) 2016-11-22 2018-05-31 セイコーエプソン株式会社 頭部装着型表示装置、頭部装着型表示装置の制御方法、コンピュータープログラム
CN109313499A (zh) 2016-06-07 2019-02-05 皇家飞利浦有限公司 用于向用户呈现触觉反馈的设备和用于操作该设备的方法
CN107589846A (zh) * 2017-09-20 2018-01-16 歌尔科技有限公司 场景切换方法、装置及电子设备

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130271360A1 (en) * 2012-04-16 2013-10-17 Qualcomm Incorporated Interacting with a device using gestures
CN103383598A (zh) * 2012-05-04 2013-11-06 三星电子株式会社 终端和基于空间交互控制所述终端的方法
CN103019378A (zh) * 2012-12-07 2013-04-03 无锡清华信息科学与技术国家实验室物联网技术中心 一种移动电子设备手势控制交互方法、装置及移动终端
CN205304923U (zh) * 2015-12-23 2016-06-08 武汉哒呤科技有限公司 一种通过手势操作实现交互的手机
CN108989553A (zh) * 2018-06-29 2018-12-11 北京微播视界科技有限公司 场景操控的方法、装置及电子设备

Also Published As

Publication number Publication date
CN108989553A (zh) 2018-12-11
JP7372945B2 (ja) 2023-11-01
JP2021530032A (ja) 2021-11-04
GB202100223D0 (en) 2021-02-24
GB2590207B (en) 2023-02-08
US11755119B2 (en) 2023-09-12
GB2590207A (en) 2021-06-23
US20200311398A1 (en) 2020-10-01

Similar Documents

Publication Publication Date Title
US11017580B2 (en) Face image processing based on key point detection
WO2020259651A1 (zh) 一种控制用户界面的方法及电子设备
WO2020019663A1 (zh) 基于人脸的特效生成方法、装置和电子设备
US10021319B2 (en) Electronic device and method for controlling image display
US11749020B2 (en) Method and apparatus for multi-face tracking of a face effect, and electronic device
US11513608B2 (en) Apparatus, method and recording medium for controlling user interface using input image
WO2020001014A1 (zh) 图像美化方法、装置及电子设备
US11366582B2 (en) Screenshot capturing method, device, electronic device and computer-readable medium
WO2020019664A1 (zh) 基于人脸的形变图像生成方法和装置
US20130234957A1 (en) Information processing apparatus and information processing method
WO2020001015A1 (zh) 场景操控的方法、装置及电子设备
WO2020037923A1 (zh) 图像合成方法和装置
WO2020019665A1 (zh) 基于人脸的三维特效生成方法、装置和电子设备
WO2017113821A1 (zh) 一种智能手机操作方法、装置及智能手机
CN110275611B (zh) 一种参数调节方法、装置和电子设备
US20150063785A1 (en) Method of overlappingly displaying visual object on video, storage medium, and electronic device
WO2020052083A1 (zh) 侵权图片的识别方法、装置和计算机可读存储介质
JP2024518333A (ja) マルチスクリーンインタラクション方法及び機器、端末装置、及び車両
WO2022111458A1 (zh) 图像拍摄方法和装置、电子设备及存储介质
EP2939411A1 (en) Image capture
WO2020037924A1 (zh) 动画生成方法和装置
WO2020029556A1 (zh) 自适应平面的方法、装置和计算机可读存储介质
WO2020000975A1 (zh) 视频拍摄方法、客户端、终端及介质
WO2020029555A1 (zh) 用于平面间无缝切换的方法、装置和计算机可读存储介质
CN110827413A (zh) 控制虚拟物体形态改变的方法、装置和计算机可读存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19825265

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2020571800

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 202100223

Country of ref document: GB

Kind code of ref document: A

Free format text: PCT FILING DATE = 20190125

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 08.04.2021)

122 Ep: pct application non-entry in european phase

Ref document number: 19825265

Country of ref document: EP

Kind code of ref document: A1