US10162515B2 - Method and electronic device for controlling display objects on a touch display based on a touch directional touch operation that both selects and executes a function - Google Patents

Method and electronic device for controlling display objects on a touch display based on a touch directional touch operation that both selects and executes a function Download PDF

Info

Publication number
US10162515B2
US10162515B2 US14/858,332 US201514858332A US10162515B2 US 10162515 B2 US10162515 B2 US 10162515B2 US 201514858332 A US201514858332 A US 201514858332A US 10162515 B2 US10162515 B2 US 10162515B2
Authority
US
United States
Prior art keywords
touch
display objects
instruction
trajectory
display
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US14/858,332
Other versions
US20160349982A1 (en
Inventor
Xiaoping Zhang
Qingdi Xie
Xueyuan Zhang
Qiang Yao
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Lenovo Beijing Ltd
Beijing Lenovo Software Ltd
Original Assignee
Lenovo Beijing Ltd
Beijing Lenovo Software Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lenovo Beijing Ltd, Beijing Lenovo Software Ltd filed Critical Lenovo Beijing Ltd
Assigned to LENOVO (BEIJING) LIMITED, BEIJING LENOVO SOFTWARE LTD. reassignment LENOVO (BEIJING) LIMITED ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: XIE, QINGDI, YAO, QIANG, ZHANG, XIAOPING, ZHANG, XUEYUAN
Publication of US20160349982A1 publication Critical patent/US20160349982A1/en
Application granted granted Critical
Publication of US10162515B2 publication Critical patent/US10162515B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04842Selection of displayed objects or displayed text elements

Definitions

  • the present disclosure relates to the field of electronic technology, and more particularly, to an information processing method and an electronic device.
  • Embodiments of the present disclosure provide an information processing method and an electronic device to select and edit some of multiple objects in the electronic device, thereby simplifying the operation and improving the efficiency in the operation.
  • an information processing method comprising:
  • N is a positive integer less than or equal to M.
  • the first touch operation is a touch operation implemented by an operation body on the touch display unit, and the operation body is a finger of a user or a stylus.
  • controlling N of the M display objects to be in a selected state in response to the first touch operation comprises:
  • controlling the N display objects to be in the selected state comprises:
  • determining a first touch trajectory of the first touch operation based on the touch start point and the touch end point comprises:
  • determining the N display objects from the M display objects based on the first touch trajectory comprises:
  • the N display objects at least comprise at least one of the M display objects which is located in the first closed region;
  • generating and executing a first operation instruction for implementing a first operation on the N display objects comprises:
  • generating and executing a first operation instruction for implementing a first operation on the N display objects comprises:
  • the first operation instruction is:
  • the M objects are file objects and/or application objects displayed on the touch display unit; or
  • the M objects are characters and/or symbols displayed on the touch display unit after a text file is opened through a text application.
  • an electronic device comprising:
  • a touch display unit arranged in the housing
  • a processor arranged in the housing and comprising hardware configured to detect and acquire a first touch operation on the touch display unit of the electronic device when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1, control N of the M display objects to be in a selected state in response to the first touch operation, and generate and execute a first operation instruction for implementing a first operation on the N display objects, wherein N is a positive integer less than or equal to M.
  • the hardware is further configured to:
  • the hardware is further configured to:
  • the hardware is further configured to:
  • the hardware is further configured to:
  • the N display objects determine the N display units from the M display objects, wherein the N display objects at least comprise at least one of the M display objects which is located in the first closed region;
  • the hardware is further configured to:
  • the hardware is further configured to:
  • an electronic device comprising:
  • a first detection unit configured to detect and acquire a first touch operation on a touch display unit of the electronic device when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1;
  • a first control unit configured to control N of the M display objects to be in a selected state in response to the first touch operation, and generate and execute a first operation instruction for implementing a first operation on the N display objects, wherein N is a positive integer less than or equal to M.
  • the technical measure comprises detecting and acquiring a first touch operation on a touch display unit of the electronic device when M display objects are displayed on the touch display unit, controlling N of the M display objects to be in a selected state in response to the first touch operation, and generating and executing a first operation instruction for implementing a first operation on the N display objects.
  • the electronic device can not only select objects corresponding to the touch operation based on the touch operation, but also can implement other operations on the selected objects based on the first operation. This simplifies the operation and improves the efficiency in the operation.
  • the technical measure comprises acquiring a first touch direction of the first touch operation and generating and executing the first operation instruction for implementing a first operation on the N display objects at least based on the first touch direction.
  • the electronic device can generate multiple operation instructions according to different touch directions, so as to satisfy various users' requirements for operations so as to be convenient for use and improve the user experience.
  • the technical measure comprises: the first operation instruction specifically being a deletion instruction of deleting the N display objects, or a copy instruction of copying the N display objects, or a cut instruction of cutting the N display objects, or a share instruction of sharing the N display objects, or a selected state canceling instruction of canceling the selected state of the N display objects.
  • the user can not only achieve an edition operation of display objects in one touch operation, but also can achieve an canceling operation without implementing other touch operations when the user wants to implement a canceling operation so as to simplify the operation.
  • FIG. 1 is a flowchart of an information processing method according to an embodiment of the present disclosure
  • FIG. 2 is a diagram of a first specific implementation of step S 101 in a first embodiment of the present disclosure
  • FIG. 3 is a diagram of a second specific implementation of step S 101 in the first embodiment of the present disclosure.
  • FIG. 4 is a flowchart of a specific implementation of a first step in step S 102 in the first embodiment of the present disclosure
  • FIG. 5 is a flowchart of a specific implementation of step S 10212 in the first embodiment of the present disclosure
  • FIG. 6 is a flowchart of a specific implementation of step S 10214 in the first embodiment of the present disclosure
  • FIG. 7 is a diagram of a specific implementation of step S 10214 in the first embodiment of the present disclosure.
  • FIG. 8A is a diagram of moving a selected bar to the left according to the first embodiment of the present disclosure.
  • FIG. 8B is a diagram of adjusting a position of a leftmost end of a selected bar according to the first embodiment of the present disclosure
  • FIG. 9A is a diagram of moving a selected bar to the right according to the first embodiment of the present disclosure.
  • FIG. 9B is a diagram of adjusting a position of a rightmost end of a selected bar according to the first embodiment of the present disclosure.
  • FIG. 10 is a flowchart of a specific implementation of a second step in step S 102 according to the first embodiment of the present disclosure
  • FIG. 11 is a structural diagram of an electronic device according to a second embodiment of the present disclosure.
  • FIG. 12 is a structural block diagram of an electronic device according to a third embodiment of the present disclosure.
  • the embodiments of the present disclosure provide an information processing method and an electronic device, to simplify the operation and improve the efficiency in the operation.
  • An information processing method may comprise:
  • N is a positive integer less than or equal to M.
  • the method comprises detecting and acquiring a first touch operation on a touch display unit of the electronic device when M display objects are displayed on the touch display unit, controlling N of the M display objects to be in a selected state in response to the first touch operation, and generating and executing a first operation instruction for implementing a first operation on the N display objects.
  • the electronic device can not only select objects corresponding to the touch operation based on the touch operation, but also can implement other operations on the selected objects based on the first operation. This simplifies the process that an operation can only be completed if a selection operation is firstly implemented on some of display objects and then corresponding operation options are selected from an operation menu when said some display objects need to be edited. Therefore, the present disclosure may effectively simplify the operation and improve the efficiency in the operation.
  • the electronic device may be a tablet computer, a touch screen mobile phone, or another electronic device having a touch unit.
  • the touch unit may be a conventional hard touch screen, or may also be a flexible touch screen, which will not be limited in the embodiments of the present disclosure.
  • FIG. 1 illustrated is a flowchart of a specific implementation of an information processing method according to the first embodiment of the present disclosure.
  • the method comprises the following steps:
  • step S 101 a first touch operation on a touch display unit of the electronic device is detected and acquired when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1;
  • N of the M display objects are controlled to be in a selected state in response to the first touch operation, and a first operation instruction for implementing a first operation on the N display objects is generated and executed, wherein N is a positive integer less than or equal to M.
  • the first touch operation is specifically a touch operation implemented by an operation body on the touch display unit, and the operation body is specifically a finger of the user or a stylus.
  • the M objects are file objects and/or application objects displayed on the touch display unit; or the M objects are characters and/or symbols displayed on the touch display unit after a text file is opened through a text application.
  • the first touch operation may be a touch operation implemented by the user on an application on a desktop of the tablet computer with a finger or a stylus on a screen of the tablet computer, or a touch operation implemented by the user on characters, symbols or the like in an opened word document with a finger or a stylus on the screen of the tablet computer, which will not be enumerated here.
  • step S 101 is firstly performed, i.e., a first touch operation on a touch display unit of the electronic device is detected and acquired when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1.
  • a user wants to implement a copy operation on characters in a word document.
  • the first touch operation being specifically a touch operation implemented by the user on characters in an opened word document with a stylus on a screen of the tablet computer as an example
  • a circle drawing operation may be implemented around “ ”, as shown in FIG. 2
  • a transverse line drawing operation may also be implemented on “ ”, as shown in FIG. 3 .
  • the touch operation of the user is detected by a sensor of the screen of the tablet computer.
  • step S 102 is performed in the method according to the embodiment of the present disclosure, i.e., N of the M display objects are controlled to be in a selected state in response to the first touch operation, and a first operation instruction for implementing a first operation on the N display objects is generated and executed, wherein N is a positive integer less than or equal to M.
  • step S 102 may be divided into two steps.
  • N of the M display objects are controlled to be in a selected state in response to the first touch operation
  • a first operation instruction for implementing a first operation on the N display objects is generated and executed.
  • a specific implementation of the first step in the embodiment of the present disclosure comprises the following steps.
  • a first touch trajectory of the first touch operation is determined based on the touch start point and the touch end point;
  • the N display objects are determined from the M display objects based on the first touch trajectory
  • the N display objects are controlled to be in the selected state.
  • an operation start point and an operation end point of the touch operation are acquired.
  • the touch operation is a circle drawing operation
  • the touch start point and the touch end point are at 1 ⁇ 3 of the character “ ” from left to right, i.e., the operation start point is the same as the operation end point.
  • the touch operation is a transverse line drawing operation
  • the touch start point is at 1 ⁇ 2 of the character “ ” from left to right
  • the touch end point is at 1 ⁇ 4 of the character “ ” from left to right, i.e., the touch start point is at a distance from the touch end point.
  • step S 10212 is performed in the method according to the embodiment of the present disclosure, i.e., a first touch trajectory of the first touch operation is determined based on the touch start point and the touch end point.
  • step S 10212 comprises the following steps.
  • the touch operation has been detected on the screen of the tablet computer.
  • the touch operation being a circle drawing operation as an example
  • the tablet computer judges that the touch start point and the touch end point of the touch operation are the same. Further, if it is judged that there is another touch point between the touch start point and the touch end point, for example, a touch point at the rightmost end of the trajectory of the circle drawing operation is at 1 ⁇ 2 of the character “ ” from left to right, it is determined that the circle drawing touch operation forms a closed touch trajectory.
  • the tablet computer judges that the touch start point and the touch end point are not in the same position, and thus determines that the transverse line drawing touch operation forms a non-closed touch trajectory.
  • step S 10213 is performed in the method according to the embodiment of the present disclosure, i.e., the N display objects are determined from the M display objects based on the first touch trajectory.
  • step S 10213 comprises:
  • the N display objects at least comprise at least one of the M display objects which is located in the first closed region;
  • the tablet computer determines that the touch trajectory is a closed touch trajectory, it is determined that characters surrounded by the closed trajectory belong to the N display objects.
  • a character surrounded by the closed trajectory is necessarily one of the N objects, such as three characters “ ” illustrated in the figure.
  • the tablet computer may be configured to determine that a character is one of the N objects only if the 2 ⁇ 3 of the character is surrounded by the closed trajectory.
  • the tablet computer determines that the N objects are “ ”.
  • which character is one of the N objects may be set by those skilled in the art according to practical usage requirements, for example, it may be determined that any character on the touch trajectory is one of the N objects, which is not limited in the present disclosure.
  • the tablet computer may be configured to determine that any character on the touch trajectory is one of the N objects.
  • the touch start point is at 1 ⁇ 2 of the character “ ” from left to right and the touch end point is at 1 ⁇ 4 of the character “ ” from left to right, characters on the touch trajectory are “ ”, and thus it is determined that the N objects are “ ”.
  • which character is one of the N objects may be set by those skilled in the art according to practical usage requirements, for example, it may be determined that a character is one of the N objects only if 1 ⁇ 2 of the character is surrounded by the touch trajectory, which is not limited in the present disclosure.
  • step S 10214 is performed in the method according to the embodiment of the present disclosure, i.e., the N display objects are controlled to be in the selected state.
  • step S 10214 comprises the following steps.
  • the N display units are controlled to be in the selected state when the detection result is yes.
  • the tablet computer detects touch duration corresponding to the end point of the touch operation, for example, 2 s.
  • Predetermined duration may be specifically set to for example 1s or 2 s by a user or those skilled in the art according to practical usage conditions.
  • the tablet computer judges that the touch duration 2 s of the touch end point is greater than 1 s, then selects the N objects “ ” or “ ”, and feeds the content of the characters which are currently selected back to the user on the screen of the tablet computer, for example, by adding a shadow on “ ”, as shown in FIG. 7 , or changing a font color of “ ”, which is not limited in the present disclosure.
  • a touch operation of moving to the left or to the right in a position of the end point of the touch operation is further detected.
  • the touch operation being a circle drawing operation as an example
  • the tablet computer detects a touch operation of moving to the left at the touch end point, i.e., there is a touch operation of moving a character to the left at 1 ⁇ 2 of the character “ ” from left to right.
  • the entire selected bar is moved by the tablet computer to the left for a distance of one character, and then “: ” are selected by the tablet computer, as shown in FIG.
  • the leftmost end of the selected bar is moved by the tablet computer to the left for a distance of one character, and then “: ” are selected by the tablet computer, as shown in FIG. 8B .
  • the entire selected bar is moved by the tablet computer to the right, and then “ ” are selected by the tablet computer, as shown in FIG. 9A ; or the rightmost end of the selected bar is moved by the tablet computer to the right, and then “ ” are selected by the tablet computer, as shown in FIG. 9B .
  • this may be set by those skilled in the art according to practical usage requirements, which is not limited in the present disclosure.
  • the second step is performed in the method according to the embodiment of the present disclosure, i.e., the first operation instruction for implementing a first operation on the N display objects is generated and executed.
  • the second step may be specifically implemented in the following two manners.
  • the second step specifically comprises the following steps.
  • the first operation instruction for implementing a first operation on the N display objects is generated and executed at least based on the first touch direction.
  • the first operation instruction is specifically:
  • the tablet computer judges that the direction of the trajectory is a clockwise direction.
  • a user may preset an operation instruction corresponding to the direction of the trajectory in the tablet computer according to practical usage requirements, for example, a copy operation instruction corresponding to a clockwise circle drawing operation, a cut operation instruction corresponding to an anticlockwise circle drawing operation or the like.
  • the operation instructions may comprise a deletion operation instruction, a copy operation instruction, a cut operation instruction, a share operation instruction, and a cancel operation instruction.
  • the tablet computer judges that the direction of the circle drawing trajectory is a clockwise direction, and determines that the operation instruction corresponding to the clockwise direction is a copy operation from the preset operation instructions. Then, the tablet computer generates an operation instruction of copying “ ”. When the touch operation ends, the tablet computer achieves an operation of copying “ ”.
  • the user When the user implements a clockwise circle drawing operation on “ ”, and does not want to implement any operation on the four characters before the stylus is lifted up, the user may implement a slide operation in an upward direction or a downward direction.
  • the tablet computer detects that the touch operation is an operation of firstly drawing a circle in a clockwise direction and then sliding at the end point of the circle in an upward direction or a downward direction, it is determined that the operation instruction corresponding to the current operation is an operation instruction of canceling a selected state of “ ”.
  • the second step specifically comprises:
  • the first operation instruction is specifically:
  • the tablet computer judges that the direction of the trajectory is a left-to-right direction.
  • a user may preset an operation instruction corresponding to the direction of the trajectory in the tablet computer according to practical usage requirements, for example, a copy operation instruction corresponding to a left-to-right transverse line drawing operation, a cut operation instruction corresponding to a right-to-left transverse line drawing operation or the like.
  • the operation instructions may comprise a deletion operation instruction, a copy operation instruction, a cut operation instruction, a share operation instruction, and a cancel operation instruction.
  • the tablet computer judges that the direction of the transverse line drawing trajectory is a left-to-right direction, and determines that the operation instruction corresponding to the left-to-right direction is a copy operation from the preset operation instructions. Then, the tablet computer generates an operation instruction of copying “ ”. When the touch operation ends, the tablet computer achieves an operation of copying “ ”.
  • the user When the user implements a left-to-right transverse line drawing operation on “ ”, and does not want to implement any operation on the five characters before the stylus is lifted up, the user may implement a slide operation in an upward direction or a downward direction.
  • the tablet computer detects that the touch operation is an operation of firstly drawing a transverse line in a left-to-right direction and then sliding at the end point of the transverse line in an upward direction or a downward direction, it is determined that the operation instruction corresponding to the current operation is an operation instruction of canceling a selected state of “ ”.
  • the second embodiment of the present disclosure provides an electronic device, comprising:
  • a housing 10 a housing 10 ;
  • a touch display unit 20 arranged in the housing 10 ;
  • a processor 30 arranged in the housing 10 and comprising hardware configured to detect and acquire a first touch operation on the touch display unit of the electronic device when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1, control N of the M display objects to be in a selected state in response to the first touch operation, and generate and execute a first operation instruction for implementing a first operation on the N display objects, wherein N is a positive integer less than or equal to M.
  • the hardware in the processor 30 is further configured to:
  • the hardware in the processor 30 is further configured to:
  • the hardware in the processor 30 is further configured to:
  • the hardware in the processor 30 is further configured to:
  • the N display objects determine the N display units from the M display objects, wherein the N display objects at least comprise at least one of the M display objects which is located in the first closed region;
  • the hardware in the processor 30 is further configured to:
  • the hardware in the processor 30 is further configured to:
  • the electronic device described above corresponds to the information processing method according to the first embodiment one by one, and therefore will not be described in detail here.
  • the third embodiment of the present disclosure provides an electronic device, comprising:
  • a first detection unit 101 configured to detect and acquire a first touch operation on a touch display unit of the electronic device when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1;
  • a first control unit 102 configured to control N of the M display objects to be in a selected state in response to the first touch operation, and generate and execute a first operation instruction for implementing a first operation on the N display objects, wherein N is a positive integer less than or equal to M.
  • the first control unit 102 specifically comprises:
  • a first acquisition sub-unit configured to acquire a touch start point and a touch end point of the first touch operation in response to the first touch operation
  • a first determination sub-unit configured to determine a first touch trajectory of the first touch operation based on the touch start point and the touch end point;
  • a second determination sub-unit configured to determine the N display objects from the M display objects based on the first touch trajectory
  • a first control sub-unit configured to control the N display objects to be in the selected state.
  • the first control sub-unit specifically comprises:
  • a first acquisition module configured to acquire touch duration in which the operation body keeps in a contact state with the touch display unit at the touch end point
  • a first detection module configured to detect whether the touch duration is greater than or equal to predetermined duration to acquire a detection result
  • a first control module configured to control the N display units to be in the selected state when the detection result is yes.
  • the first determination sub-unit specifically comprises:
  • a first judgment module configured to judge whether the touch start point is the same as the touch end point and there is another touch point between the touch start point and the touch end point to acquire a judgment result
  • a first determination module configured to determine that the touch trajectory of the first touch operation is a closed touch trajectory having a first closed region on the touch display unit when the judgment result is yes;
  • a second determination module configured to determine that the touch trajectory of the first touch operation is a non-closed touch trajectory when the judgment result is no.
  • the second determination sub-unit specifically comprises:
  • a third determination module configured to determine the N display objects from the M display objects, wherein the N display objects at least comprise at least one of the M display objects which is located in the first closed region;
  • a fourth determination module configured to determine that at least one of the M display objects which is on the non-closed touch trajectory belongs to the N display objects.
  • the first control unit 102 when the first touch trajectory is the closed touch trajectory, specifically comprises:
  • a second acquisition sub-unit configured to acquire a first touch direction of the first touch operation
  • a first execution sub-unit configured to generate and execute the first operation instruction for implementing a first operation on the N display objects at least based on the first touch direction.
  • the first control unit 102 when the first touch trajectory is the non-closed touch trajectory, specifically comprises:
  • a second execution sub-unit configured to generate and execute the first operation instruction for implementing a first operation on the N display objects based on the non-closed touch trajectory.
  • each flow and/or block in the flowcharts and/or block diagrams as well as a combination of the flows and/or blocks in the flowcharts and/or block diagrams can be implemented by computer program instructions.
  • These computer program instructions can be provided to a processor of a general-purpose computer, a dedicated-purpose computer, an embedded processor, or other programmable data processing devices to generate a machine. Thereby, the instructions executed by the processor of the computer or other programmable data processing devices generate means for implementing functions specified in one or more flows in the flowcharts and/or one or more blocks in the block diagrams.
  • These computer program instructions can also be stored in a computer readable memory capable of introducing a computer or other programmable data processing devices to operate in a particular mode.
  • the instructions stored in the computer readable memory generate an article of manufacture comprising instruction means for implementing functions specified in one or more flows in the flowcharts and/or one or more blocks in the block diagrams.
  • These computer program instructions can also be loaded onto a computer or other programmable data processing devices, so as to enable a series of operation steps to be performed on the computer or other programmable devices to generate a computer-implemented process.
  • the instructions executed on the computer or other programmable devices provide a step of implementing functions specified in one or more flows in the flowcharts and/or one or more blocks in the block diagrams.
  • computer program instructions corresponding to the information processing method may be stored on a storage medium such as an optical disk, a hard disk, a USB device etc.
  • a storage medium such as an optical disk, a hard disk, a USB device etc.
  • N is a positive integer less than or equal to M.
  • the N display objects at least comprise at least one of the M display objects which is located in the first closed region;
  • the first touch trajectory is the closed touch trajectory
  • computer instructions stored in the storage medium which correspond to the step of generating and executing a first operation instruction for implementing a first operation on the N display objects, are executed, the following steps are further performed:
  • the first touch trajectory is the non-closed touch trajectory
  • computer instructions stored in the storage medium which correspond to the step of generating and executing a first operation instruction for implementing a first operation on the N display objects, are executed, the following steps are further performed:

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The present disclosure discloses an information processing method and an electronic device. The method comprises: detecting and acquiring a first touch operation on a touch display unit of the electronic device when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1; and controlling N of the M display objects to be in a selected state in response to the first touch operation, and generating and executing a first operation instruction for implementing a first operation on the N display objects, wherein N is a positive integer less than or equal to M.

Description

CROSS-REFERENCE TO RELATED APPLICATION(S)
This application claims priority to the Chinese Patent Application No. 201510275191.7, filed on May 26, 2015, entitled “INFORMATION PROCESSING METHOD AND ELECTRONIC DEVICE” which is incorporated herein by reference in its entirety.
TECHNICAL FIELD
The present disclosure relates to the field of electronic technology, and more particularly, to an information processing method and an electronic device.
BACKGROUND
With continuous development of science and technology, there are also increasing types and functions of electronics, and the electronics gradually become necessities of people's life.
Conventionally, when a user edits display objects on a touch screen of an electronic device with a stylus, and wants to implement a copy operation on some of the display objects, for example, some of characters in a text object, the user firstly needs to select a required text by dragging the stylus, and then lift up the stylus. In this case, an operation menu comprising “copy”, “cut”, “paste” and “delete” appears on the touch screen of the electronic device. Then, the user clicks on the touch screen again with the stylus to select “copy” in the operation menu. Then, a text selected by the user is copied by the electronic device, and is stored in a clipboard of the electronic device for use by the user.
The inventor of the present disclosure discovers the following technical problem in the related art in the process of implementing the technical solutions according to the embodiments of the present disclosure:
As it needs to achieve the purpose of edition by implementing multiple operations such as an object selection operation, an edition operation or the like when the user needs to edit the display objects of the electronic device, there is a technical problem in the related art that the operation is complex and the efficiency is low when there is a need to select and edit some of multiple objects in the electronic device.
SUMMARY
Embodiments of the present disclosure provide an information processing method and an electronic device to select and edit some of multiple objects in the electronic device, thereby simplifying the operation and improving the efficiency in the operation.
According to a first aspect of the present disclosure, an information processing method is provided, comprising:
detecting and acquiring a first touch operation on a touch display unit of an electronic device when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1; and
controlling N of the M display objects to be in a selected state in response to the first touch operation, and generating and executing a first operation instruction for implementing a first operation on the N display objects, wherein N is a positive integer less than or equal to M.
In a possible implementation, the first touch operation is a touch operation implemented by an operation body on the touch display unit, and the operation body is a finger of a user or a stylus.
In a possible implementation, controlling N of the M display objects to be in a selected state in response to the first touch operation comprises:
acquiring a touch start point and a touch end point of the first touch operation in response to the first touch operation;
determining a first touch trajectory of the first touch operation based on the touch start point and the touch end point;
determining the N display objects from the M display objects based on the first touch trajectory; and
controlling the N display objects to be in the selected state.
In a possible implementation, controlling the N display objects to be in the selected state comprises:
acquiring touch duration in which the operation body keeps in a contact state with the touch display unit at the touch end point;
detecting whether the touch duration is greater than or equal to predetermined duration to acquire a detection result; and
controlling the N display units to be in the selected state when the detection result is yes.
In a possible implementation, determining a first touch trajectory of the first touch operation based on the touch start point and the touch end point comprises:
judging whether the touch start point is the same as the touch end point and there is another touch point between the touch start point and the touch end point to acquire a judgment result;
determining that the touch trajectory of the first touch operation is a closed touch trajectory having a first closed region on the touch display unit when the judgment result is yes; and
determining that the touch trajectory of the first touch operation is a non-closed touch trajectory when the judgment result is no.
In a possible implementation, determining the N display objects from the M display objects based on the first touch trajectory comprises:
determining the N display objects from the M display objects, wherein the N display objects at least comprise at least one of the M display objects which is located in the first closed region; or
determining that at least one of the M display objects which is on the non-closed touch trajectory belongs to the N display objects.
In a possible implementation, when the first touch trajectory is the closed touch trajectory, generating and executing a first operation instruction for implementing a first operation on the N display objects comprises:
acquiring a first touch direction of the first touch operation; and
generating and executing the first operation instruction for implementing a first operation on the N display objects at least based on the first touch direction.
In a possible implementation, when the first touch trajectory is the non-closed touch trajectory, generating and executing a first operation instruction for implementing a first operation on the N display objects comprises:
generating and executing the first operation instruction for implementing a first operation on the N display objects based on the non-closed touch trajectory.
In a possible implementation, the first operation instruction is:
a deletion instruction of deleting the N display objects; or
a copy instruction of copying the N display objects; or
a cut instruction of cutting the N display objects; or
a share instruction of sharing the N display objects; or
a selected state canceling instruction of canceling the selected state of the N display objects.
In a possible implementation, the M objects are file objects and/or application objects displayed on the touch display unit; or
the M objects are characters and/or symbols displayed on the touch display unit after a text file is opened through a text application.
According to another aspect of the present disclosure, an electronic device is provided, comprising:
a housing;
a touch display unit arranged in the housing; and
a processor arranged in the housing and comprising hardware configured to detect and acquire a first touch operation on the touch display unit of the electronic device when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1, control N of the M display objects to be in a selected state in response to the first touch operation, and generate and execute a first operation instruction for implementing a first operation on the N display objects, wherein N is a positive integer less than or equal to M.
In a possible implementation, the hardware is further configured to:
acquire a touch start point and a touch end point of the first touch operation in response to the first touch operation;
determine a first touch trajectory of the first touch operation based on the touch start point and the touch end point;
determine the N display objects from the M display objects based on the first touch trajectory; and
control the N display objects to be in the selected state.
In a possible implementation, the hardware is further configured to:
acquire touch duration in which the operation body keeps in a contact state with the touch display unit at the touch end point;
detect whether the touch duration is greater than or equal to predetermined duration to acquire a detection result; and
control the N display units to be in the selected state when the detection result is yes.
In a possible implementation, the hardware is further configured to:
judge whether the touch start point is the same as the touch end point and there is another touch point between the touch start point and the touch end point to acquire a judgment result;
determine that the touch trajectory of the first touch operation is a closed touch trajectory having a first closed region on the touch display unit when the judgment result is yes; and
determine that the touch trajectory of the first touch operation is a non-closed touch trajectory when the judgment result is no.
In a possible implementation, the hardware is further configured to:
determine the N display units from the M display objects, wherein the N display objects at least comprise at least one of the M display objects which is located in the first closed region; or
determine that at least one of the M display objects which is on the non-closed touch trajectory belongs to the N display objects.
In a possible implementation, the hardware is further configured to:
acquire a first touch direction of the first touch operation; and
generate and execute the first operation instruction for implementing a first operation on the N display objects at least based on the first touch direction.
In a possible implementation, the hardware is further configured to:
generate and execute the first operation instruction for implementing a first operation on the N display objects based on the non-closed touch trajectory.
According to another aspect of the present application, an electronic device is further provided, comprising:
a first detection unit configured to detect and acquire a first touch operation on a touch display unit of the electronic device when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1; and
a first control unit configured to control N of the M display objects to be in a selected state in response to the first touch operation, and generate and execute a first operation instruction for implementing a first operation on the N display objects, wherein N is a positive integer less than or equal to M.
According to the embodiments of the present disclosure, the technical measure comprises detecting and acquiring a first touch operation on a touch display unit of the electronic device when M display objects are displayed on the touch display unit, controlling N of the M display objects to be in a selected state in response to the first touch operation, and generating and executing a first operation instruction for implementing a first operation on the N display objects. In this way, when a touch operation is detected by the electronic device, the electronic device can not only select objects corresponding to the touch operation based on the touch operation, but also can implement other operations on the selected objects based on the first operation. This simplifies the operation and improves the efficiency in the operation.
In addition, according to the embodiments of the present disclosure, the technical measure comprises acquiring a first touch direction of the first touch operation and generating and executing the first operation instruction for implementing a first operation on the N display objects at least based on the first touch direction. In this way, the electronic device can generate multiple operation instructions according to different touch directions, so as to satisfy various users' requirements for operations so as to be convenient for use and improve the user experience.
In addition, according to the embodiments of the present disclosure, the technical measure comprises: the first operation instruction specifically being a deletion instruction of deleting the N display objects, or a copy instruction of copying the N display objects, or a cut instruction of cutting the N display objects, or a share instruction of sharing the N display objects, or a selected state canceling instruction of canceling the selected state of the N display objects. In this way, the user can not only achieve an edition operation of display objects in one touch operation, but also can achieve an canceling operation without implementing other touch operations when the user wants to implement a canceling operation so as to simplify the operation.
BRIEF DESCRIPTION OF THE DRAWINGS
In order to illustrate technical solutions in the embodiments of the present disclosure or in the related art more clearly, accompanying drawings needed to be used in the description of the embodiments will be described below in brief. Obviously, the accompanying drawings described below are merely some embodiments of the present disclosure.
FIG. 1 is a flowchart of an information processing method according to an embodiment of the present disclosure;
FIG. 2 is a diagram of a first specific implementation of step S101 in a first embodiment of the present disclosure;
FIG. 3 is a diagram of a second specific implementation of step S101 in the first embodiment of the present disclosure;
FIG. 4 is a flowchart of a specific implementation of a first step in step S102 in the first embodiment of the present disclosure;
FIG. 5 is a flowchart of a specific implementation of step S10212 in the first embodiment of the present disclosure;
FIG. 6 is a flowchart of a specific implementation of step S10214 in the first embodiment of the present disclosure;
FIG. 7 is a diagram of a specific implementation of step S10214 in the first embodiment of the present disclosure;
FIG. 8A is a diagram of moving a selected bar to the left according to the first embodiment of the present disclosure;
FIG. 8B is a diagram of adjusting a position of a leftmost end of a selected bar according to the first embodiment of the present disclosure;
FIG. 9A is a diagram of moving a selected bar to the right according to the first embodiment of the present disclosure;
FIG. 9B is a diagram of adjusting a position of a rightmost end of a selected bar according to the first embodiment of the present disclosure;
FIG. 10 is a flowchart of a specific implementation of a second step in step S102 according to the first embodiment of the present disclosure;
FIG. 11 is a structural diagram of an electronic device according to a second embodiment of the present disclosure; and
FIG. 12 is a structural block diagram of an electronic device according to a third embodiment of the present disclosure.
DETAILED DESCRIPTION
The embodiments of the present disclosure provide an information processing method and an electronic device, to simplify the operation and improve the efficiency in the operation.
The general concept of the technical solutions according to the embodiments of the present application is as follows.
An information processing method may comprise:
detecting and acquiring a first touch operation on a touch display unit of the electronic device when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1; and
controlling N of the M display objects to be in a selected state in response to the first touch operation, and generating and executing a first operation instruction for implementing a first operation on the N display objects, wherein N is a positive integer less than or equal to M.
In the above technical solutions, the method comprises detecting and acquiring a first touch operation on a touch display unit of the electronic device when M display objects are displayed on the touch display unit, controlling N of the M display objects to be in a selected state in response to the first touch operation, and generating and executing a first operation instruction for implementing a first operation on the N display objects. In this way, when a touch operation is detected by the electronic device, the electronic device can not only select objects corresponding to the touch operation based on the touch operation, but also can implement other operations on the selected objects based on the first operation. This simplifies the process that an operation can only be completed if a selection operation is firstly implemented on some of display objects and then corresponding operation options are selected from an operation menu when said some display objects need to be edited. Therefore, the present disclosure may effectively simplify the operation and improve the efficiency in the operation.
In order to make the technical solutions, and advantages of the embodiments of the present disclosure more clear and obvious, the technical solutions according to the present disclosure will be clearly described below with reference to the drawings of the embodiments of the present disclosure. Obviously, the embodiments as described are merely a part of the embodiments in the technical solutions of the present disclosure. All other embodiments acquired by those skilled in the art based on the embodiments recited in the documents of the present disclosure without any creative effort should belong to the scope protected by the technical solutions of the present disclosure.
Before describing an information processing method according to the embodiments of the present disclosure in detail, an electronic device corresponding to the method will firstly be basically described. The electronic device according to the embodiments of the present disclosure may be a tablet computer, a touch screen mobile phone, or another electronic device having a touch unit. The touch unit may be a conventional hard touch screen, or may also be a flexible touch screen, which will not be limited in the embodiments of the present disclosure.
First Embodiment
In order to enable those skilled in the art to clearly understand the technical solution according to the embodiment of the present disclosure, the technical solution will be set forth in the following specific description by taking the electronic device being a tablet computer as an example. A specific implementation of the method according to the embodiment of the present disclosure will be described below by way of examples.
With reference to FIG. 1, illustrated is a flowchart of a specific implementation of an information processing method according to the first embodiment of the present disclosure. The method comprises the following steps:
In step S101, a first touch operation on a touch display unit of the electronic device is detected and acquired when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1; and
in step S102, N of the M display objects are controlled to be in a selected state in response to the first touch operation, and a first operation instruction for implementing a first operation on the N display objects is generated and executed, wherein N is a positive integer less than or equal to M.
In the first embodiment of the present disclosure, the first touch operation is specifically a touch operation implemented by an operation body on the touch display unit, and the operation body is specifically a finger of the user or a stylus. The M objects are file objects and/or application objects displayed on the touch display unit; or the M objects are characters and/or symbols displayed on the touch display unit after a text file is opened through a text application.
Specifically, by taking the electronic device being a tablet computer as an example, the first touch operation may be a touch operation implemented by the user on an application on a desktop of the tablet computer with a finger or a stylus on a screen of the tablet computer, or a touch operation implemented by the user on characters, symbols or the like in an opened word document with a finger or a stylus on the screen of the tablet computer, which will not be enumerated here.
When information is processed with the method according to the first embodiment of the present disclosure, step S101 is firstly performed, i.e., a first touch operation on a touch display unit of the electronic device is detected and acquired when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1.
In a specific implementation, by taking the electronic device being a tablet computer as an example, a user wants to implement a copy operation on characters in a word document. In this case, by taking the first touch operation being specifically a touch operation implemented by the user on characters in an opened word document with a stylus on a screen of the tablet computer as an example, when content of a word document is displayed on the screen and the user wants to edit “
Figure US10162515-20181225-P00001
” in the content of the document, a circle drawing operation may be implemented around “
Figure US10162515-20181225-P00002
”, as shown in FIG. 2, or a transverse line drawing operation may also be implemented on “
Figure US10162515-20181225-P00003
Figure US10162515-20181225-P00004
”, as shown in FIG. 3. Then, the touch operation of the user is detected by a sensor of the screen of the tablet computer.
After step S101 is performed, step S102 is performed in the method according to the embodiment of the present disclosure, i.e., N of the M display objects are controlled to be in a selected state in response to the first touch operation, and a first operation instruction for implementing a first operation on the N display objects is generated and executed, wherein N is a positive integer less than or equal to M.
In the embodiment of the present disclosure, step S102 may be divided into two steps. In a first step, N of the M display objects are controlled to be in a selected state in response to the first touch operation, and in a second step, a first operation instruction for implementing a first operation on the N display objects is generated and executed. With reference to FIG. 4, a specific implementation of the first step in the embodiment of the present disclosure comprises the following steps.
In S10211, a touch start point and a touch end point of the first touch operation are acquired in response to the first touch operation;
in S10212, a first touch trajectory of the first touch operation is determined based on the touch start point and the touch end point;
in S10213, the N display objects are determined from the M display objects based on the first touch trajectory; and
in S10214, the N display objects are controlled to be in the selected state.
In a specific implementation, by still taking the above example, after the touch operation of the user is detected on the screen of the tablet computer, an operation start point and an operation end point of the touch operation are acquired. For example, when the touch operation is a circle drawing operation, the touch start point and the touch end point are at ⅓ of the character “
Figure US10162515-20181225-P00005
” from left to right, i.e., the operation start point is the same as the operation end point. When the touch operation is a transverse line drawing operation, the touch start point is at ½ of the character “
Figure US10162515-20181225-P00006
” from left to right, and the touch end point is at ¼ of the character “
Figure US10162515-20181225-P00007
” from left to right, i.e., the touch start point is at a distance from the touch end point.
After the step is performed, step S10212 is performed in the method according to the embodiment of the present disclosure, i.e., a first touch trajectory of the first touch operation is determined based on the touch start point and the touch end point.
In the embodiment of the present disclosure, with reference to FIG. 5, a specific implementation of step S10212 comprises the following steps.
In S102121, it is judged whether the touch start point is the same as the touch end point and there is another touch point between the touch start point and the touch end point to acquire a judgment result;
in S102122, it is determined that the touch trajectory of the first touch operation is a closed touch trajectory having a first closed region on the touch display unit when the judgment result is yes; and
in S102123, it is determined that the touch trajectory of the first touch operation is a non-closed touch trajectory when the judgment result is no.
In a specific implementation, by still taking the above example, the touch operation has been detected on the screen of the tablet computer. In this case, by taking the touch operation being a circle drawing operation as an example, as the touch start point and the touch end point are at ⅓ of the character “
Figure US10162515-20181225-P00008
” from left to right, the tablet computer judges that the touch start point and the touch end point of the touch operation are the same. Further, if it is judged that there is another touch point between the touch start point and the touch end point, for example, a touch point at the rightmost end of the trajectory of the circle drawing operation is at ½ of the character “
Figure US10162515-20181225-P00009
” from left to right, it is determined that the circle drawing touch operation forms a closed touch trajectory. Alternatively, by taking the touch operation being a transverse line drawing operation as an example, as the touch start point is at ½ of the character “
Figure US10162515-20181225-P00010
” from left to right and the touch end point is at ¼ of the character “
Figure US10162515-20181225-P00011
” from left to right, the tablet computer judges that the touch start point and the touch end point are not in the same position, and thus determines that the transverse line drawing touch operation forms a non-closed touch trajectory.
After step S10212 is performed, step S10213 is performed in the method according to the embodiment of the present disclosure, i.e., the N display objects are determined from the M display objects based on the first touch trajectory.
In the embodiment of the present disclosure, a specific implementation of step S10213 comprises:
determining the N display objects from the M display objects, wherein the N display objects at least comprise at least one of the M display objects which is located in the first closed region; or
determining that at least one of the M display objects which is on the non-closed touch trajectory belongs to the N display objects.
In a specific implementation, by still taking the above example, when the tablet computer determines that the touch trajectory is a closed touch trajectory, it is determined that characters surrounded by the closed trajectory belong to the N display objects. In the specific implementation, a character surrounded by the closed trajectory is necessarily one of the N objects, such as three characters “
Figure US10162515-20181225-P00012
” illustrated in the figure. As for characters on the boundary of the trajectory, such as characters “
Figure US10162515-20181225-P00013
” and “
Figure US10162515-20181225-P00014
” illustrated in the figure, the tablet computer may be configured to determine that a character is one of the N objects only if the ⅔ of the character is surrounded by the closed trajectory. In this case, as the touch start point is at ⅓ of the character “
Figure US10162515-20181225-P00015
” from left to right, ⅔ of the character is necessarily surrounded by the closed region, and thus it is determined that “
Figure US10162515-20181225-P00016
” is one of the N objects. As the touch point at the rightmost end of the touch trajectory is at ½ of the character “
Figure US10162515-20181225-P00017
” from left to right, only ½ of the character “
Figure US10162515-20181225-P00018
” is surrounded by the closed region, and thus the character “
Figure US10162515-20181225-P00019
” is not one of the N objects. As a result, the tablet computer determines that the N objects are “
Figure US10162515-20181225-P00020
”. Of course, which character is one of the N objects may be set by those skilled in the art according to practical usage requirements, for example, it may be determined that any character on the touch trajectory is one of the N objects, which is not limited in the present disclosure.
When the tablet computer determines that the touch trajectory is a non-closed touch trajectory, the tablet computer may be configured to determine that any character on the touch trajectory is one of the N objects. As the touch start point is at ½ of the character “
Figure US10162515-20181225-P00021
” from left to right and the touch end point is at ¼ of the character “
Figure US10162515-20181225-P00022
” from left to right, characters on the touch trajectory are “
Figure US10162515-20181225-P00023
”, and thus it is determined that the N objects are “
Figure US10162515-20181225-P00024
Figure US10162515-20181225-P00025
”. Of course, which character is one of the N objects may be set by those skilled in the art according to practical usage requirements, for example, it may be determined that a character is one of the N objects only if ½ of the character is surrounded by the touch trajectory, which is not limited in the present disclosure.
After step S10213 is performed, step S10214 is performed in the method according to the embodiment of the present disclosure, i.e., the N display objects are controlled to be in the selected state.
In the embodiment of the present disclosure, with reference to FIG. 6, a specific implementation of step S10214 comprises the following steps.
In S102141, touch duration in which the operation body keeps in a contact state with the touch display unit at the touch end point is acquired;
in S102142, it is detected whether the touch duration is greater than or equal to predetermined duration to acquire a detection result; and
in S102143, the N display units are controlled to be in the selected state when the detection result is yes.
In a specific implementation, by still taking the above example, after the N objects are determined through the touch trajectory, the tablet computer detects touch duration corresponding to the end point of the touch operation, for example, 2 s. Predetermined duration may be specifically set to for example 1s or 2 s by a user or those skilled in the art according to practical usage conditions. By taking the predetermined duration being 1 s as an example, the tablet computer judges that the touch duration 2 s of the touch end point is greater than 1 s, then selects the N objects “
Figure US10162515-20181225-P00026
” or “
Figure US10162515-20181225-P00027
”, and feeds the content of the characters which are currently selected back to the user on the screen of the tablet computer, for example, by adding a shadow on “
Figure US10162515-20181225-P00028
Figure US10162515-20181225-P00029
”, as shown in FIG. 7, or changing a font color of “
Figure US10162515-20181225-P00030
”, which is not limited in the present disclosure.
After the N objects are selected by the tablet computer, before the stylus is lifted up, a touch operation of moving to the left or to the right in a position of the end point of the touch operation is further detected. By taking the touch operation being a circle drawing operation as an example, after the tablet computer selects “
Figure US10162515-20181225-P00031
”, the tablet computer detects a touch operation of moving to the left at the touch end point, i.e., there is a touch operation of moving a character to the left at ½ of the character “
Figure US10162515-20181225-P00032
” from left to right. In this case, the entire selected bar is moved by the tablet computer to the left for a distance of one character, and then “:
Figure US10162515-20181225-P00033
” are selected by the tablet computer, as shown in FIG. 8A; or the leftmost end of the selected bar is moved by the tablet computer to the left for a distance of one character, and then “:
Figure US10162515-20181225-P00034
Figure US10162515-20181225-P00035
” are selected by the tablet computer, as shown in FIG. 8B. Correspondingly, when it is detected that there is a touch operation of moving to the right at the touch end point, the entire selected bar is moved by the tablet computer to the right, and then “
Figure US10162515-20181225-P00036
” are selected by the tablet computer, as shown in FIG. 9A; or the rightmost end of the selected bar is moved by the tablet computer to the right, and then “
Figure US10162515-20181225-P00037
” are selected by the tablet computer, as shown in FIG. 9B. In a specific implementation, this may be set by those skilled in the art according to practical usage requirements, which is not limited in the present disclosure.
After the first step is performed, the second step is performed in the method according to the embodiment of the present disclosure, i.e., the first operation instruction for implementing a first operation on the N display objects is generated and executed.
In the embodiment of the present disclosure, as there may be different touch trajectories, the second step may be specifically implemented in the following two manners.
First Manner:
When the first touch trajectory is the closed touch trajectory, with reference to FIG. 10, the second step specifically comprises the following steps.
In S102211, a first touch direction of the first touch operation is acquired; and
In S102212, the first operation instruction for implementing a first operation on the N display objects is generated and executed at least based on the first touch direction.
The first operation instruction is specifically:
a deletion instruction of deleting the N display objects; or
a copy instruction of copying the N display objects; or
a cut instruction of cutting the N display objects; or
a share instruction of sharing the N display objects; or
a selected state cancel instruction of cancel the selected state of the N display objects.
In a specific implementation, as the circle drawing trajectory is a trajectory from a touch start point “
Figure US10162515-20181225-P00038
” to a touch point “
Figure US10162515-20181225-P00039
” and then back to “
Figure US10162515-20181225-P00040
”, the tablet computer judges that the direction of the trajectory is a clockwise direction. A user may preset an operation instruction corresponding to the direction of the trajectory in the tablet computer according to practical usage requirements, for example, a copy operation instruction corresponding to a clockwise circle drawing operation, a cut operation instruction corresponding to an anticlockwise circle drawing operation or the like. In the embodiment of the present disclosure, the operation instructions may comprise a deletion operation instruction, a copy operation instruction, a cut operation instruction, a share operation instruction, and a cancel operation instruction. Of course, there may be other operation instructions, which are not limited in the embodiment of the present disclosure. In this case, the tablet computer judges that the direction of the circle drawing trajectory is a clockwise direction, and determines that the operation instruction corresponding to the clockwise direction is a copy operation from the preset operation instructions. Then, the tablet computer generates an operation instruction of copying “
Figure US10162515-20181225-P00041
”. When the touch operation ends, the tablet computer achieves an operation of copying “
Figure US10162515-20181225-P00042
”. When the user implements a clockwise circle drawing operation on “
Figure US10162515-20181225-P00043
”, and does not want to implement any operation on the four characters before the stylus is lifted up, the user may implement a slide operation in an upward direction or a downward direction. When the tablet computer detects that the touch operation is an operation of firstly drawing a circle in a clockwise direction and then sliding at the end point of the circle in an upward direction or a downward direction, it is determined that the operation instruction corresponding to the current operation is an operation instruction of canceling a selected state of “
Figure US10162515-20181225-P00044
”.
Second Manner:
When the first touch trajectory is the non-closed touch trajectory, the second step specifically comprises:
generating and executing the first operation instruction for implementing a first operation on the N display objects based on the non-closed touch trajectory.
The first operation instruction is specifically:
a deletion instruction of deleting the N display objects; or
a copy instruction of copying the N display objects; or
a cut instruction of cutting the N display objects; or
a share instruction of sharing the N display objects; or
a selected state cancel instruction of cancel the selected state of the N display objects.
In a specific implementation, as the transverse line drawing trajectory is a trajectory from a character “
Figure US10162515-20181225-P00045
” at a touch start point to a character “
Figure US10162515-20181225-P00046
” at a touch end point in a rightward direction, the tablet computer judges that the direction of the trajectory is a left-to-right direction. A user may preset an operation instruction corresponding to the direction of the trajectory in the tablet computer according to practical usage requirements, for example, a copy operation instruction corresponding to a left-to-right transverse line drawing operation, a cut operation instruction corresponding to a right-to-left transverse line drawing operation or the like. In the embodiment of the present disclosure, the operation instructions may comprise a deletion operation instruction, a copy operation instruction, a cut operation instruction, a share operation instruction, and a cancel operation instruction. Of course, there may be other operation instructions, which are not limited in the embodiment of the present disclosure. In this case, the tablet computer judges that the direction of the transverse line drawing trajectory is a left-to-right direction, and determines that the operation instruction corresponding to the left-to-right direction is a copy operation from the preset operation instructions. Then, the tablet computer generates an operation instruction of copying “
Figure US10162515-20181225-P00047
”. When the touch operation ends, the tablet computer achieves an operation of copying “
Figure US10162515-20181225-P00048
”. When the user implements a left-to-right transverse line drawing operation on “
Figure US10162515-20181225-P00049
”, and does not want to implement any operation on the five characters before the stylus is lifted up, the user may implement a slide operation in an upward direction or a downward direction. When the tablet computer detects that the touch operation is an operation of firstly drawing a transverse line in a left-to-right direction and then sliding at the end point of the transverse line in an upward direction or a downward direction, it is determined that the operation instruction corresponding to the current operation is an operation instruction of canceling a selected state of “
Figure US10162515-20181225-P00050
”.
Second Embodiment
With reference to FIG. 11, based on the same inventive concept as that of the first embodiment of the present disclosure, the second embodiment of the present disclosure provides an electronic device, comprising:
a housing 10;
a touch display unit 20 arranged in the housing 10; and
a processor 30 arranged in the housing 10 and comprising hardware configured to detect and acquire a first touch operation on the touch display unit of the electronic device when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1, control N of the M display objects to be in a selected state in response to the first touch operation, and generate and execute a first operation instruction for implementing a first operation on the N display objects, wherein N is a positive integer less than or equal to M.
Alternatively, in the second embodiment of the present disclosure, the hardware in the processor 30 is further configured to:
acquire a touch start point and a touch end point of the first touch operation in response to the first touch operation;
determine a first touch trajectory of the first touch operation based on the touch start point and the touch end point;
determine the N display objects from the M display objects based on the first touch trajectory; and
control the N display objects to be in the selected state.
Alternatively, in the second embodiment of the present disclosure, the hardware in the processor 30 is further configured to:
acquire touch duration in which the operation body keeps in a contact state with the touch display unit at the touch end point;
detect whether the touch duration is greater than or equal to predetermined duration to acquire a detection result; and
control the N display units to be in the selected state when the detection result is yes.
Alternatively, in the second embodiment of the present disclosure, the hardware in the processor 30 is further configured to:
judge whether the touch start point is the same as the touch end point and there is another touch point between the touch start point and the touch end point to acquire a judgment result;
determine that the touch trajectory of the first touch operation is a closed touch trajectory having a first closed region on the touch display unit when the judgment result is yes; and
determine that the touch trajectory of the first touch operation is a non-closed touch trajectory when the judgment result is no.
Alternatively, in the second embodiment of the present disclosure, the hardware in the processor 30 is further configured to:
determine the N display units from the M display objects, wherein the N display objects at least comprise at least one of the M display objects which is located in the first closed region; or
determine that at least one of the M display objects which is on the non-closed touch trajectory belongs to the N display objects.
Alternatively, in the second embodiment of the present disclosure, the hardware in the processor 30 is further configured to:
acquire a first touch direction of the first touch operation; and
generate and execute the first operation instruction for implementing a first operation on the N display objects at least based on the first touch direction.
Alternatively, in the second embodiment of the present disclosure, the hardware in the processor 30 is further configured to:
generate and execute the first operation instruction for implementing a first operation on the N display objects based on the non-closed touch trajectory.
The electronic device described above corresponds to the information processing method according to the first embodiment one by one, and therefore will not be described in detail here.
Third Embodiment
With reference to FIG. 12, based on the same inventive concept as that of the first embodiment of the present disclosure, the third embodiment of the present disclosure provides an electronic device, comprising:
a first detection unit 101 configured to detect and acquire a first touch operation on a touch display unit of the electronic device when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1; and
a first control unit 102 configured to control N of the M display objects to be in a selected state in response to the first touch operation, and generate and execute a first operation instruction for implementing a first operation on the N display objects, wherein N is a positive integer less than or equal to M.
In the third embodiment of the present disclosure, the first control unit 102 specifically comprises:
a first acquisition sub-unit configured to acquire a touch start point and a touch end point of the first touch operation in response to the first touch operation;
a first determination sub-unit configured to determine a first touch trajectory of the first touch operation based on the touch start point and the touch end point;
a second determination sub-unit configured to determine the N display objects from the M display objects based on the first touch trajectory; and
a first control sub-unit configured to control the N display objects to be in the selected state.
In the third embodiment of the present disclosure, the first control sub-unit specifically comprises:
a first acquisition module configured to acquire touch duration in which the operation body keeps in a contact state with the touch display unit at the touch end point;
a first detection module configured to detect whether the touch duration is greater than or equal to predetermined duration to acquire a detection result; and
a first control module configured to control the N display units to be in the selected state when the detection result is yes.
In the third embodiment of the present disclosure, the first determination sub-unit specifically comprises:
a first judgment module configured to judge whether the touch start point is the same as the touch end point and there is another touch point between the touch start point and the touch end point to acquire a judgment result;
a first determination module configured to determine that the touch trajectory of the first touch operation is a closed touch trajectory having a first closed region on the touch display unit when the judgment result is yes; and
a second determination module configured to determine that the touch trajectory of the first touch operation is a non-closed touch trajectory when the judgment result is no.
In the third embodiment of the present disclosure, the second determination sub-unit specifically comprises:
a third determination module configured to determine the N display objects from the M display objects, wherein the N display objects at least comprise at least one of the M display objects which is located in the first closed region; or
a fourth determination module configured to determine that at least one of the M display objects which is on the non-closed touch trajectory belongs to the N display objects.
In the third embodiment of the present disclosure, when the first touch trajectory is the closed touch trajectory, the first control unit 102 specifically comprises:
a second acquisition sub-unit configured to acquire a first touch direction of the first touch operation; and
a first execution sub-unit configured to generate and execute the first operation instruction for implementing a first operation on the N display objects at least based on the first touch direction.
In the third embodiment of the present disclosure, when the first touch trajectory is the non-closed touch trajectory, the first control unit 102 specifically comprises:
a second execution sub-unit configured to generate and execute the first operation instruction for implementing a first operation on the N display objects based on the non-closed touch trajectory.
As the above electronic device corresponds to the information processing method in the first embodiment one by one, the electronic device will not be described in detail here.
Those skilled in the art should appreciate that the embodiments of the present disclosure can be provided as methods, systems, or computer program products. Therefore, forms such as hardware-only embodiments, software-only embodiments, or embodiments combining software and hardware can be used in the present disclosure. In addition, forms such as a computer program product which is implemented on one or more of computer usable storage media (comprising but not limited to a disk memory, a CD-ROM, an optical memory etc.) with computer usable program codes can be used in the present disclosure.
The present disclosure is described with reference to the flowcharts and/or block diagrams of the methods, devices (systems) and computer program products according to the embodiments of the present disclosure. It should be understood that each flow and/or block in the flowcharts and/or block diagrams as well as a combination of the flows and/or blocks in the flowcharts and/or block diagrams can be implemented by computer program instructions. These computer program instructions can be provided to a processor of a general-purpose computer, a dedicated-purpose computer, an embedded processor, or other programmable data processing devices to generate a machine. Thereby, the instructions executed by the processor of the computer or other programmable data processing devices generate means for implementing functions specified in one or more flows in the flowcharts and/or one or more blocks in the block diagrams.
These computer program instructions can also be stored in a computer readable memory capable of introducing a computer or other programmable data processing devices to operate in a particular mode. Thereby, the instructions stored in the computer readable memory generate an article of manufacture comprising instruction means for implementing functions specified in one or more flows in the flowcharts and/or one or more blocks in the block diagrams.
These computer program instructions can also be loaded onto a computer or other programmable data processing devices, so as to enable a series of operation steps to be performed on the computer or other programmable devices to generate a computer-implemented process. Thereby, the instructions executed on the computer or other programmable devices provide a step of implementing functions specified in one or more flows in the flowcharts and/or one or more blocks in the block diagrams.
Specifically, computer program instructions corresponding to the information processing method according to the embodiments of the present disclosure may be stored on a storage medium such as an optical disk, a hard disk, a USB device etc. When the computer program instructions in the storage medium corresponding to the information processing method are read or executed by an electronic device, the following steps are performed:
detecting and acquiring a first touch operation on a touch display unit of the electronic device when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1; and
controlling N of the M display objects to be in a selected state in response to the first touch operation, and generating and executing a first operation instruction for implementing a first operation on the N display objects, wherein N is a positive integer less than or equal to M.
Alternatively, when computer instructions stored in the storage medium, which correspond to the step of controlling N of the M display objects to be in a selected state in response to the first touch operation, are executed, the following steps are further performed:
acquiring a touch start point and a touch end point of the first touch operation in response to the first touch operation;
determining a first touch trajectory of the first touch operation based on the touch start point and the touch end point;
determining the N display objects from the M display objects based on the first touch trajectory; and
controlling the N display objects to be in the selected state.
Alternatively, when computer instructions stored in the storage medium, which correspond to the step of controlling the N display objects to be in the selected state, are executed, the following steps are further performed:
acquiring touch duration in which the operation body keeps in a contact state with the touch display unit at the touch end point;
detecting whether the touch duration is greater than or equal to predetermined duration to acquire a detection result; and
controlling the N display units to be in the selected state when the detection result is yes.
Alternatively, when computer instructions stored in the storage medium, which correspond to the step of determining a first touch trajectory of the first touch operation based on the touch start point and the touch end point, are executed, the following steps are further performed:
judging whether the touch start point is the same as the touch end point and there is another touch point between the touch start point and the touch end point to acquire a judgment result;
determining that the touch trajectory of the first touch operation is a closed touch trajectory having a first closed region on the touch display unit when the judgment result is yes; and
determining that the touch trajectory of the first touch operation is a non-closed touch trajectory when the judgment result is no.
Alternatively, when computer instructions stored in the storage medium, which correspond to the step of determining the N display objects from the M display objects based on the first touch trajectory, are executed, the following steps are further performed:
determining the N display objects from the M display objects, wherein the N display objects at least comprise at least one of the M display objects which is located in the first closed region; or
determining that at least one of the M display objects which is on the non-closed touch trajectory belongs to the N display objects.
Alternatively, when the first touch trajectory is the closed touch trajectory, when computer instructions stored in the storage medium, which correspond to the step of generating and executing a first operation instruction for implementing a first operation on the N display objects, are executed, the following steps are further performed:
acquiring a first touch direction of the first touch operation; and
generating and executing the first operation instruction for implementing a first operation on the N display objects at least based on the first touch direction.
Alternatively, when the first touch trajectory is the non-closed touch trajectory, when computer instructions stored in the storage medium, which correspond to the step of generating and executing a first operation instruction for implementing a first operation on the N display objects, are executed, the following steps are further performed:
generating and executing the first operation instruction for implementing a first operation on the N display objects based on the non-closed touch trajectory.
Alternatively, when computer instructions stored in the storage medium, which correspond to the first operation instruction, are executed, the following steps are further performed:
a deletion instruction of deleting the N display objects; or
a copy instruction of copying the N display objects; or
a cut instruction of cutting the N display objects; or
a share instruction of sharing the N display objects; or
a selected state canceling instruction of canceling the selected state of the N display objects.
Although preferable embodiments of the present disclosure have been described, additional changes and modifications can be made to these embodiments by those skilled in the art upon learning the basic creative concepts. Therefore, the appended claims are intended to be construed as comprising the preferable embodiments and all changes and modifications that fall into the scope of the present disclosure.
Obviously, various modifications and variants can be made to the present disclosure by those skilled in the art without departing from the spirit and scope of the present disclosure. Therefore, these modifications and variants are to be encompassed by the present disclosure if they fall within the scope of the present disclosure as defined by the claims and their equivalents.

Claims (13)

We claim:
1. An information processing method, comprising:
detecting and acquiring a first touch operation on a touch display unit of the electronic device when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1; and
controlling N of the M display objects to be in a selected state in response to the first touch operation, wherein a first operation instruction for implementing a first operation on the N display objects is generated and executed based on a first touch direction of the first touch operation, wherein a second operation instruction for implementing a second operation on the N display objects is generated and executed based on a second touch direction of the first touch operation, wherein the first touch direction of the first touch operation is different from the second touch direction of the first touch operation, wherein the first operation instruction is different from the second operation instruction, wherein N is a positive integer less than or equal to M, and wherein each of the first and second operation instructions is one of a deletion instruction of deleting the N display objects, a copy instruction of copying the N display objects, a cut instruction of cutting the N display objects, or a share instruction of sharing the N display objects.
2. The method according to claim 1, wherein the first touch operation is specifically a touch operation implemented by an operation body on the touch display unit, and the operation body is specifically a finger of a user or a stylus.
3. The method according to claim 2, wherein controlling N of the M display objects to be in a selected state in response to the first touch operation comprises:
acquiring a touch start point and a touch end point of the first touch operation in response to the first touch operation;
determining a first touch trajectory of the first touch operation based on the touch start point and the touch end point;
determining the N display objects from the M display objects based on the first touch trajectory; and
controlling the N display objects to be in the selected state.
4. The method according to claim 3, wherein controlling the N display objects to be in the selected state comprises:
acquiring a touch duration in which the operation body keeps in a contact state with the touch display unit at the touch end point;
detecting whether the touch duration is greater than or equal to predetermined duration to acquire a detection result; and
controlling the N display units to be in the selected state when the detection result is yes.
5. The method according to claim 4, wherein determining a first touch trajectory of the first touch operation based on the touch start point and the touch end point comprises:
judging whether the touch start point is the same as the touch end point and there is another touch point between the touch start point and the touch end point to acquire a judgment result; and
determining that the touch trajectory of the first touch operation is a closed touch trajectory having a first closed region on the touch display unit when the judgment result is yes, and determining that the touch trajectory of the first touch operation is a non-closed touch trajectory when the judgment result is no.
6. The method according to claim 5, wherein determining the N display objects from the M display objects based on the first touch trajectory comprises:
determining the N display objects from the M display objects, wherein the N display objects at least comprise at least one of the M display objects which is located in the first closed region, or
determining that at least one of the M display objects which is on the non-closed touch trajectory belongs to the N display objects.
7. The method according to claim 1, further comprising:
generating and executing a second operation instruction for implementing a second operation on the N display objects based on a second touch direction after the first touch direction of the first touch operation, wherein the second operation instruction is an instruction to cancel the selected state of the N display objects.
8. An electronic device, comprising:
a housing;
a touch display unit arranged in the housing; and
a processor arranged in the housing and comprising hardware configured to detect and acquire a first touch operation on the touch display unit of the electronic device when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1, control N of the M display objects to be in a selected state in response to the first touch operation, wherein a first operation instruction for implementing a first operation on the N display objects is generated and executed based on a first touch direction of the first touch operation, wherein a second operation instruction for implementing a second operation on the N display objects is generated and executed based on a second touch direction of the first touch operation, wherein the first touch direction of the first touch operation is different from the second touch direction of the first touch operation, wherein the first operation instruction is different from the second operation instruction, wherein N is a positive integer less than or equal to M, and wherein each of the first and second operation instructions is one of a deletion instruction of deleting the N display objects, a copy instruction of copying the N display objects, a cut instruction of cutting the N display objects, or a share instruction of sharing the N display objects.
9. The electronic device according to claim 8, wherein the hardware is further configured to:
acquire a touch start point and a touch end point of the first touch operation in response to the first touch operation;
determine a first touch trajectory of the first touch operation based on the touch start point and the touch end point;
determine the N display objects from the M display objects based on the first touch trajectory; and
control the N display objects to be in the selected state.
10. The electronic device according to claim 9, wherein the hardware is further configured to:
acquire a touch duration in which the operation body keeps in a contact state with the touch display unit at the touch end point;
detect whether the touch duration is greater than or equal to predetermined duration to acquire a detection result; and
control the N display units to be in the selected state when the detection result is yes.
11. The electronic device according to claim 10, wherein the hardware is further configured to:
judge whether the touch start point is the same as the touch end point and there is another touch point between the touch start point and the touch end point to acquire a judgment result;
determine that the touch trajectory of the first touch operation is a closed touch trajectory having a first closed region on the touch display unit when the judgment result is yes; and
determine that the touch trajectory of the first touch operation is a non-closed touch trajectory when the judgment result is no.
12. The electronic device according to claim 11, wherein the hardware is further configured to:
determine the N display units from the M display objects, wherein the N display objects at least comprise at least one of the M display objects which is located in the first closed region; or
determine that at least one of the M display objects which is on the non-closed touch trajectory belongs to the N display objects.
13. An electronic device, comprising:
a processor comprising hardware configured to
detect and acquire a first touch operation on a touch display unit of the electronic device when M display objects are displayed on the touch display unit, wherein M is an integer greater than or equal to 1; and
control N of the M display objects to be in a selected state in response to the first touch operation, wherein a first operation instruction for implementing a first operation on the N display objects is generated and executed based on a first touch direction of the first touch operation, wherein a second operation instruction for implementing a second operation on the N display objects is generated and executed based on a second touch direction of the first touch operation, wherein the first touch direction of the first touch operation is different from the second touch direction of the first touch operation, wherein the first operation instruction is different from the second operation instruction, wherein N is a positive integer less than or equal to M, and wherein each of the first and second operation instructions is one of a deletion instruction of deleting the N display objects, a copy instruction of copying the N display objects, a cut instruction of cutting the N display objects, or a share instruction of sharing the N display objects.
US14/858,332 2015-05-26 2015-09-18 Method and electronic device for controlling display objects on a touch display based on a touch directional touch operation that both selects and executes a function Active 2036-02-12 US10162515B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
CN201510275191.7 2015-05-26
CN201510275191.7A CN106293433A (en) 2015-05-26 2015-05-26 A kind of information processing method and electronic equipment
CN201510275191 2015-05-26

Publications (2)

Publication Number Publication Date
US20160349982A1 US20160349982A1 (en) 2016-12-01
US10162515B2 true US10162515B2 (en) 2018-12-25

Family

ID=57281986

Family Applications (1)

Application Number Title Priority Date Filing Date
US14/858,332 Active 2036-02-12 US10162515B2 (en) 2015-05-26 2015-09-18 Method and electronic device for controlling display objects on a touch display based on a touch directional touch operation that both selects and executes a function

Country Status (3)

Country Link
US (1) US10162515B2 (en)
CN (1) CN106293433A (en)
DE (1) DE102015117659A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109032465B (en) * 2018-05-29 2020-09-22 Oppo广东移动通信有限公司 Data processing method and device and mobile terminal
CN111008080A (en) * 2018-10-08 2020-04-14 中兴通讯股份有限公司 Information processing method, device, terminal equipment and storage medium

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110163944A1 (en) * 2010-01-05 2011-07-07 Apple Inc. Intuitive, gesture-based communications with physics metaphors
US20110258537A1 (en) * 2008-12-15 2011-10-20 Rives Christopher M Gesture based edit mode
US20120188164A1 (en) * 2009-10-16 2012-07-26 Prasenjit Dey Gesture processing
US20130167019A1 (en) * 2010-10-15 2013-06-27 Sharp Kabushiki Kaisha Information-processing device and control method for information-processing device
US20140109004A1 (en) 2012-10-12 2014-04-17 Cellco Partnership D/B/A Verizon Wireless Flexible selection tool for mobile devices
US20140123049A1 (en) * 2012-10-30 2014-05-01 Microsoft Corporation Keyboard with gesture-redundant keys removed
US20140223345A1 (en) * 2013-02-04 2014-08-07 Samsung Electronics Co., Ltd. Method for initiating communication in a computing device having a touch sensitive display and the computing device
US20140325410A1 (en) * 2013-04-26 2014-10-30 Samsung Electronics Co., Ltd. User terminal device and controlling method thereof
US20150135112A1 (en) 2013-11-08 2015-05-14 Microsoft Corporation Two step content selection
US20160070460A1 (en) * 2014-09-04 2016-03-10 Adobe Systems Incorporated In situ assignment of image asset attributes
US20160117141A1 (en) * 2014-10-22 2016-04-28 Lg Electronics Inc. Watch type terminal and method for controlling the same
US20170147199A1 (en) * 2013-09-13 2017-05-25 BK Company Ltd. Method for changing user-originating information through interaction between mobile device and information display device

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106445322A (en) * 2012-11-26 2017-02-22 中兴通讯股份有限公司 Text processing method and terminal
CN103645857B (en) * 2013-12-27 2018-08-31 联想(北京)有限公司 A kind of method and electronic equipment of control electronics
CN103885696A (en) * 2014-03-17 2014-06-25 联想(北京)有限公司 Information processing method and electronic device
CN104133627A (en) * 2014-07-28 2014-11-05 联想(北京)有限公司 Zooming display method and electronic equipment

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110258537A1 (en) * 2008-12-15 2011-10-20 Rives Christopher M Gesture based edit mode
US20120188164A1 (en) * 2009-10-16 2012-07-26 Prasenjit Dey Gesture processing
US20110163944A1 (en) * 2010-01-05 2011-07-07 Apple Inc. Intuitive, gesture-based communications with physics metaphors
US20130167019A1 (en) * 2010-10-15 2013-06-27 Sharp Kabushiki Kaisha Information-processing device and control method for information-processing device
US20140109004A1 (en) 2012-10-12 2014-04-17 Cellco Partnership D/B/A Verizon Wireless Flexible selection tool for mobile devices
US20140123049A1 (en) * 2012-10-30 2014-05-01 Microsoft Corporation Keyboard with gesture-redundant keys removed
US20140223345A1 (en) * 2013-02-04 2014-08-07 Samsung Electronics Co., Ltd. Method for initiating communication in a computing device having a touch sensitive display and the computing device
US20140325410A1 (en) * 2013-04-26 2014-10-30 Samsung Electronics Co., Ltd. User terminal device and controlling method thereof
US20170147199A1 (en) * 2013-09-13 2017-05-25 BK Company Ltd. Method for changing user-originating information through interaction between mobile device and information display device
US20150135112A1 (en) 2013-11-08 2015-05-14 Microsoft Corporation Two step content selection
US20160070460A1 (en) * 2014-09-04 2016-03-10 Adobe Systems Incorporated In situ assignment of image asset attributes
US20160117141A1 (en) * 2014-10-22 2016-04-28 Lg Electronics Inc. Watch type terminal and method for controlling the same

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
"German Application Serial No. 10 2015 117 659.9, Office Action dated Feb. 28, 2018", w/English Translation, (dated Feb. 28, 2018), 15 pgs.

Also Published As

Publication number Publication date
CN106293433A (en) 2017-01-04
US20160349982A1 (en) 2016-12-01
DE102015117659A1 (en) 2016-12-01

Similar Documents

Publication Publication Date Title
US11681866B2 (en) Device, method, and graphical user interface for editing screenshot images
US11783117B2 (en) Device, method, and graphical user interface for sharing a content object in a document
US11487426B2 (en) Apparatus and method for cursor control and text selection and editing based on gesture-based touch inputs received in a virtual keyboard display area
US9939992B2 (en) Methods and systems for navigating a list with gestures
CN106575196B (en) Electronic device and method for displaying user interface thereof
US9619435B2 (en) Methods and apparatus for modifying typographic attributes
US9851896B2 (en) Edge swiping gesture for home navigation
EP2577435B1 (en) A user interface for a touch sensitive display on an electronic device
US20140098142A1 (en) System and method for generation and manipulation of a curve in a dynamic graph based on user input
US10282219B2 (en) Consolidated orthogonal guide creation
US11340755B2 (en) Moving a position of interest on a display
WO2017023844A1 (en) User interface for a touch screen device in communication with a physical keyboard
EP3437035A1 (en) Techniques to control notifications for content items in a collaboration platform
EP2650797A2 (en) Electronic device and method for annotating data
US10162515B2 (en) Method and electronic device for controlling display objects on a touch display based on a touch directional touch operation that both selects and executes a function
TWI416369B (en) Data selection methods and systems, and computer program products thereof
US9460362B2 (en) Method and apparatus for identifying a desired object of an image using a suggestive marking
CN104423885A (en) Information processing device and control method
US9176940B2 (en) System and method for text editor text alignment control
WO2020227402A1 (en) Systems and methods for delete key gestures and providing continuous-path gestures at a touch-sensitive keyboard
CN104391834A (en) Method for annotating electronic book content
KR20160010993A (en) Object editing method and image display device using thereof
US20150248229A1 (en) Electronic devices and methods for controlling user interface

Legal Events

Date Code Title Description
AS Assignment

Owner name: BEIJING LENOVO SOFTWARE LTD., CHINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ZHANG, XIAOPING;XIE, QINGDI;ZHANG, XUEYUAN;AND OTHERS;REEL/FRAME:036614/0081

Effective date: 20150908

Owner name: LENOVO (BEIJING) LIMITED, CHINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ZHANG, XIAOPING;XIE, QINGDI;ZHANG, XUEYUAN;AND OTHERS;REEL/FRAME:036614/0081

Effective date: 20150908

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4