CN113703631A - Writing control method and device, electronic equipment and storage medium - Google Patents

Writing control method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN113703631A
CN113703631A CN202111012762.XA CN202111012762A CN113703631A CN 113703631 A CN113703631 A CN 113703631A CN 202111012762 A CN202111012762 A CN 202111012762A CN 113703631 A CN113703631 A CN 113703631A
Authority
CN
China
Prior art keywords
input
interface
layer
writing
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111012762.XA
Other languages
Chinese (zh)
Inventor
丁鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vivo Mobile Communication Co Ltd
Original Assignee
Vivo Mobile Communication Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vivo Mobile Communication Co Ltd filed Critical Vivo Mobile Communication Co Ltd
Priority to CN202111012762.XA priority Critical patent/CN113703631A/en
Publication of CN113703631A publication Critical patent/CN113703631A/en
Priority to PCT/CN2022/114410 priority patent/WO2023030117A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/04812Interaction techniques based on cursor appearance or behaviour, e.g. being affected by the presence of displayed objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/60Editing figures and text; Combining figures or text

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The application discloses a writing control method and device, electronic equipment and a storage medium, and belongs to the technical field of intelligent control. The writing control method comprises the following steps: receiving a first input of a first interface by a touch control pen, wherein the first interface is an interface which does not support writing; displaying a first layer on a first interface in response to a first input; receiving a second input; and displaying an input track corresponding to the second input on the first layer or executing corresponding operation on the first interface based on the input parameter of the second input.

Description

Writing control method and device, electronic equipment and storage medium
Technical Field
The application belongs to the technical field of intelligent control, and particularly relates to a writing control method and device, electronic equipment and a storage medium.
Background
With the wide application of touch screens and the improvement of users' demands for precise control, touch pens have come into play.
The existing stylus can write in a specific area, for example, a writing-supporting interface, and then record a writing track to recognize the writing track as text. Since writing does not require mastering the input method, it is very practical for users who do not like to use a keyboard or are not used to use the Chinese input method. The stylus can also be used for precise drawing in a specific application program interface, for example, for circuit design, CAD design, graphic design, free drawing, etc., and for text and data input.
However, most application interfaces do not support the writing and marking functions, so that users cannot write and mark on the interfaces which do not support writing.
Disclosure of Invention
An embodiment of the present application provides a writing control method, a writing control apparatus, an electronic device, and a storage medium, which can solve a problem that writing cannot be performed on an interface that does not support writing.
In a first aspect, an embodiment of the present application provides a writing control method, where the method includes:
receiving a first input of a first interface by a touch control pen, wherein the first interface is an interface which does not support writing;
displaying a first layer on a first interface in response to a first input;
receiving a second input;
and displaying an input track corresponding to the second input on the first layer or executing corresponding operation on the first interface based on the input parameter of the second input.
In a second aspect, an embodiment of the present application provides a writing control apparatus, including:
the first input receiving module is used for receiving first input of a first interface by a touch control pen; the first interface is an interface which does not support writing;
the layer display module is used for responding to a first input and displaying a first layer on a first interface;
the second input receiving module is used for receiving a second input;
and the input response module is used for displaying an input track corresponding to the second input on the first layer or executing corresponding operation on the first interface based on the input parameter of the second input.
In a third aspect, an embodiment of the present application provides an electronic device, which includes a processor, a memory, and a program or instructions stored on the memory and executable on the processor, and when executed by the processor, the program or instructions implement the steps of the method according to the first aspect.
In a fourth aspect, embodiments of the present application provide a readable storage medium, on which a program or instructions are stored, which when executed by a processor implement the steps of the method according to the first aspect.
In a fifth aspect, an embodiment of the present application provides a chip, where the chip includes a processor and a communication interface, where the communication interface is coupled to the processor, and the processor is configured to execute a program or instructions to implement the method according to the first aspect.
In the embodiment of the application, a first input of a stylus to a first interface is received, a first image layer is displayed on the first interface in response to the first input, a second input is received, and an input track corresponding to the second input is displayed on the first image layer or a corresponding operation is executed on the first interface based on an input parameter of the second input.
Drawings
FIG. 1 is one of the schematic diagrams of a writing control application scenario shown in accordance with an exemplary embodiment;
FIG. 2 is one of the flow diagrams of a writing control method shown in accordance with an exemplary embodiment;
FIG. 3a is a second diagram illustrating a writing control application scenario in accordance with an illustrative embodiment;
FIG. 3b is a third diagram illustrating a writing control application scenario in accordance with an illustrative embodiment;
FIG. 4 illustrates a second flow chart of a writing control method according to an exemplary embodiment;
FIG. 5 is a fourth illustration of a writing control application scenario shown in accordance with an exemplary embodiment;
FIG. 6 is a block diagram illustrating the structure of a writing control apparatus according to an exemplary embodiment;
FIG. 7 is a block diagram illustrating the structure of an electronic device in accordance with an exemplary embodiment;
fig. 8 is a schematic diagram of a hardware structure of an electronic device implementing an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be described clearly below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some, but not all, embodiments of the present application. All other embodiments that can be derived by one of ordinary skill in the art from the embodiments given herein are intended to be within the scope of the present disclosure.
The terms first, second and the like in the description and in the claims of the present application are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It will be appreciated that the data so used may be interchanged under appropriate circumstances such that embodiments of the application may be practiced in sequences other than those illustrated or described herein, and that the terms "first," "second," and the like are generally used herein in a generic sense and do not limit the number of terms, e.g., the first term can be one or more than one. In addition, "and/or" in the specification and claims means at least one of connected objects, a character "/" generally means that a preceding and succeeding related objects are in an "or" relationship.
The writing control method, the writing control apparatus, the electronic device, and the storage medium according to the embodiments of the present application are described in detail below with reference to the accompanying drawings through specific embodiments and application scenarios thereof.
The writing control method provided by the application can be applied to the scene shown in fig. 1, and is specifically described in detail with reference to fig. 1.
FIG. 1 is a schematic diagram illustrating a writing control application scenario in accordance with an illustrative embodiment.
The writing control method provided by the embodiment of the application can be applied to an application scenario in which a user writes on an electronic device, as shown in fig. 1, the writing control method can include a stylus 110 and an electronic device 120. The electronic device 120 may be a mobile phone or a tablet computer, or may be other devices capable of being communicatively connected to the stylus 110. The stylus 110 may be wired or wirelessly connected to the electronic device 120, for example, via a Universal Serial Bus (USB) interface or via bluetooth.
For example, when the user uses the stylus 110 to write on the interface 130 displayed on the electronic device 120, if the interface 130 supports writing, the corresponding writing track may be directly displayed, or the writing track may be recorded and recognized as text and input to the corresponding position, which is not limited herein. If the interface 130 displayed on the electronic device 120 does not support writing, for example, a web page or some Application (APP) interface that does not support writing or marking, writing and marking cannot be performed.
In order to solve the above problem, an embodiment of the present application provides a writing control method, that is, receiving a first input from a stylus to a first interface, where the first interface is an interface that does not support writing; displaying a first layer on a first interface in response to a first input; receiving a second input; and displaying an input track corresponding to the second input on the first layer or executing corresponding operation on the first interface based on the input parameter of the second input.
Therefore, the first input of the stylus to the first interface is received, the first layer is displayed on the first interface in response to the first input, the second input is received, the input track corresponding to the second input is displayed on the first layer or the corresponding operation is executed on the first interface based on the input parameter of the second input, and the input track corresponding to the second input can be displayed on the first layer.
According to the application scenario, the writing control method provided by the embodiment of the application is described in detail below with reference to fig. 2 to 5. In the writing control method provided by the embodiment of the present application, the execution main body may be a writing control device, or a control module of the writing control device for executing the writing control method. In the embodiment of the present application, a writing control method executed by a writing control apparatus is taken as an example, and the writing control method provided in the embodiment of the present application is described.
FIG. 2 is a flow chart illustrating a writing control method according to an exemplary embodiment.
As shown in fig. 2, the writing control method may include the steps of:
step 210, a first input to a first interface by a stylus is received.
In this embodiment of the application, the first interface may be an interface that does not support writing, for example, a web interface and some display interfaces of APPs that do not support writing, and may also be other interfaces that do not support writing, which is not limited herein. The stylus may be in communication connection with the electronic device corresponding to the first interface, for example, the stylus may be in wired connection through a USB interface, or may be in wireless connection through bluetooth. The electronic device may be a mobile phone or a tablet computer, or may be other devices capable of being in communication connection with the stylus. The first input may be an input for instructing entry into a writing mode.
For example, after a user opens a first interface that does not support writing on the electronic device, if the user needs to perform writing operation in the first interface, a first input may be performed on a screen of the electronic device by operating a stylus to enter a writing mode.
Step 220, responding to the first input, displaying the first layer on the first interface.
Optionally, the transparency of the background of the first layer may be greater than 0, so that a user may see the content displayed by the first interface through the first layer.
Step 230, a second input is received.
Here, the second input may be writing input or manipulation input, and the second input may be input by a stylus or input by a finger, which is not limited herein.
For example, the user may use a stylus to write or manipulate on a screen of a tablet computer displaying a web interface, or may use a finger to write or manipulate, which is not limited herein.
And step 240, displaying an input track corresponding to the second input on the first layer or executing corresponding operation on the first interface based on the input parameter of the second input.
Here, the input parameter of the second input may be a gesture of the stylus, and it may be determined through the input parameter that the stylus is in a writing mode or a manipulation mode, so as to determine whether to display an input track corresponding to the second input on the first layer or to perform a corresponding operation on the first interface. For example, when the front end of the stylus is used for inputting, the user can be determined to be in a writing mode, and then an input track is displayed on the first image layer; when the rear end of the touch pen is used for inputting, the touch pen can be determined to be in the control mode, and then the first interface is controlled.
Therefore, the first input of the stylus to the first interface is received, the first layer is displayed on the first interface in response to the first input, the second input is received, the input track corresponding to the second input is displayed on the first layer or the corresponding operation is executed on the first interface based on the input parameter of the second input, the input track corresponding to the second input can be displayed on the first layer, in addition, the first interface can be controlled through the second input, and therefore the user can write and mark on the interface which does not support writing while the user can normally operate the interface.
Based on this, in an alternative embodiment, step 230 may include:
receiving a second input to the first image layer;
based on this, step 240 may include:
and displaying an input track of the second input on the first layer according to the first touch event corresponding to the second input.
Here, the second input may be an input of writing on the screen using a tip of a stylus pen, or a slide input on the screen using a finger. The first touch event may be a touch event for writing, and the first layer may be configured to respond to the touch event, so that a corresponding input track, that is, a writing track, may be displayed on the first layer according to track information of the first touch event, and an input track corresponding to a second input may be saved by saving the first layer.
For example, a writing input of a user on the first layer through a front end of a stylus or a finger may be received, and a first touch event may be generated according to the writing input, and the first touch event may be intercepted by the first layer, so that a writing track corresponding to the first touch event corresponding to the writing input may be displayed in the first layer.
In this way, the first layer is set to respond to the touch event, the input track of the second input can be displayed on the first layer, and therefore the writing track can be displayed on the interface which does not support writing.
Furthermore, in another alternative embodiment, step 230 may include:
receiving a second input to the first interface;
accordingly, step 240 may include:
setting the first layer to not respond to the touch event based on the input parameter of the first sub-input;
transmitting a second touch event corresponding to the second sub-input to the first interface based on the second sub-input;
and executing the operation corresponding to the second sub-input on the first interface according to the second touch event.
Here, the second input may include a first sub-input and a second sub-input, wherein the first sub-input may be an input of a stylus and the input parameter of the first sub-input may be a gesture of the stylus. When a user needs to control the first interface, the electronic equipment can be subjected to first sub-input through the rear end of the touch pen so as to exit the writing mode and enter the control mode.
The second sub-input may be an input operated on the screen by using the rear end of the stylus pen, or an input operated on the screen by using a finger. A second touch event may be generated according to the second sub-input, and the second touch event may be a touch event for performing an interface operation. Because the first layer is set to not respond to the touch event, for example, the setting of the first layer to intercept the touch event is cancelled or the first layer is set to intercept but not respond to the touch event, the second touch event can be transmitted to the first interface through the first layer, so that the control of the first interface is realized. At this time, the existing writing track in the first image layer can be continuously displayed, but the writing mode is exited, and writing cannot be performed.
For example, in a web interface that is displayed on a tablet computer and does not support writing, if a user needs to control the web interface after writing is completed, a stylus may be turned to enable a rear end to input towards a screen of the tablet computer, that is, a first sub-input is performed, after receiving the input, the tablet computer may set a first image layer as a non-response touch event, and then the user may use the rear end of the stylus or directly perform a click or slide operation on the web interface through a finger, that is, perform a second sub-input. After receiving the click or sliding input of the touch pen, the web interface can transmit a second touch event corresponding to the click or sliding input to the web interface through the first layer, so that the click or sliding operation of the web interface is realized.
In addition, the operation mode can be exited after the operation is finished, the writing mode is entered again, the specific writing process is not repeated herein, and after the final writing is finished, the writing mode can be exited completely by clicking a specific button or switching other interfaces, and certainly, the writing mode can be exited completely by other modes without limitation. For example, a "save" button may be clicked to save an existing writing trace and exit the writing mode completely, and accordingly, the existing writing trace is no longer displayed on the first interface.
Therefore, the first layer is set to be not in response to the touch event, the first interface can be controlled through the first layer, and therefore the electronic equipment can be switched from the writing mode to the control mode quickly.
Based on this, in an alternative embodiment, the first input may be an input of the stylus in case of the first posture, and the first sub-input may be an input of the stylus in case of the second posture.
Here, the first posture may be a posture in which an end of the stylus for writing is directed downward, and the second posture may be a posture in which an end of the stylus for writing is directed upward. The writing function and the control function of the touch pen can be switched by turning the touch pen.
For example, as shown in fig. 3a and 3b, an orientation sensor may be disposed in the stylus to detect a gesture of the stylus, and if the stylus tip 310 is downward, the stylus is detected to be in a first gesture, and the stylus may generate a first input and send the first input to the connected electronic device to control the electronic device to enter a writing mode; if the stylus pen cap 320 is downward, it is detected that the stylus pen is in the second posture, and the stylus pen may generate the first sub-input and send the first sub-input to the connected electronic device to control the electronic device to enter the operation mode.
Therefore, the touch pen can be triggered to generate different inputs by detecting the gesture of the touch pen, so that the electronic equipment connected with the electronic equipment is indicated to set the first image layer to respond or not respond to a touch event, different functions of the touch pen are switched rapidly, and user experience is improved.
In addition, in another alternative embodiment, the first input may be an input of the stylus pen when the first end is pressed, and the first sub-input may be an input of the stylus pen when the second end is pressed.
Here, the first end may be an end of the stylus for writing, such as a pen tip; the second end may be an end of the stylus for manipulation, such as a cap.
For example, as shown in fig. 3a and 3b, a pressing sensor may be disposed at both ends of the stylus pen, and if the stylus pen point 310, i.e. the first end, is pressed, accordingly, the stylus pen may perform a first input; if the stylus pen cap 320 is pressed, i.e. the second end, the stylus pen can perform the first sub-input accordingly.
Therefore, different input is performed by pressing different ends of the touch pen, so that the touch pen can be accurately triggered to generate different instructions without being influenced by the using posture of a user. For example, if the user lies down using the stylus pen, it is impossible to determine whether the user wants to write or manipulate by detecting the posture of the stylus pen, and it is possible to determine whether the user wants to write or manipulate by detecting the pressing end.
In addition, a key or a touch area can be arranged on the touch pen, and the user can trigger the touch pen to perform first input or first sub-input by pressing the key or clicking the touch area. For example, pressing the key once, a first input is made; and pressing the key twice to perform the first sub-input.
Based on this, in an optional implementation manner, in a case that the first interface is an interface extending along a first direction, and the second touch event is a sliding touch event along a second direction, the performing, according to the second touch event, an operation corresponding to the second sub-input on the first interface may specifically include:
and controlling the first interface to slide to a first position along the first direction and controlling the first layer to slide to a second position along the first direction according to the second touch event.
Here, the first position of the first interface may correspond to the second position of the first layer, and the first direction may be opposite to the second direction. Under the condition that a sliding input of a user is received, the first interface and the first layer can synchronously slide for the same distance, so that the handwriting on the first layer can always correspond to the content displayed in the first interface.
For example, when the user slides the web interface upwards, that is, in the second direction, using the stylus pen cap or the finger, the web interface and the first layer may be controlled to slide downwards in synchronization, that is, in the first direction, by the same distance.
Therefore, the first interface and the first image layer can slide synchronously through the second touch event corresponding to the second sub-input, and writing and marking on the first image layer can be continued after the first interface is slid.
Based on the above step 210 and 240, in a possible embodiment, as shown in fig. 4, after the step 210, the writing control method may further include: step 211-212, wherein:
and step 211, performing screen capture operation on the first interface to obtain a first image.
Here, the screen capturing operation may be performed at any time after the first input is received and before the first image layer and the second image layer are superimposed to obtain the second image, that is, may be performed at any time after step 210 and before step 241, and step 211, step 212, and step 220 may be performed simultaneously, which is not limited herein.
Based on this, in an optional implementation manner, in the case that the first interface slides to the first position along the first direction, the step 211 may specifically include:
and performing scrolling screen capture operation on the first interface along the first direction until the first interface slides to the first position, and ending the scrolling screen capture operation to obtain a first image.
Here, if the first interface is slid, the first interface may be subjected to scrolling and screen capturing to obtain the first image. The first image may be a long shot of a position displayed by the first interface from the position at which the electronic device receives the first input to the position displayed at the end of the sliding, that is, the first position. The scroll screen capture operation may be performed during the sliding of the first interface or after the sliding of the first interface to the first position.
For example, the web interface may be scrolled and captured while the user is sliding down the web interface via the stylus.
Therefore, through scrolling screen capture, the content targeted by the handwriting of the user can be not limited to the currently displayed content, and the user can also continue to write the new content displayed after the interface slides, so that the writing range is expanded.
Step 212, displaying a first image between the first interface and the first image layer.
Here, a first image obtained by performing a screen capture operation on the first interface may be displayed between the first interface and the first layer.
Based on this, after step 240, the writing control method may further include: step 241-242, wherein:
and 241, overlapping the first image layer and the second image layer to obtain a second image.
Here, the second layer may be a layer corresponding to the first image, the first layer may be a layer displaying the input track, and the second image may be an image displaying the first image content and the input track.
For example, when the user completes writing and exits from the current writing track display process, or jumps from the first interface to another interface through user manipulation in the manipulation mode, or when a third image layer is newly created for writing in the writing mode, the writing operation process may be ended, and at this time, the first image layer and the second image layer may be synthesized to obtain the second image.
Step 242, save the second image.
Here, in addition to saving the second image, images corresponding to the first layer and the second layer may be saved separately.
For example, after the multiple writing operation process is finished, if there are multiple second images, the multiple second images may be stored in one album, or the multiple second images may be combined into an image group and stored in a PDF document, which is not limited herein.
In a specific example, when the user uses the stylus pen with the pen point facing downward, the tablet may create a first layer and perform a screen capture operation on the currently displayed web interface to obtain a screen capture of the web interface, and then display the screen capture between the web interface and the first layer, and after the user uses the stylus pen to write on the screen, may display a corresponding writing track on the first layer. When the user quits the writing operation process, the second image layer corresponding to the screenshot of the webpage interface and the first image layer displaying the writing track can be stored after being overlapped, and the second image displaying the content of the webpage interface and the writing track corresponding to the content of the webpage interface is obtained so as to be convenient for sharing, recording and the like.
Therefore, a first image can be obtained by performing screen capture operation on the first interface, and the second image layer corresponding to the first image is superposed with the first image layer to obtain a second image containing the display content of the first interface and the corresponding writing track of the display content, so that a user can store the content to be marked conveniently.
In addition, the stylus in the writing control method may further include an erasing function, when a user uses one end of the stylus to write in the input frame supporting writing, the electronic device recognizes a writing track and converts the writing track into an input frame character, if the user needs to erase the input frame character, the stylus may be turned, the other end of the stylus is used to paint the input frame character needing to be erased, and the painted character is deleted. The touch control pen can erase the characters at any position in the plurality of characters, and after the characters are erased, the positions corresponding to the erased characters can be left vacant.
Illustratively, as shown in fig. 5, after the user writes with the stylus tip, the tablet computer recognizes the writing trace and converts it into an input box text 510: when the user wants to delete the input two characters, the user can turn the touch pen, the position 530 of the input two characters is smeared by using the circular end 520 of the touch pen, and the position 530 of the input two characters is displayed as blank after smearing.
Besides erasing characters, the erasing function can be used for drawing scenes, for example, when lines are erased by using a touch pen, erasing traces which are light in the middle and deep on two sides after a real eraser is erased can be simulated, so that the effect of drawing on the electronic equipment is closer to the effect of drawing on paper.
Based on the same inventive concept, the application also provides a writing control device. The writing control device provided in the embodiment of the present application will be described in detail below with reference to fig. 6.
Fig. 6 is a block diagram illustrating a structure of a writing control apparatus according to an exemplary embodiment.
As shown in fig. 6, the writing control apparatus 600 may include:
the first input receiving module 601 is configured to receive a first instruction, which is sent by a stylus and is for a first interface; the first interface is an interface which does not support writing;
the layer display module 602 is configured to display a first layer on a first interface in response to a first input;
a second input receiving module 603, configured to receive a second input;
the input response module 604 is configured to display an input track corresponding to the second input on the first layer or perform a corresponding operation on the first interface based on the input parameter of the second input.
The writing control device 600 is described in detail below, specifically as follows:
in one embodiment, the second input receiving module 603 may specifically include:
the first receiving submodule is used for receiving a second input of the first image layer; the first layer is set to respond to a touch event;
an input response module 604, comprising:
and the track display submodule is used for displaying an input track of the second input on the first layer according to the first touch event corresponding to the second input.
In one embodiment, the second input receiving module 603 may specifically include:
the second receiving submodule is used for receiving a second input to the first interface, and the second input comprises a first sub-input and a second sub-input; wherein, the first sub-input is the input of a touch control pen;
the input response module 604 may include:
the layer setting submodule is used for setting the first layer as a non-response touch event based on the input parameters input by the first sub-input module;
the event transmission submodule is used for transmitting a second touch event corresponding to the second sub-input to the first interface based on the second sub-input;
and the operation execution sub-module is used for executing the operation corresponding to the second sub-input on the first interface according to the second touch event.
In one embodiment, in a case where the first interface is an interface extending along a first direction and the second touch event is a sliding touch event along a second direction, the operating the execution submodule may include:
the sliding control unit is used for controlling the first interface to slide to a first position along a first direction according to the second touch event and controlling the first layer to slide to a second position along the first direction;
the first position of the first interface corresponds to the second position of the first layer, and the first direction is opposite to the second direction.
In one embodiment, the writing control apparatus 600 may further include:
the screen capture module is used for carrying out screen capture operation on the first interface after receiving first input of the first interface by the touch control pen to obtain a first image;
the first image display module is used for displaying a first image between the first interface and the first image layer;
the writing control apparatus 600 may further include:
the superposition module is used for superposing the first image layer and the second image layer to obtain a second image after an input track corresponding to the second input is displayed on the first image layer or corresponding operation is executed on the first interface based on the input parameter of the second input; the second image layer is a layer corresponding to the first image;
and the storage module is used for storing the second image.
In one embodiment, in the case that the first interface slides to the first position along the first direction, the screen capture module may include:
and the screen capture submodule is used for performing scrolling screen capture operation on the first interface along the first direction until the first interface slides to the first position, and finishing the scrolling screen capture operation to obtain a first image.
In one embodiment, the first input is an input of the stylus in a first posture, and the first sub-input is an instruction generated by the stylus in a second posture.
In one embodiment, the first input is an input of the stylus with the first end pressed, and the first sub-input is an input of the stylus with the second end pressed.
Therefore, the first input of the stylus to the first interface is received, the first layer is displayed on the first interface in response to the first input, the second input is received, the input track corresponding to the second input is displayed on the first layer or the corresponding operation is executed on the first interface based on the input parameter of the second input, the input track corresponding to the second input can be displayed on the first layer, in addition, the first interface can be controlled through the second input, and therefore the user can write and mark on the interface which does not support writing while the user can normally operate the interface.
The writing control device in the embodiment of the present application may be a device, or may be a component, an integrated circuit, or a chip in a terminal. The device can be mobile electronic equipment or non-mobile electronic equipment. By way of example, the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a palm top computer, a vehicle-mounted electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook or a Personal Digital Assistant (PDA), and the like, and the non-mobile electronic device may be a server, a Network Attached Storage (NAS), a Personal Computer (PC), a Television (TV), a teller machine or a self-service machine, and the like, and the embodiments of the present application are not particularly limited.
The writing control device in the embodiment of the present application may be a device having an operating system. The operating system may be an Android (Android) operating system, an ios operating system, or other possible operating systems, and embodiments of the present application are not limited specifically.
The writing control device provided in the embodiment of the present application can implement each process implemented by the method embodiments of fig. 1 to 5, and is not described here again to avoid repetition.
Optionally, as shown in fig. 7, an electronic device 700 is further provided in this embodiment of the present application, and includes a processor 701, a memory 702, and a program or an instruction stored in the memory 702 and executable on the processor 701, where the program or the instruction is executed by the processor 701 to implement each process of the writing control method embodiment, and can achieve the same technical effect, and no further description is provided here to avoid repetition.
It should be noted that the electronic device in the embodiment of the present application includes the mobile electronic device and the non-mobile electronic device described above.
Fig. 8 is a schematic diagram of a hardware structure of an electronic device implementing an embodiment of the present application.
The electronic device 800 includes, but is not limited to: a radio frequency unit 801, a network module 802, an audio output unit 803, an input unit 804, a sensor 805, a display unit 806, a user input unit 807, an interface unit 808, a memory 809, and a processor 810.
Those skilled in the art will appreciate that the electronic device 800 may further comprise a power source (e.g., a battery) for supplying power to the various components, and the power source may be logically connected to the processor 810 via a power management system, so as to manage charging, discharging, and power consumption management functions via the power management system. The electronic device structure shown in fig. 8 does not constitute a limitation of the electronic device, and the electronic device may include more or less components than those shown, or combine some components, or arrange different components, and thus, the description is omitted here.
The input unit 804 is configured to receive a first input of a first interface by a stylus, where the first interface is an interface that does not support writing;
a processor 810 for displaying a first layer on a first interface in response to a first input;
a user input unit 807 for receiving a second input;
the display unit 806 is configured to display an input track corresponding to the second input on the first layer or perform a corresponding operation on the first interface based on the input parameter of the second input.
Therefore, the first input of the stylus to the first interface is received, the first layer is displayed on the first interface in response to the first input, the second input is received, the input track corresponding to the second input is displayed on the first layer or the corresponding operation is executed on the first interface based on the input parameter of the second input, the input track corresponding to the second input can be displayed on the first layer, in addition, the first interface can be controlled through the second input, and therefore the user can write and mark on the interface which does not support writing while the user can normally operate the interface.
Optionally, the user input unit 807 is specifically configured to receive a second input to the first layer; the first layer is set to respond to a touch event;
the display unit 806 is specifically configured to display an input track of the second input on the first layer according to the first touch event corresponding to the second input;
optionally, the user input unit 807 is specifically configured to receive a second input to the first interface, where the second input includes a first sub-input and a second sub-input; wherein, the first sub-input is the input of a touch control pen;
the processor 810 is further configured to set the first layer to not respond to the touch event based on the input parameter of the first sub-input;
transmitting a second touch event corresponding to the second sub-input to the first interface based on the second sub-input;
and executing the operation corresponding to the second sub-input on the first interface according to the second touch event.
Optionally, the processor 810 is further configured to, when the first interface is an interface extending along a first direction and the second touch event is a sliding touch event along a second direction, control the first interface to slide to a first position along the first direction and control the first layer to slide to a second position along the first direction according to the second touch event;
the first position of the first interface corresponds to the second position of the first layer, and the first direction is opposite to the second direction.
Optionally, the processor 810 is further configured to perform a screen capture operation on the first interface after receiving a first input to the first interface by the stylus, so as to obtain a first image;
the display unit 806 is further configured to display a first image between the first interface and the first layer;
the processor 810 is further configured to, after an input track corresponding to a second input is displayed on the first layer or a corresponding operation is performed on the first interface based on the input parameter of the second input, superimpose the first layer and the second layer to obtain a second image; the second image layer is a layer corresponding to the first image;
the second image is saved.
Optionally, the processor 810 is specifically configured to, when the first interface slides to the first position along the first direction, perform a scrolling and screen-capturing operation on the first interface along the first direction until the first interface slides to the first position, and end the scrolling and screen-capturing operation to obtain the first image.
From this, through the both ends of transferring the touch-control pen, can fast switch write and control two kinds of functions, make the user can write and control at will through the touch-control pen at the interface that does not support writing, promote user experience.
It should be understood that in the embodiment of the present application, the input Unit 804 may include a Graphics Processing Unit (GPU) 8041 and a microphone 8042, and the Graphics Processing Unit 8041 processes image data of a still picture or a video obtained by an image capturing device (such as a camera) in a video capturing mode or an image capturing mode. The display unit 806 may include a display panel 8061, and the display panel 8061 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like. The user input unit 807 includes a touch panel 8071 and other input devices 8072. A touch panel 8071, also referred to as a touch screen. The touch panel 8071 may include two portions of a touch detection device and a touch controller. Other input devices 8072 may include, but are not limited to, a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, and a joystick, which are not described in detail herein. The memory 809 may be used to store software programs as well as various data including, but not limited to, application programs and operating systems. The processor 810 may integrate an application processor, which mainly handles operating systems, user interfaces, application programs, etc., and a modem processor, which mainly handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into processor 810.
The embodiment of the present application further provides a readable storage medium, where a program or an instruction is stored on the readable storage medium, and when the program or the instruction is executed by a processor, the program or the instruction implements each process of the writing control method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
The processor is the processor in the electronic device described in the above embodiment. The readable storage medium includes a computer readable storage medium, such as a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and so on.
The embodiment of the present application further provides a chip, where the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to execute a program or an instruction to implement each process of the above writing control method embodiment, and can achieve the same technical effect, and the description is omitted here to avoid repetition.
It should be understood that the chips mentioned in the embodiments of the present application may also be referred to as system-on-chip, system-on-chip or system-on-chip, etc.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element. Further, it should be noted that the scope of the methods and apparatus of the embodiments of the present application is not limited to performing the functions in the order illustrated or discussed, but may include performing the functions in a substantially simultaneous manner or in a reverse order based on the functions involved, e.g., the methods described may be performed in an order different than that described, and various steps may be added, omitted, or combined. In addition, features described with reference to certain examples may be combined in other examples.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present application may be embodied in the form of a computer software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal (such as a mobile phone, a computer, a server, or a network device) to execute the method according to the embodiments of the present application.
While the present embodiments have been described with reference to the accompanying drawings, it is to be understood that the invention is not limited to the precise embodiments described above, which are meant to be illustrative and not restrictive, and that various changes may be made therein by those skilled in the art without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (11)

1. A writing control method, comprising:
receiving a first input of a first interface by a stylus, wherein the first interface is an interface which does not support writing;
displaying a first layer on the first interface in response to the first input;
receiving a second input;
and displaying an input track corresponding to the second input on the first layer or executing corresponding operation on the first interface based on the input parameter of the second input.
2. The method of claim 1, wherein receiving a second input comprises:
receiving a second input to the first image layer; the first layer is set to respond to a touch event;
the displaying an input track corresponding to the second input on the first layer or executing an operation corresponding to the second input on the first interface based on the input parameter of the second input includes:
and displaying an input track of the second input on the first layer according to a first touch event corresponding to the second input.
3. The method of claim 1, wherein receiving a second input comprises:
receiving a second input to the first interface, the second input comprising a first sub-input and a second sub-input; wherein the first sub-input is an input of the stylus;
the displaying an input track corresponding to the second input on the first layer or executing an operation corresponding to the second input on the first interface based on the input parameter of the second input includes:
setting the first layer to not respond to a touch event based on the input parameter of the first sub-input;
based on the second sub-input, transmitting a second touch event corresponding to the second sub-input to the first interface;
and executing the operation corresponding to the second sub-input on the first interface according to the second touch event.
4. The method according to claim 3, wherein in a case where the first interface is an interface extending in a first direction and the second touch event is a sliding touch event in a second direction, the performing, according to the second touch event, an operation corresponding to the second sub-input on the first interface comprises:
according to the second touch event, controlling the first interface to slide to a first position along the first direction, and controlling the first layer to slide to a second position along the first direction;
the first position of the first interface corresponds to the second position of the first layer, and the first direction is opposite to the second direction.
5. The method of any of claims 1-4, wherein after receiving the first input to the first interface by the stylus, the method further comprises:
performing screen capture operation on the first interface to obtain a first image;
displaying the first image between the first interface and the first image layer;
after the input parameter based on the second input displays an input track corresponding to the second input on the first layer or performs a corresponding operation on the first interface, the method further includes:
superposing the first image layer and the second image layer to obtain a second image; the second image layer is a layer corresponding to the first image;
and saving the second image.
6. The method according to claim 5, wherein the performing a screen capture operation on the first interface to obtain a first image when the first interface slides to a first position along a first direction comprises:
and performing scrolling screen capture operation on the first interface along the first direction until the first interface slides to the first position, and ending the scrolling screen capture operation to obtain the first image.
7. The method of claim 3, wherein the first input is an input of the stylus in a first gesture and the first sub-input is an input of the stylus in a second gesture.
8. The method of claim 3, wherein the first input is an input of the stylus with a first end pressed, and wherein the first sub-input is an input of the stylus with a second end pressed.
9. A writing control apparatus, comprising:
the first input receiving module is used for receiving first input of a first interface by a touch control pen; wherein the first interface is an interface which does not support writing;
the layer display module is used for responding to the first input and displaying a first layer on the first interface;
the second input receiving module is used for receiving a second input;
and the input response module is used for displaying an input track corresponding to the second input on the first layer or executing corresponding operation on the first interface based on the input parameter of the second input.
10. An electronic device comprising a processor, a memory, and a program or instructions stored on the memory and executable on the processor, the program or instructions when executed by the processor implementing the steps of the writing control method according to any one of claims 1-8.
11. A readable storage medium, on which a program or instructions are stored, which when executed by a processor, implement the steps of the writing control method according to any one of claims 1 to 8.
CN202111012762.XA 2021-08-31 2021-08-31 Writing control method and device, electronic equipment and storage medium Pending CN113703631A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202111012762.XA CN113703631A (en) 2021-08-31 2021-08-31 Writing control method and device, electronic equipment and storage medium
PCT/CN2022/114410 WO2023030117A1 (en) 2021-08-31 2022-08-24 Writing control method and apparatus, and electronic device and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111012762.XA CN113703631A (en) 2021-08-31 2021-08-31 Writing control method and device, electronic equipment and storage medium

Publications (1)

Publication Number Publication Date
CN113703631A true CN113703631A (en) 2021-11-26

Family

ID=78658021

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111012762.XA Pending CN113703631A (en) 2021-08-31 2021-08-31 Writing control method and device, electronic equipment and storage medium

Country Status (2)

Country Link
CN (1) CN113703631A (en)
WO (1) WO2023030117A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114610188A (en) * 2022-02-09 2022-06-10 广东能龙教育股份有限公司 Interface handwriting writing method and system, control device and storage medium
WO2023030117A1 (en) * 2021-08-31 2023-03-09 维沃移动通信有限公司 Writing control method and apparatus, and electronic device and storage medium

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115185444A (en) * 2019-04-17 2022-10-14 华为技术有限公司 Method for adding annotations and electronic equipment
CN112783408A (en) * 2021-01-28 2021-05-11 维沃移动通信有限公司 Gesture navigation method and device of electronic equipment, equipment and readable storage medium
CN113703631A (en) * 2021-08-31 2021-11-26 维沃移动通信有限公司 Writing control method and device, electronic equipment and storage medium

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023030117A1 (en) * 2021-08-31 2023-03-09 维沃移动通信有限公司 Writing control method and apparatus, and electronic device and storage medium
CN114610188A (en) * 2022-02-09 2022-06-10 广东能龙教育股份有限公司 Interface handwriting writing method and system, control device and storage medium
CN114610188B (en) * 2022-02-09 2024-02-13 广东能龙教育股份有限公司 Interface handwriting writing method and system, control device and storage medium

Also Published As

Publication number Publication date
WO2023030117A1 (en) 2023-03-09

Similar Documents

Publication Publication Date Title
US8244233B2 (en) Systems and methods for operating a virtual whiteboard using a mobile phone device
JP6233314B2 (en) Information processing apparatus, information processing method, and computer-readable recording medium
CN110297550B (en) Label display method and device, screen throwing equipment, terminal and storage medium
WO2023030117A1 (en) Writing control method and apparatus, and electronic device and storage medium
CN111475097B (en) Handwriting selection method and device, computer equipment and storage medium
CN104239266B (en) Graphics device and graphic display method
CN104199552A (en) Multi-screen display method, device and system
JP5306528B1 (en) Electronic device and handwritten document processing method
US20150347000A1 (en) Electronic device and handwriting-data processing method
JP2015158900A (en) Information processing device, information processing method and information processing program
CN102830812A (en) Pen-based terminal, controlled device and man-machine interactive system
CN104298437A (en) Text selection method, device and terminal
CN114518820A (en) Icon sorting method and device and electronic equipment
CN111665984A (en) Information processing method and device, electronic equipment and readable storage medium
JP5994224B2 (en) Terminal device, electronic pen system, and program
CN109656435B (en) Display control device and recording medium
CN112269523B (en) Object editing processing method and device and electronic equipment
EP2712433B1 (en) User interface for drawing with electronic devices
CN113360035A (en) Icon sorting method and device and electronic equipment
CN114518824A (en) Note recording method and device and electronic equipment
CN113986080A (en) Multimedia file editing method and device and electronic equipment
CN113485590A (en) Touch operation method and device
CN110032317B (en) Wireless demonstration equipment and marking control method and device thereof
CN112269481A (en) Method and device for controlling friction force adjustment and electronic equipment
CN113778277A (en) Application icon management method and device and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination