US20230315282A1 - Writing interaction method, smart interactive display device and writing interaction system - Google Patents

Writing interaction method, smart interactive display device and writing interaction system Download PDF

Info

Publication number
US20230315282A1
US20230315282A1 US17/913,853 US202017913853A US2023315282A1 US 20230315282 A1 US20230315282 A1 US 20230315282A1 US 202017913853 A US202017913853 A US 202017913853A US 2023315282 A1 US2023315282 A1 US 2023315282A1
Authority
US
United States
Prior art keywords
writing
smart
touch event
handwriting
generated
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US17/913,853
Other versions
US11861160B2 (en
Inventor
Zipeng YU
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Honghe Innovation Information Technology Co Ltd
Original Assignee
Shenzhen Honghe Innovation Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Honghe Innovation Information Technology Co Ltd filed Critical Shenzhen Honghe Innovation Information Technology Co Ltd
Publication of US20230315282A1 publication Critical patent/US20230315282A1/en
Application granted granted Critical
Publication of US11861160B2 publication Critical patent/US11861160B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0354Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
    • G06F3/03545Pens or stylus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/038Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04886Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/038Indexing scheme relating to G06F3/038
    • G06F2203/0382Plural input, i.e. interface arrangements in which a plurality of input device of the same type are in communication with a PC
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04803Split screen, i.e. subdividing the display area or the window area into separate subareas
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04808Several contacts: gestures triggering a specific function, e.g. scrolling, zooming, right-click, when the user establishes several contacts with the surface simultaneously; e.g. using several fingers or a combination of fingers and pen
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/14Image acquisition
    • G06V30/142Image acquisition using hand-held instruments; Constructional details of the instruments
    • G06V30/1423Image acquisition using hand-held instruments; Constructional details of the instruments the instrument generating sequences of position coordinates corresponding to handwriting

Definitions

  • the present application relates to a technical field of interaction between users and computers, and in particular to a writing interaction method, a smart interactive display device and a writing interaction system.
  • Smart interactive display devices such as smart interactive tablets
  • the existing smart interactive display device is also used together with a smart pen, and the writing control can be realized through the communication connection between the smart pen and the smart 25 interactive display device.
  • multiple smart pens actually share the same writing region, so there are some problems as follows. For example, when several people are required to hold smart pens and write on the blackboard at the same time, since everyone's handwriting, writing habits and writing layout are different, some people are more casual about the writing layout, thereby generating a chaotic writing layout, and occupying other people's writing space.
  • the embodiments of the present application provide a writing interaction method applied to the interaction between a smart interactive display device and a smart pen, wherein the writing interaction method includes:
  • the embodiments of the present application further provide a smart interactive display device for interacting with a smart pen, the smart interactive display device including a touch event matching module, a touch event response module, and a writing region generation module, wherein:
  • the embodiments of the present application further provide a smart interactive display device, including a capacitive touch screen, a processor and a computer-readable storage medium, wherein the computer-readable storage medium stores a writing interaction program, and the writing interaction program, when being executed, implements the writing interaction method as described above.
  • the embodiments of the present application further provide a writing interaction system, including a smart pen and the smart interactive display device as described above, wherein the smart interactive display device includes a touch event matching module, a touch event response module, and a writing region generation module, wherein:
  • the embodiments of the present application further provide a writing interaction system, including a smart pen and the smart interactive display device as described above, wherein the smart interactive display device includes a capacitive touch screen, a processor and a computer-readable storage medium, wherein the computer-readable storage medium stores a writing interaction program, and the writing interaction program, when being executed, implements the writing interaction method as described above.
  • FIG. 1 is a flowchart of a first embodiment of a writing interaction method according to the present application
  • FIG. 2 is a flowchart of a second embodiment of a writing interaction method according to the present application.
  • FIG. 3 is a flowchart of a third embodiment of a writing interaction method according to the present application.
  • FIG. 4 is a flowchart of a fourth embodiment of a writing interaction method according to the present application.
  • FIG. 5 is a flowchart of a fifth embodiment of a writing interaction method according to the present application.
  • FIG. 6 is a flowchart of a sixth embodiment of a writing interaction method according to the present application.
  • FIG. 7 is a flowchart of a seventh embodiment of a writing interaction method according to the present application.
  • FIG. 8 is a flowchart of an eighth embodiment of a writing interaction method according to the present application.
  • FIG. 9 is a first state schematic diagram of an implementation process of a writing interaction method according to the present application.
  • FIG. 10 is a second state schematic diagram of an implementation process of a writing interaction method according to the present application.
  • FIG. 11 is a module structure of an embodiment of a writing interaction system according to the present application.
  • FIG. 1 is a flowchart of a first embodiment of a writing interaction method according to the present application.
  • the writing interaction method includes step 100 .
  • Step 100 detecting a first touch event of one or more smart pens and acquiring one or more corresponding identifiers of the one or more smart pens.
  • the smart pens may generate a touch event on a screen of the smart interactive display device, the touch event includes the above first touch event, and the type of the touch event can be capacitive touch, infrared touch and the like.
  • a communication connection is established between the smart pens and the smart interactive device, and the connection between the smart pens and the smart interactive device may transmit the identifiers of the smart pens.
  • an active capacitive pen may establish a communication connection with a touch panel card of a capacitive touch screen through Microsoft Pen Protocol (MPP) of Microsoft or proprietary Active Electrostatic Solution (AES) of Wacom, and transmit its own identifier.
  • MPP Microsoft Pen Protocol
  • AES Active Electrostatic Solution
  • the smart pen may also transmit its own identifier through the electromagnetic induction input technology or the infrared communication technology.
  • the touch events of different smart pens may occur simultaneously or sequentially. From the perspective of application, although there are multiple people operating the smart pens at the same time, the start time of the writing of each person is generally not exactly the same. Technically speaking, it is easier to match touch events that occur one after the other with the corresponding identifiers of the smart pens.
  • the initial stroke of the character “ ” is a horizontal line
  • the initial stroke of the character “ ” is a dot.
  • the time when the initial positions of the two strokes are generated can be simultaneous or sequential.
  • the writing interaction method further includes step 102 : generating handwriting according to the first touch event, and determining whether the generated handwriting corresponds to multiple smart pens of the smart pens according to the identifiers of the smart pens.
  • step 102 A of generating handwriting is arranged before step 102 B of determining the number of the smart pens, determining the number of the smart pens is actually determining the occurred first touch event belongs to which smart pen(s) and then determining the number of the smart pens according to the identifiers of the smart pens. Therefore, apparently, the handwriting is associated with the identifiers of the smart pens, and substantially, the first touch event is associated with the identifiers of the smart pens. Therefore, the sequence between step 102 A and step 102 B may not be limited.
  • step 102 B can also be performed after the first touch event occurs. For example, in FIG. 9 , the determination of step 102 B is completed after the writing of the words “ ” and “ ” is completed.
  • the writing interaction method further includes step 104 : under a condition that the handwriting corresponds to the multiple smart pens, generating multiple non-overlapping writing regions, wherein the multiple writing regions are in one-to-one correspondence with the multiple smart pens, each writing region covers handwriting of a corresponding smart pen, and each writing region only responds to a touch event of the corresponding smart pen for generating handwriting.
  • the size and position of the writing region may be determined according to the customary size and the position where the initial first touch event occurs.
  • the two writing regions generated in FIG. 10 are set adjacent to the upper side of the display region, and according to the default left-aligned typesetting mode, the left border of the writing region is placed close to the position where the initial first touch event occurs, that is, the position where the initial handwriting is generated.
  • the left and right writing regions in FIG. 10 do not overlap with each other, but in order to make full use of the display region, the two writing regions can share a common border.
  • the writing region can be determined in the form of a graphic window, or can be defined by only colored lines. After multiple writing regions are generated, since all writing regions are independent of each other, suppose that the left writing region in FIG. 10 belongs to the smart pen marked 010, and the right writing region belongs to the smart pen marked 011. Then, under a condition that the smart pen marked 011 is writing in the left writing region, it will not be responded as displaying handwriting. It is understandable that if the handwriting corresponds to only one smart pen, the default writing region, such as the display region of the entire display screen, is used; or only a suitable size of writing region is generated for the smart pen.
  • the generation source of the first touch event (that is, the smart pens) may be identified. Under a condition that it is detected that the written handwriting comes from multiple smart pens, a corresponding number of writing regions may be automatically generated. Each writing region covers handwriting of a corresponding smart pen, and each writing region only responds to a touch event of the corresponding smart pen for generating handwriting. Therefore, it is ensured that when multiple people use multiple pens to write, the generated multiple writing regions are independent of each other, the content generated by different smart pens will not be cluttered, and the writing content will be clear, thereby improving the user experience of multiple people operating smart pens for writing.
  • the writing region is generated by matching the initial handwriting of each smart pen, in other words, the writing region is generated with the initial handwriting of each smart pen, and is positioned according to the initial handwriting. Therefore, a more flexible region layout can be realized, thereby adapting to different application scenarios.
  • FIG. 2 is a flowchart of a second embodiment of a writing interaction method according to the present application. After the step 104 , the writing interaction method further includes step 206 .
  • Step 206 detecting a second touch event of a smart pen and acquiring a corresponding identifier of the smart pen.
  • This step is used to detect the touch event of the smart pen again after the latest generation of the writing regions, that is, to detect the second touch event of the smart pen and match the identifier of the smart pen corresponding to the second touch event.
  • a data basis can be provided for subsequent smart control.
  • the writing interaction method further includes step 208 : determining whether the second touch event is generated by a newly added smart pen according to the identifier of the smart pen.
  • each touch event is associated with the identifier of the smart pen, by querying the data of the identifiers of the smart pens, whether an identifier of a new smart pen is added can be determined, and thus whether the new smart pen or an operator of the new smart pen is added can be indirectly determined.
  • the writing interaction method further includes step 210 : under a condition that the second touch event is generated by the newly added smart pen, determining whether the second touch event is generated in the generated writing regions according to a location where the second touch event is generated.
  • the writing interaction method further includes step 212 : under a condition that the second touch event is generated in the generated writing regions, generating no handwriting according to the second touch event.
  • the principle is that the existing writing regions are prioritized and their sizes are not adjusted.
  • the second touch event occurs in the existing writing regions, in order to ensure the independence of the existing writing regions relative to the newly added smart pen, the second touch event of the newly added smart pen is not responded and no handwriting is generated.
  • no handwriting is generated if the smart pen with the identification number of 001 writes in any of the left and right writing regions in FIG. 10 .
  • the writing interaction method further includes step 214 : under a condition that the second touch event is not generated in the generated writing regions, generating handwriting according to the second touch event, and generating a new writing region corresponding to the newly added smart pen, wherein the new writing region is not overlapped with the existing multiple writing regions, the new writing region covers the handwriting corresponding to the newly added smart pen, and the new writing region only responds to a touch event of the newly added smart pen for generating handwriting.
  • the principle is that the existing writing regions are prioritized and their sizes are not adjusted.
  • a new writing region is generated for the newly added smart pen. For example, referring to FIG. 10 , under a condition that a third smart pen with an identification number of 001 is newly added, a new writing region is generated if the second touch event of the smart pen is generated near the lower edge of the display region and outside the existing left and right writing regions.
  • FIG. 3 is a flowchart of a third embodiment of a writing interaction method according to the present application. After the step 104 , the writing interaction method further includes steps 306 , 308 and 310 .
  • Step 306 detecting a third touch event of a smart pen and acquiring a corresponding identifier of the smart pen.
  • Step 308 determining whether the third touch event is generated by a newly added smart pen according to the identifier of the smart pen.
  • Step 310 under a condition that the third touch event is generated by the newly added smart pen, determining whether the third touch event is generated in the generated writing regions according to a location where the third touch event is generated.
  • the writing interaction method further includes step 312 : under a condition that the third touch event is generated in the generated writing regions, regenerating multiple non-overlapping writing regions, wherein the regenerated multiple writing regions are in one-to-one correspondence with the multiple smart pens and the newly added smart pen, each writing region covers a touch track of a corresponding smart pen, and each writing region only responds to a touch event of the corresponding smart pen for generating handwriting.
  • the multiple writing regions need to be regenerated. For example, under a condition that the third touch event of the newly added smart pen falls in the left or right writing region, the multiple writing regions need to be regenerated. For example, under a condition that the newly added smart pen writes “ ” between the words “ ” and “ ”, the regenerated writing regions can include three writing regions (left, middle and right), with the word “ ” in the middle writing region.
  • the writing interaction method further includes step 314 : under a condition that third touch event is not generated in the generated writing regions, generating a new writing region corresponding to the newly added smart pen, wherein the existing multiple writing areas remain unchanged, wherein the new writing region is not overlapped with the existing multiple writing regions, the new writing region covers handwriting corresponding to the newly added smart pen, and the new writing region only responds to a touch event of the newly added smart pen for generating handwriting.
  • a new writing region is generated for the newly added smart pen, and the existing multiple writing regions remain unchanged.
  • the third touch event of the newly added smart pen can be flexibly responded, and the independence of the existing writing regions can be maintained. For example, under a condition that the third touch event of the newly added smart pen occurs at the lower right corner of the display region, then a new writing region is generated for the newly added smart pen at the lower right corner, while keeping the original left and right writing regions unchanged.
  • FIG. 4 is a flowchart of a fourth embodiment of a writing interaction method according to the present application.
  • the writing interaction method further includes steps 406 , 408 and 410 .
  • Step 406 detecting a fourth touch event of a smart pen and acquiring a corresponding identifier of the smart pen.
  • Step 408 determining whether the fourth touch event is generated by a newly added smart pen according to the identifier of the smart pen.
  • Step 410 under a condition that the fourth touch event is not generated by the newly added smart pen, determining whether the fourth touch event is generated in the generated writing regions according to a location where the fourth touch event is generated.
  • the fourth touch event generated by it may fall within the generated writing regions or may fall outside the generated writing regions.
  • the writing interaction method further includes step 412 : under a condition that the fourth touch event is not generated in the generated writing regions, tentatively expanding the writing region corresponding to the smart pen according to a preset rule, so as to cover a track corresponding to the fourth touch event.
  • the writing region is usually a rectangle, but in some embodiments, the writing region can also be other shapes, such as an ellipse.
  • the preset rule is to expand the corresponding writing region according to the preset rule to cover the track of the fourth touch event.
  • the right writing region belongs to the smart pen marked as 011. Under a condition that the smart pen marked as 011 generates a fourth touch event near the lower edge of the writing region, the right writing region can be tentatively expanded downward. Under a condition that the smart pen marked as 011 generates a fourth touch event near the left edge of the writing region, the right writing region can be tentatively expanded to the left.
  • the writing interaction method further includes step 414 : determining whether the tentatively expanded writing region overlaps with other writing regions.
  • the writing interaction method further includes step 416 : under a condition that the tentatively expanded writing region overlaps with the other writing regions, canceling expansion of the writing region corresponding to the smart pen and generating no handwriting according to the fourth touch event.
  • step 412 for example, following the example in step 412 , under a condition that the right writing region is extended to the left, the right writing region will overlap the left writing region. Therefore, the expansion of the right writing region should be canceled, and no handwriting will be generated according to the fourth touch event.
  • the writing interaction method further includes step 418 : under a condition that the tentatively expanded writing region is not overlapped with the other writing regions, expanding the writing region corresponding to the smart pen and generating handwriting according to the fourth touch event.
  • the right writing region is extended downward, for example, the right writing region expands downward until the lower edge is flush with the lower edge of the left writing region, thereby actually extending the writing region corresponding to the smart pen, and generating handwriting in respond to the fourth touch event.
  • the initially generated writing region may not be enough for writing. Therefore, under a condition that the touch event of the corresponding smart pen falls outside the original writing region, the original writing region may be tentatively expanded. Under the premise of not overlapping with other writing regions, the original writing region may be expanded, so as to more flexibly adapt to the needs of the users.
  • the generating multiple non-overlapping writing regions specifically includes:
  • generating the multiple non-overlapping writing regions on a screen display region includes:
  • FIG. 5 is a flowchart of a fifth embodiment of a writing interaction method according to the present application.
  • the generating multiple non-overlapping writing regions specifically includes step 541 .
  • Step 541 acquiring a location of a leftmost coordinate point of the handwriting corresponding to each smart pen.
  • the leftmost coordinate point of the handwriting of each smart pen is obtained as a reference point for generating the writing region.
  • the leftmost point of the word “ ” is P 2 , and its coordinates are (c, d), and the leftmost point of the word “ ” is P 1 , and its coordinates are (a, b).
  • the generating multiple non-overlapping writing regions further includes step 543 : determining whether a distance between two adjacent leftmost coordinate points is greater than or equal to a product of ⁇ and L, wherein L is a horizontal length of the screen display region, and ⁇ is less than 0.5.
  • the distance can cover the possible horizontal length of the left handwriting among the two adjacent handwritings in the horizontal direction, thereby ensuring that a vertical dividing line can be set between the two handwritings.
  • the horizontal lengths of the two handwritings are generally determined by the time required for determining the number of the smart pens, and the horizontal length is usually less than L/3. Of course, the time required for the determination can also be extended a little longer, so as to extend the length occupied by the initial handwriting of each smart pen.
  • the generating multiple non-overlapping writing regions further includes step 545 : under a condition that the distance between the two adjacent leftmost coordinate points is greater than or equal to the product of ⁇ and L, except for a leftmost coordinate point close to a left side of the screen display region, forming the writing regions by setting vertical dividing lines referring to the other leftmost coordinate points.
  • the writing region corresponding to the handwriting can be bounded by the left side of the display region, like the left writing region in FIG. 10 .
  • the touch screen of the smart interactive display device is usually a wide screen, such as a 16:9 wide screen, and multiple people are standing along the horizontal direction in sequence when writing. Therefore, under a condition that multiple people are writing at the same time, it is also appropriate to only consider dividing the display region horizontally to generate the writing regions, and the algorithm required for this setup is relatively simpler.
  • FIG. 6 is a flowchart of a sixth embodiment of a writing interaction method according to the present application.
  • the step 104 specifically includes step 642 .
  • Step 642 under a condition that the handwriting corresponds to the multiple smart pens, and under a condition that the generated handwriting corresponds to two smart pens, acquiring a location of a leftmost coordinate point of the handwriting corresponding to each smart pen.
  • step 541 since left alignment is the most commonly used typesetting method, in order to adapt to this typesetting method, the leftmost coordinate point of the handwriting of each smart pen is obtained as a reference point for generating the writing region.
  • the step 104 further includes step 644 : determining whether a distance between two adjacent leftmost coordinate points is greater than or equal to L/2, wherein L is a horizontal length of the screen display region.
  • the distance between P 1 and P 2 is greater than or equal to L/2, that is, the value of a-c is greater than or equal to L/2. It is indicated that the distance between P 1 and P 2 is large enough to completely cover the possible horizontal length of the left handwriting among the two adjacent handwritings in the horizontal direction, thereby ensuring greater flexibility in setting a vertical dividing line between the two handwritings.
  • the step 104 further includes step 646 : under a condition that the distance between the two adjacent leftmost coordinate points is greater than or equal to L/2, forming the writing regions by using a vertical centerline of the screen display region as a dividing line.
  • the vertical centerline of the display region is used as the boundary to form two writing regions of the same size.
  • the writing region corresponding to different handwriting may not be formed by setting the vertical dividing line referring to the leftmost coordinate point of the handwriting, but the vertical centerline of the display region may be used as the boundary to form two writing regions of the same size.
  • FIG. 7 is a flowchart of a seventh embodiment of a writing interaction method according to the present application.
  • the step 104 includes step 703 A.
  • Step 703 A under a condition that the handwriting corresponds to the multiple smart pens, popping up an instruction window requesting a user to confirm whether to generate the multiple writing regions.
  • a pop-up window is set to request the user to confirm.
  • the step 104 further includes step 703 B: monitoring an instruction command input by the user in the instruction window.
  • the input instruction command may be a touch event, a voice command, or a character command input from a keyboard.
  • the step 104 further includes step 703 C: under a condition that a command of determining to generate the multiple writing regions is received from the user, generating the multiple non-overlapping writing regions, wherein the multiple writing regions are in one-to-one correspondence with the multiple smart pens, each writing region covers handwriting of a corresponding smart pen, and each writing region only responds to a touch event of the corresponding smart pen for generating handwriting.
  • the writing interaction method After monitoring the instruction command input by the user in the instruction window, the writing interaction method further includes step 705 .
  • Step 705 under a condition that a command of cancelling generation of the multiple writing regions is received from the user, maintaining original writing regions.
  • the original writing regions are maintained, for example, multiple smart pens write in a same writing region.
  • the identifier of the newly added smart pen may be included in that writing region, that is, the writing region can accept the writing input of two smart pens.
  • a user confirmation process is set before determining whether to generate multiple writing regions, so that the smart interactive display device can more accurately meet the needs of the user, thereby achieving a better user experience.
  • each writing region includes an editing region and a menu bar region, and the editing region of each writing region covers the handwriting corresponding to the smart pen and corresponds to the touch event of the smart pen;
  • the menu bar region includes a writing main menu, and submenus of the writing main menu include color, eraser, and stroke thickness submenus.
  • the upper parts of the left and right writing regions are provided with a menu bar region
  • the square patterns in the menu bar in the figure represents graphic function buttons or icons
  • the outside of the menu bar is the editing region.
  • the menu bar region can also be set to be hidden and can be called up when needed.
  • the user can set the color of the handwriting by invoking the color command, or erase the generated handwriting by invoking the eraser, and can also set the thickness of the handwriting by invoking the stroke thickness.
  • the user can independently edit the handwriting in each writing region.
  • FIG. 8 is a flowchart of an eighth embodiment of a writing interaction method according to the present application.
  • the menu bar region includes a partition canceling main menu; the writing interaction method further includes step 806 .
  • Step 806 monitoring the input command of the user in the menu bar.
  • the input instruction command may be a touch event, a voice command, or a character command input from a keyboard.
  • the writing interaction method further includes step 810 : under a condition that the command of canceling the partitions is detected, restoring the writing regions to the original single writing region, and deleting the content in each of the writing regions.
  • a corresponding trigger command can be conveniently input in the menu bar, which restores the writing regions to the original single region, and deletes the content to quickly clear the screen.
  • the math class is coming to an end and the next class is English class.
  • the display region of the smart interactive display device has multiple writing regions, all operators have finished writing, and the teacher or any operator can trigger the command of canceling the partitions.
  • the screen can be quickly cleared, and it is convenient to use the writing interaction method of the smart interactive display device of the present application in English class.
  • the user when the user wants to cancel multiple writing regions that have been generated and clear the screen, the user can use the partition canceling main menu to quickly achieve the goal.
  • the present application also provides a smart interactive display device for interacting with a smart pen.
  • the smart interactive display device includes a touch event matching module, a touch event response module, and a writing area generation module, wherein,
  • the matching generation unit is configured to generate multiple non-overlapping writing regions
  • the matching generation unit is configured to generate multiple non-overlapping writing regions
  • the matching generation unit is configured to generate multiple non-overlapping writing regions
  • the matching generation unit includes a reference point acquisition unit, a distance determination unit and a region generation unit, wherein,
  • region generation unit is specifically configured to:
  • the writing region generation module also includes a partition pop-up window unit and a command monitoring unit, wherein,
  • each writing region includes an editing region and a menu bar region, and the editing region of each writing region covers the handwriting corresponding to the smart pen and corresponds to the touch event of the smart pen;
  • the menu bar region includes a writing main menu, and submenus of the writing main menu include color, eraser, and stroke thickness submenus.
  • the writing region generation module further includes a command monitoring unit, the command monitoring unit monitors the input command of the user in the menu bar; under a condition that the command monitoring unit detects the command of canceling the partitions, the matching generation module is further configured to restore the writing regions to the original single writing region, and delete the content in each of the writing regions.
  • the present application also provides a smart interactive display device, including a capacitive touch screen, a processor, and a computer-readable storage medium, wherein the computer-readable storage medium stores a writing interaction program, and the writing interaction program, when being executed, implements the writing interaction method as described above.
  • the specific steps of the writing interaction method may refer to the above embodiments. Since the smart interactive display device adopts all technical solutions of the above embodiments, it has at least all beneficial effects brought by the technical solutions of the above embodiments, will not be repeated here.
  • the present application also provides a writing interactive system, including a smart pen and the above smart interactive display device including various modules.
  • the specific structure of the smart interactive display device may refer to the above embodiments. Since the smart interactive display device adopts all technical solutions of the above embodiments, it has at least all beneficial effects brought by the technical solutions of the above embodiments, will not be repeated here.
  • the present application also provides a writing interactive system, including an active capacitive smart pen and the above smart interactive display device including the computer storage medium.
  • a writing interactive system including an active capacitive smart pen and the above smart interactive display device including the computer storage medium.
  • the specific structure of the smart interactive display device and the steps implemented by executing the writing interactive program may refer to the above embodiments. Since the writing interactive system adopts all technical solutions of the above embodiments, it has at least all beneficial effects brought by the technical solutions of the above embodiments, will not be repeated here.
  • the terms “comprise”, “include”, or any other variants thereof are intended to represent a non-exclusive inclusion, such that a process, method, article or device including a series of elements includes not only those elements, but also other elements that are not explicitly listed or elements inherent to such a process, method, article or device. Without more constraints, the elements following an expression “comprise/include . . . ” do not exclude the existence of additional identical elements in the process, method, article or device that includes the elements.
  • relational terms such as first and second, are used merely to distinguish one entity or operation from another entity or operation, without necessarily requiring or implying any actual such relationships or orders of these entities or operations.

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

A writing interaction method, a smart interactive display device and a writing interaction system are provided. The writing interaction method includes: detecting a first touch event of one or more smart pens and acquiring one or more corresponding identifiers of the one or more smart pens (S100); generating handwriting according to the first touch event (S102A), and determining whether the generated handwriting corresponds to multiple smart pens of the smart pens according to the identifiers of the smart pens (S102B); under a condition that the handwriting corresponds to the multiple smart pens, generating multiple non-overlapping writing regions, wherein the multiple writing regions are in one-to-one correspondence with the multiple smart pens, each writing region covers handwriting of a corresponding smart pen, and each writing region only responds to a touch event of the corresponding smart pen for generating handwriting (S104).

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • The present application is a National Stage of International Application No. PCT/CN2020/100379 filed on Jul. 6, 2020, which claims priority to Chinese patent application No. 202010213966.9, filed on Mar. 24, 2020 and entitled “WRITING INTERACTION METHOD, SMART INTERACTIVE DISPLAY DEVICE AND WRITING INTERACTION SYSTEM”, both of which are hereby incorporated by reference in their entireties.
  • TECHNICAL FIELD
  • The present application relates to a technical field of interaction between users and computers, and in particular to a writing interaction method, a smart interactive display device and a writing interaction system.
  • BACKGROUND
  • Smart interactive display devices, such as smart interactive tablets, are increasingly being used in various fields, such as education and teaching, corporate meetings, business presentations and other fields. In order to better realize the interactive function, the existing smart interactive display device is also used together with a smart pen, and the writing control can be realized through the communication connection between the smart pen and the smart 25 interactive display device. However, when the application of writing on the blackboard is provided for multiple people at the same time, multiple smart pens actually share the same writing region, so there are some problems as follows. For example, when several people are required to hold smart pens and write on the blackboard at the same time, since everyone's handwriting, writing habits and writing layout are different, some people are more casual about the writing layout, thereby generating a chaotic writing layout, and occupying other people's writing space. Thus, due to the problem of chaotic writing layout, their own writing space is wasted, and their writing content may occupy other people's writing space, resulting in a chaotic writing layout of the entire page, and the viewers can't even distinguish the writing content written by different people. It can be seen that the current smart interactive display device can provide poor user experience in dealing with multiple people operating smart pens for writing.
  • SUMMARY
  • The embodiments of the present application provide a writing interaction method applied to the interaction between a smart interactive display device and a smart pen, wherein the writing interaction method includes:
      • detecting a first touch event of one or more smart pens and acquiring one or more corresponding identifiers of the one or more smart pens;
      • generating handwriting according to the first touch event, and determining whether the generated handwriting corresponds to multiple smart pens of the smart pens according to the identifiers of the smart pens;
      • under a condition that the handwriting corresponds to the multiple smart pens, generating multiple non-overlapping writing regions, wherein the multiple writing regions are in one-to-one correspondence with the multiple smart pens, each writing region covers handwriting of a corresponding smart pen, and each writing region only responds to a touch event of the corresponding smart pen for generating handwriting.
  • The embodiments of the present application further provide a smart interactive display device for interacting with a smart pen, the smart interactive display device including a touch event matching module, a touch event response module, and a writing region generation module, wherein:
      • the touch event matching module includes a touch detection unit and a touch matching unit, the touch detection unit is configured to detect a first touch event of one or more smart pens, and the touch matching unit is configured to acquire one or more identifiers of the one or more smart pens and match the corresponding first touch event;
      • the touch event response module includes a handwriting generation unit and a quantity monitoring unit, the handwriting generation unit is configured to generate handwriting corresponding to the first touch event; the quantity monitoring unit is configured to determine whether the generated handwriting corresponds to multiple smart pens of the smart pens according to the identifiers of the smart pens;
      • the writing region generation module includes a matching generation unit and a response control unit, the matching generation unit is configured to generate multiple non-overlapping writing regions under a condition that the generated handwriting corresponds to the multiple smart pens, wherein the multiple writing regions are in one-to-one correspondence with the multiple smart pens, and each writing region covers handwriting of a corresponding smart pen; and the response control unit is configured to control each writing region to only respond to a touch event of the corresponding smart pen for generating handwriting.
  • The embodiments of the present application further provide a smart interactive display device, including a capacitive touch screen, a processor and a computer-readable storage medium, wherein the computer-readable storage medium stores a writing interaction program, and the writing interaction program, when being executed, implements the writing interaction method as described above.
  • The embodiments of the present application further provide a writing interaction system, including a smart pen and the smart interactive display device as described above, wherein the smart interactive display device includes a touch event matching module, a touch event response module, and a writing region generation module, wherein:
      • the touch event matching module includes a touch detection unit and a touch matching unit, the touch detection unit is configured to detect a first touch event of one or more smart pens, and the touch matching unit is configured to acquire one or more identifiers of the one or more smart pens and match the corresponding first touch event;
      • the touch event response module includes a handwriting generation unit and a quantity monitoring unit, the handwriting generation unit is configured to generate handwriting corresponding to the first touch event; the quantity monitoring unit is configured to determine whether the generated handwriting corresponds to multiple smart pens of the smart pens according to the identifiers of the smart pens;
      • the writing region generation module includes a matching generation unit and a response control unit, the matching generation unit is configured to generate multiple non-overlapping writing regions under a condition that the generated handwriting corresponds to the multiple smart pens, wherein the multiple writing regions are in one-to-one correspondence with the multiple smart pens, and each writing region covers handwriting of a corresponding smart pen; and the response control unit is configured to control each writing region to only respond to a touch event of the corresponding smart pen for generating handwriting.
  • The embodiments of the present application further provide a writing interaction system, including a smart pen and the smart interactive display device as described above, wherein the smart interactive display device includes a capacitive touch screen, a processor and a computer-readable storage medium, wherein the computer-readable storage medium stores a writing interaction program, and the writing interaction program, when being executed, implements the writing interaction method as described above.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Preferred embodiments of the writing interaction method, the smart interactive display device and the writing interaction system according to the present application will be described below with reference to the accompanying drawings, in which:
  • FIG. 1 is a flowchart of a first embodiment of a writing interaction method according to the present application;
  • FIG. 2 is a flowchart of a second embodiment of a writing interaction method according to the present application;
  • FIG. 3 is a flowchart of a third embodiment of a writing interaction method according to the present application;
  • FIG. 4 is a flowchart of a fourth embodiment of a writing interaction method according to the present application;
  • FIG. 5 is a flowchart of a fifth embodiment of a writing interaction method according to the present application;
  • FIG. 6 is a flowchart of a sixth embodiment of a writing interaction method according to the present application;
  • FIG. 7 is a flowchart of a seventh embodiment of a writing interaction method according to the present application;
  • FIG. 8 is a flowchart of an eighth embodiment of a writing interaction method according to the present application;
  • FIG. 9 is a first state schematic diagram of an implementation process of a writing interaction method according to the present application;
  • FIG. 10 is a second state schematic diagram of an implementation process of a writing interaction method according to the present application;
  • FIG. 11 is a module structure of an embodiment of a writing interaction system according to the present application.
  • DETAILED DESCRIPTION
  • The embodiments of the present application provide a writing interaction method applied to the interaction between a smart interactive display device and a smart pen. Please refer to FIG. 1 , FIG. 1 is a flowchart of a first embodiment of a writing interaction method according to the present application. The writing interaction method includes step 100.
  • Step 100: detecting a first touch event of one or more smart pens and acquiring one or more corresponding identifiers of the one or more smart pens.
  • In this step, the smart pens may generate a touch event on a screen of the smart interactive display device, the touch event includes the above first touch event, and the type of the touch event can be capacitive touch, infrared touch and the like. At the same time, a communication connection is established between the smart pens and the smart interactive device, and the connection between the smart pens and the smart interactive device may transmit the identifiers of the smart pens. For example, an active capacitive pen may establish a communication connection with a touch panel card of a capacitive touch screen through Microsoft Pen Protocol (MPP) of Microsoft or proprietary Active Electrostatic Solution (AES) of Wacom, and transmit its own identifier. It can be understood that, the smart pen may also transmit its own identifier through the electromagnetic induction input technology or the infrared communication technology. The touch events of different smart pens may occur simultaneously or sequentially. From the perspective of application, although there are multiple people operating the smart pens at the same time, the start time of the writing of each person is generally not exactly the same. Technically speaking, it is easier to match touch events that occur one after the other with the corresponding identifiers of the smart pens. As shown in FIG. 9 , the initial stroke of the character “
    Figure US20230315282A1-20231005-P00001
    ” is a horizontal line, and the initial stroke of the character “
    Figure US20230315282A1-20231005-P00002
    ” is a dot. The time when the initial positions of the two strokes are generated can be simultaneous or sequential.
  • The writing interaction method further includes step 102: generating handwriting according to the first touch event, and determining whether the generated handwriting corresponds to multiple smart pens of the smart pens according to the identifiers of the smart pens.
  • In this step, in the example shown in FIG. 1 , although step 102A of generating handwriting is arranged before step 102B of determining the number of the smart pens, determining the number of the smart pens is actually determining the occurred first touch event belongs to which smart pen(s) and then determining the number of the smart pens according to the identifiers of the smart pens. Therefore, apparently, the handwriting is associated with the identifiers of the smart pens, and substantially, the first touch event is associated with the identifiers of the smart pens. Therefore, the sequence between step 102A and step 102B may not be limited. In addition, step 102B can also be performed after the first touch event occurs. For example, in FIG. 9 , the determination of step 102B is completed after the writing of the words “
    Figure US20230315282A1-20231005-P00003
    ” and “
    Figure US20230315282A1-20231005-P00004
    ” is completed.
  • The writing interaction method further includes step 104: under a condition that the handwriting corresponds to the multiple smart pens, generating multiple non-overlapping writing regions, wherein the multiple writing regions are in one-to-one correspondence with the multiple smart pens, each writing region covers handwriting of a corresponding smart pen, and each writing region only responds to a touch event of the corresponding smart pen for generating handwriting.
  • In this step, under a condition that multiple writing regions are generated, it is not necessary to completely fill the entire display region of the smart interactive display device. Optionally, the size and position of the writing region may be determined according to the customary size and the position where the initial first touch event occurs. For example, the two writing regions generated in FIG. 10 are set adjacent to the upper side of the display region, and according to the default left-aligned typesetting mode, the left border of the writing region is placed close to the position where the initial first touch event occurs, that is, the position where the initial handwriting is generated. The left and right writing regions in FIG. 10 do not overlap with each other, but in order to make full use of the display region, the two writing regions can share a common border. It is understandable that the writing region can be determined in the form of a graphic window, or can be defined by only colored lines. After multiple writing regions are generated, since all writing regions are independent of each other, suppose that the left writing region in FIG. 10 belongs to the smart pen marked 010, and the right writing region belongs to the smart pen marked 011. Then, under a condition that the smart pen marked 011 is writing in the left writing region, it will not be responded as displaying handwriting. It is understandable that if the handwriting corresponds to only one smart pen, the default writing region, such as the display region of the entire display screen, is used; or only a suitable size of writing region is generated for the smart pen.
  • According to the writing interaction method, the smart interactive display device and the writing interaction system provided by the embodiments of the present application, the generation source of the first touch event (that is, the smart pens) may be identified. Under a condition that it is detected that the written handwriting comes from multiple smart pens, a corresponding number of writing regions may be automatically generated. Each writing region covers handwriting of a corresponding smart pen, and each writing region only responds to a touch event of the corresponding smart pen for generating handwriting. Therefore, it is ensured that when multiple people use multiple pens to write, the generated multiple writing regions are independent of each other, the content generated by different smart pens will not be cluttered, and the writing content will be clear, thereby improving the user experience of multiple people operating smart pens for writing. In addition, the writing region is generated by matching the initial handwriting of each smart pen, in other words, the writing region is generated with the initial handwriting of each smart pen, and is positioned according to the initial handwriting. Therefore, a more flexible region layout can be realized, thereby adapting to different application scenarios.
  • Further, please refer to FIG. 2 . FIG. 2 is a flowchart of a second embodiment of a writing interaction method according to the present application. After the step 104, the writing interaction method further includes step 206.
  • Step 206: detecting a second touch event of a smart pen and acquiring a corresponding identifier of the smart pen.
  • This step is used to detect the touch event of the smart pen again after the latest generation of the writing regions, that is, to detect the second touch event of the smart pen and match the identifier of the smart pen corresponding to the second touch event. By associating each touch event of the smart pen with the corresponding identifier of the smart pen, a data basis can be provided for subsequent smart control.
  • The writing interaction method further includes step 208: determining whether the second touch event is generated by a newly added smart pen according to the identifier of the smart pen.
  • In this step, although the second touch event has been detected in the previous step, it is not necessary to display corresponding handwriting immediately in response to the second touch event. Since each touch event is associated with the identifier of the smart pen, by querying the data of the identifiers of the smart pens, whether an identifier of a new smart pen is added can be determined, and thus whether the new smart pen or an operator of the new smart pen is added can be indirectly determined.
  • The writing interaction method further includes step 210: under a condition that the second touch event is generated by the newly added smart pen, determining whether the second touch event is generated in the generated writing regions according to a location where the second touch event is generated.
  • In this step, since the identifier of the smart pen corresponding to the second touch event is the identifier of the newly added smart pen, there is no existing writing region matching the second touch event. In order to determine an appropriate response, it is necessary to firstly know whether the location where the second touch event is generated is located in the remaining region outside the writing regions in the display region or in any existing writing regions.
  • The writing interaction method further includes step 212: under a condition that the second touch event is generated in the generated writing regions, generating no handwriting according to the second touch event.
  • In this step, the principle is that the existing writing regions are prioritized and their sizes are not adjusted. Under a condition that the second touch event occurs in the existing writing regions, in order to ensure the independence of the existing writing regions relative to the newly added smart pen, the second touch event of the newly added smart pen is not responded and no handwriting is generated. For example, referring to FIG. 10 , under a condition that a third smart pen with an identification number of 001 is added, then no handwriting is generated if the smart pen with the identification number of 001 writes in any of the left and right writing regions in FIG. 10 .
  • The writing interaction method further includes step 214: under a condition that the second touch event is not generated in the generated writing regions, generating handwriting according to the second touch event, and generating a new writing region corresponding to the newly added smart pen, wherein the new writing region is not overlapped with the existing multiple writing regions, the new writing region covers the handwriting corresponding to the newly added smart pen, and the new writing region only responds to a touch event of the newly added smart pen for generating handwriting.
  • In this step, the principle is that the existing writing regions are prioritized and their sizes are not adjusted. Under a condition that the second touch event is generated in the remaining region of the display region, in order to comprehensively respond to the second touch event of the newly added smart pen and generate corresponding handwriting, a new writing region is generated for the newly added smart pen. For example, referring to FIG. 10 , under a condition that a third smart pen with an identification number of 001 is newly added, a new writing region is generated if the second touch event of the smart pen is generated near the lower edge of the display region and outside the existing left and right writing regions.
  • In this embodiment, by establishing the principle of giving priority to the existing writing regions, the high independence of the existing writing regions can be ensured. Further, within the remaining region of the display region, a new writing region can be generated for the second touch event of the newly added smart pen, thereby ensuring a certain flexibility.
  • Further, please refer to FIG. 3 . FIG. 3 is a flowchart of a third embodiment of a writing interaction method according to the present application. After the step 104, the writing interaction method further includes steps 306, 308 and 310.
  • Step 306: detecting a third touch event of a smart pen and acquiring a corresponding identifier of the smart pen.
  • Step 308: determining whether the third touch event is generated by a newly added smart pen according to the identifier of the smart pen.
  • Step 310: under a condition that the third touch event is generated by the newly added smart pen, determining whether the third touch event is generated in the generated writing regions according to a location where the third touch event is generated.
  • In this step, on the basis that the response strategy for the newly added smart pen is determined, in order to execute the corresponding strategy smoothly, it is necessary to firstly know whether the location where the third touch event is generated is located in the remaining region outside the writing regions in the display region or in any existing writing regions.
  • The writing interaction method further includes step 312: under a condition that the third touch event is generated in the generated writing regions, regenerating multiple non-overlapping writing regions, wherein the regenerated multiple writing regions are in one-to-one correspondence with the multiple smart pens and the newly added smart pen, each writing region covers a touch track of a corresponding smart pen, and each writing region only responds to a touch event of the corresponding smart pen for generating handwriting.
  • In this step, since it is determined that the newly added smart pen has priority to the existing writing regions, each time a new smart pen is added, the multiple writing regions need to be regenerated. For example, under a condition that the third touch event of the newly added smart pen falls in the left or right writing region, the multiple writing regions need to be regenerated. For example, under a condition that the newly added smart pen writes “
    Figure US20230315282A1-20231005-P00005
    ” between the words “
    Figure US20230315282A1-20231005-P00006
    ” and “
    Figure US20230315282A1-20231005-P00007
    ”, the regenerated writing regions can include three writing regions (left, middle and right), with the word “
    Figure US20230315282A1-20231005-P00008
    ” in the middle writing region.
  • The writing interaction method further includes step 314: under a condition that third touch event is not generated in the generated writing regions, generating a new writing region corresponding to the newly added smart pen, wherein the existing multiple writing areas remain unchanged, wherein the new writing region is not overlapped with the existing multiple writing regions, the new writing region covers handwriting corresponding to the newly added smart pen, and the new writing region only responds to a touch event of the newly added smart pen for generating handwriting.
  • In this step, under a condition that the third touch event of the newly added smart pen falls in the remaining region of the display region, a new writing region is generated for the newly added smart pen, and the existing multiple writing regions remain unchanged. In this way, the third touch event of the newly added smart pen can be flexibly responded, and the independence of the existing writing regions can be maintained. For example, under a condition that the third touch event of the newly added smart pen occurs at the lower right corner of the display region, then a new writing region is generated for the newly added smart pen at the lower right corner, while keeping the original left and right writing regions unchanged.
  • In this embodiment, by establishing the principle of giving priority to the newly added smart pen, it can be ensured that a writing region can be generated for the newly added smart pen. It is understandable that by using this principle, the sizes of the existing writing regions can be adjusted if necessary.
  • Further, please refer to FIG. 4 . FIG. 4 is a flowchart of a fourth embodiment of a writing interaction method according to the present application. The writing interaction method further includes steps 406, 408 and 410.
  • Step 406: detecting a fourth touch event of a smart pen and acquiring a corresponding identifier of the smart pen.
  • Step 408: determining whether the fourth touch event is generated by a newly added smart pen according to the identifier of the smart pen.
  • Step 410: under a condition that the fourth touch event is not generated by the newly added smart pen, determining whether the fourth touch event is generated in the generated writing regions according to a location where the fourth touch event is generated.
  • In this step, it is understandable that, in some application scenarios, whether it is an existing smart pen or a newly added smart pen, the fourth touch event generated by it may fall within the generated writing regions or may fall outside the generated writing regions. However, when considering whether to expand the generated writing regions, according to the principle of not overlapping with other writing regions, it is especially necessary to consider touch events that fall in the remaining region of the display region.
  • The writing interaction method further includes step 412: under a condition that the fourth touch event is not generated in the generated writing regions, tentatively expanding the writing region corresponding to the smart pen according to a preset rule, so as to cover a track corresponding to the fourth touch event.
  • In this step, the writing region is usually a rectangle, but in some embodiments, the writing region can also be other shapes, such as an ellipse. The preset rule is to expand the corresponding writing region according to the preset rule to cover the track of the fourth touch event. For example, please refer to FIG. 10 , the right writing region belongs to the smart pen marked as 011. Under a condition that the smart pen marked as 011 generates a fourth touch event near the lower edge of the writing region, the right writing region can be tentatively expanded downward. Under a condition that the smart pen marked as 011 generates a fourth touch event near the left edge of the writing region, the right writing region can be tentatively expanded to the left.
  • The writing interaction method further includes step 414: determining whether the tentatively expanded writing region overlaps with other writing regions.
  • In this step, under the principle of ensuring the independence of the existing writing regions, not all tentative expansions are suitable, and the expanded region still needs to keep from overlapping with other writing regions.
  • The writing interaction method further includes step 416: under a condition that the tentatively expanded writing region overlaps with the other writing regions, canceling expansion of the writing region corresponding to the smart pen and generating no handwriting according to the fourth touch event.
  • In this step, for example, following the example in step 412, under a condition that the right writing region is extended to the left, the right writing region will overlap the left writing region. Therefore, the expansion of the right writing region should be canceled, and no handwriting will be generated according to the fourth touch event.
  • The writing interaction method further includes step 418: under a condition that the tentatively expanded writing region is not overlapped with the other writing regions, expanding the writing region corresponding to the smart pen and generating handwriting according to the fourth touch event.
  • In this step, for example, following the example in step 412, the right writing region is extended downward, for example, the right writing region expands downward until the lower edge is flush with the lower edge of the left writing region, thereby actually extending the writing region corresponding to the smart pen, and generating handwriting in respond to the fourth touch event.
  • In this embodiment, it is considered that the initially generated writing region may not be enough for writing. Therefore, under a condition that the touch event of the corresponding smart pen falls outside the original writing region, the original writing region may be tentatively expanded. Under the premise of not overlapping with other writing regions, the original writing region may be expanded, so as to more flexibly adapt to the needs of the users.
  • Further, under a condition that the handwriting corresponds to the multiple smart pens, the generating multiple non-overlapping writing regions specifically includes:
      • acquiring a location of a leftmost coordinate point of the handwriting corresponding to each smart pen;
      • determining whether a distance between two adjacent leftmost coordinate points is greater than or equal to a first preset distance;
      • under a condition that the distance between the two adjacent leftmost coordinate points is greater than or equal to the first preset distance, generating the multiple non-overlapping writing regions on a screen display region.
  • Further, under a condition that the distance between the two adjacent leftmost coordinate points is greater than or equal to the first preset distance, generating the multiple non-overlapping writing regions on a screen display region includes:
      • under a condition that the distance between the two adjacent leftmost coordinate points is greater than or equal to the first preset distance, except for a leftmost coordinate point close to a left side of the screen display region, forming the writing regions by setting vertical dividing lines referring to the other leftmost coordinate points, wherein the first preset distance is equal to a product of δ and L, L is a horizontal length of the screen display region, and δ is less than 0.5; or
      • under a condition that the distance between the two adjacent leftmost coordinate points is greater than or equal to the first preset distance, forming the writing regions by using a vertical centerline of the screen display region as a dividing line, wherein the first preset distance is equal to L/2, where L is the horizontal length of the screen display region.
  • Further, please refer to FIG. 5 . FIG. 5 is a flowchart of a fifth embodiment of a writing interaction method according to the present application. The generating multiple non-overlapping writing regions specifically includes step 541.
  • Step 541: acquiring a location of a leftmost coordinate point of the handwriting corresponding to each smart pen.
  • In this step, since left alignment is the most commonly used typesetting method, in order to adapt to this typesetting method, the leftmost coordinate point of the handwriting of each smart pen is obtained as a reference point for generating the writing region. For example, in FIG. 9 , the leftmost point of the word “
    Figure US20230315282A1-20231005-P00009
    ” is P2, and its coordinates are (c, d), and the leftmost point of the word “
    Figure US20230315282A1-20231005-P00010
    ” is P1, and its coordinates are (a, b).
  • The generating multiple non-overlapping writing regions further includes step 543: determining whether a distance between two adjacent leftmost coordinate points is greater than or equal to a product of δ and L, wherein L is a horizontal length of the screen display region, and δ is less than 0.5.
  • In this step, following the example in step 541, δ=⅓, it is determined whether the distance between P1 and P2 is greater than or equal to L/3, that is, whether the value of a-c is greater than or equal to L/3. The distance can cover the possible horizontal length of the left handwriting among the two adjacent handwritings in the horizontal direction, thereby ensuring that a vertical dividing line can be set between the two handwritings. The horizontal lengths of the two handwritings are generally determined by the time required for determining the number of the smart pens, and the horizontal length is usually less than L/3. Of course, the time required for the determination can also be extended a little longer, so as to extend the length occupied by the initial handwriting of each smart pen.
  • The generating multiple non-overlapping writing regions further includes step 545: under a condition that the distance between the two adjacent leftmost coordinate points is greater than or equal to the product of δ and L, except for a leftmost coordinate point close to a left side of the screen display region, forming the writing regions by setting vertical dividing lines referring to the other leftmost coordinate points.
  • In this step, for the leftmost coordinate point close to the left side of the screen display region, the writing region corresponding to the handwriting can be bounded by the left side of the display region, like the left writing region in FIG. 10 . For the handwriting of the other smart pens, the corresponding writing regions may be formed by setting vertical dividing lines referring to the leftmost coordinate points in turn. It is understandable that the most basic principle of “referring” is not to divide handwriting. Therefore, for example, the most extreme value of the boundary line X in the left writing region in FIG. 10 is X=a, and a is the abscissa of P1. On this basis, the dividing line X can be shifted to the left by a certain distance, that is, X=a−a1, a1 is a constant, such as a1=5 cm, which can be different according to the size of the screen in practice.
  • In this embodiment, the touch screen of the smart interactive display device is usually a wide screen, such as a 16:9 wide screen, and multiple people are standing along the horizontal direction in sequence when writing. Therefore, under a condition that multiple people are writing at the same time, it is also appropriate to only consider dividing the display region horizontally to generate the writing regions, and the algorithm required for this setup is relatively simpler.
  • Further, please refer to FIG. 6 . FIG. 6 is a flowchart of a sixth embodiment of a writing interaction method according to the present application. The step 104 specifically includes step 642.
  • Step 642: under a condition that the handwriting corresponds to the multiple smart pens, and under a condition that the generated handwriting corresponds to two smart pens, acquiring a location of a leftmost coordinate point of the handwriting corresponding to each smart pen.
  • In this step, which is the same as step 541, since left alignment is the most commonly used typesetting method, in order to adapt to this typesetting method, the leftmost coordinate point of the handwriting of each smart pen is obtained as a reference point for generating the writing region.
  • The step 104 further includes step 644: determining whether a distance between two adjacent leftmost coordinate points is greater than or equal to L/2, wherein L is a horizontal length of the screen display region.
  • In this step, please refer to FIG. 9 , the distance between P1 and P2 is greater than or equal to L/2, that is, the value of a-c is greater than or equal to L/2. It is indicated that the distance between P1 and P2 is large enough to completely cover the possible horizontal length of the left handwriting among the two adjacent handwritings in the horizontal direction, thereby ensuring greater flexibility in setting a vertical dividing line between the two handwritings.
  • The step 104 further includes step 646: under a condition that the distance between the two adjacent leftmost coordinate points is greater than or equal to L/2, forming the writing regions by using a vertical centerline of the screen display region as a dividing line.
  • In this step, considering the greater flexibility in setting a vertical dividing line between the two handwritings, in order to facilitate the typesetting of the two handwritings, the vertical centerline of the display region is used as the boundary to form two writing regions of the same size.
  • In this embodiment, under a condition that the generated handwriting corresponds to two smart pens, and there is a sufficient distance between the two handwritings, that is, the distance between the two handwritings is L/2 or more, in order to better typeset the display region and coordinate the space utilization of the display region, the writing region corresponding to different handwriting may not be formed by setting the vertical dividing line referring to the leftmost coordinate point of the handwriting, but the vertical centerline of the display region may be used as the boundary to form two writing regions of the same size.
  • Further, please refer to FIG. 7 . FIG. 7 is a flowchart of a seventh embodiment of a writing interaction method according to the present application. The step 104 includes step 703A.
  • Step 703A: under a condition that the handwriting corresponds to the multiple smart pens, popping up an instruction window requesting a user to confirm whether to generate the multiple writing regions.
  • In this step, in order to give the user the right to choose whether to generate multiple writing regions, a pop-up window is set to request the user to confirm.
  • The step 104 further includes step 703B: monitoring an instruction command input by the user in the instruction window.
  • In this step, the input instruction command may be a touch event, a voice command, or a character command input from a keyboard. By monitoring the instruction command, it is convenient for the machine to perform subsequent actions according to the user's instruction.
  • The step 104 further includes step 703C: under a condition that a command of determining to generate the multiple writing regions is received from the user, generating the multiple non-overlapping writing regions, wherein the multiple writing regions are in one-to-one correspondence with the multiple smart pens, each writing region covers handwriting of a corresponding smart pen, and each writing region only responds to a touch event of the corresponding smart pen for generating handwriting.
  • In this step, multiple writing regions are generated only when a confirmed command is received, so as to more accurately meet the needs of the user.
  • After monitoring the instruction command input by the user in the instruction window, the writing interaction method further includes step 705.
  • Step 705: under a condition that a command of cancelling generation of the multiple writing regions is received from the user, maintaining original writing regions.
  • In this step, under a condition that a canceling command is received, the original writing regions are maintained, for example, multiple smart pens write in a same writing region. It is understandable that, optionally, under a condition that the newly added smart pen writes in one of the generated multiple writing regions, and the user chooses to cancel the instruction command of regenerating multiple writing regions, the identifier of the newly added smart pen may be included in that writing region, that is, the writing region can accept the writing input of two smart pens.
  • In this embodiment, a user confirmation process is set before determining whether to generate multiple writing regions, so that the smart interactive display device can more accurately meet the needs of the user, thereby achieving a better user experience.
  • Further, each writing region includes an editing region and a menu bar region, and the editing region of each writing region covers the handwriting corresponding to the smart pen and corresponds to the touch event of the smart pen; the menu bar region includes a writing main menu, and submenus of the writing main menu include color, eraser, and stroke thickness submenus.
  • In this embodiment, as shown in FIG. 10 , the upper parts of the left and right writing regions are provided with a menu bar region, the square patterns in the menu bar in the figure represents graphic function buttons or icons, and the outside of the menu bar is the editing region. Of course, in some embodiments, the menu bar region can also be set to be hidden and can be called up when needed. Specifically, the user can set the color of the handwriting by invoking the color command, or erase the generated handwriting by invoking the eraser, and can also set the thickness of the handwriting by invoking the stroke thickness. By setting a menu bar region in each writing region, the user can independently edit the handwriting in each writing region.
  • Further, please refer to FIG. 8 . FIG. 8 is a flowchart of an eighth embodiment of a writing interaction method according to the present application. The menu bar region includes a partition canceling main menu; the writing interaction method further includes step 806.
  • Step 806: monitoring the input command of the user in the menu bar.
  • In this step, which is similar to step 703B, the input instruction command may be a touch event, a voice command, or a character command input from a keyboard. By monitoring the instruction command, it is convenient for the machine to perform subsequent actions according to the user's instruction.
  • The writing interaction method further includes step 810: under a condition that the command of canceling the partitions is detected, restoring the writing regions to the original single writing region, and deleting the content in each of the writing regions.
  • In this step, under a condition that the user wants to cancel the partitions, a corresponding trigger command can be conveniently input in the menu bar, which restores the writing regions to the original single region, and deletes the content to quickly clear the screen. For example, in the classroom scene, the math class is coming to an end and the next class is English class. The display region of the smart interactive display device has multiple writing regions, all operators have finished writing, and the teacher or any operator can trigger the command of canceling the partitions. The screen can be quickly cleared, and it is convenient to use the writing interaction method of the smart interactive display device of the present application in English class.
  • In this embodiment, when the user wants to cancel multiple writing regions that have been generated and clear the screen, the user can use the partition canceling main menu to quickly achieve the goal.
  • The present application also provides a smart interactive display device for interacting with a smart pen. Please refer to FIG. 11 , the smart interactive display device includes a touch event matching module, a touch event response module, and a writing area generation module, wherein,
      • the touch event matching module includes a touch detection unit and a touch matching unit, the touch detection unit is configured to detect a first touch event of one or more smart pens, and the touch matching unit is configured to acquire one or more identifiers of the one or more smart pens and match the corresponding first touch event;
      • the touch event response module includes a handwriting generation unit and a quantity monitoring unit, the handwriting generation unit is configured to generate handwriting corresponding to the first touch event; the quantity monitoring unit is configured to determine whether the generated handwriting corresponds to multiple smart pens of the smart pens according to the identifiers of the smart pens;
      • the writing region generation module includes a matching generation unit and a response control unit, the matching generation unit is configured to generate multiple non-overlapping writing regions under a condition that the generated handwriting corresponds to the multiple smart pens, wherein the multiple writing regions are in one-to-one correspondence with the multiple smart pens, and each writing region covers handwriting of a corresponding smart pen; and the response control unit is configured to control each writing region to only respond to a touch event of the corresponding smart pen for generating handwriting.
  • Further, after the matching generation unit is configured to generate multiple non-overlapping writing regions,
      • the touch detection unit is further configured to detect a second touch event of a smart pen, and the touch matching unit is further configured to acquire a identifier of the smart pen and match the corresponding second touch event;
      • the quantity monitoring unit is further configured to determine whether the second touch event is generated by a newly added smart pen according to the identifier of the smart pen;
      • the response control unit is further configured to determine whether the second touch event is generated in the generated writing regions according to a location where the second touch event is generated;
      • under a condition that the second touch event is generated in the generated writing regions, the handwriting generation unit is configured to generate no handwriting according to the second touch event, that is, the handwriting generation unit does not respond to generate handwriting;
      • under a condition that the second touch event is not generated in the generated writing regions, the handwriting generation unit is configured to generate handwriting according to the second touch event; the matching generation unit is further configured to generate a new writing region corresponding to the newly added smart pen, wherein the new writing region is not overlapped with the existing multiple writing regions, the new writing region covers the handwriting corresponding to the newly added smart pen, and the new writing region only responds to a touch event of the newly added smart pen for generating handwriting.
  • Further, after the matching generation unit is configured to generate multiple non-overlapping writing regions,
      • the touch detection unit is further configured to detect a third touch event of a smart pen, and the touch matching unit is further configured to acquire an identifier of the smart pen and match the corresponding third touch event;
      • the quantity monitoring unit is further configured to determine whether the third touch event is generated by a newly added smart pen according to the identifier of the smart pen;
      • under a condition that the third touch event is generated by the newly added smart pen, the response control unit is further configured to determine whether the third touch event is generated in the generated writing regions according to a location where the third touch event is generated;
      • under a condition that the third touch event is generated in the generated writing regions, the matching generation unit is further configured to regenerate multiple non-overlapping writing regions, wherein the regenerated multiple writing regions are in one-to-one correspondence with the multiple smart pens and the newly added smart pen, each writing region covers a touch track of a corresponding smart pen, and each writing region only responds to a touch event of the corresponding smart pen for generating handwriting;
      • under a condition that third touch event is not generated in the generated writing regions, the matching generation unit is further configured to generate a new writing region corresponding to the newly added smart pen, wherein the new writing region is not overlapped with the existing multiple writing regions, the new writing region covers handwriting corresponding to the newly added smart pen, and the new writing region only responds to a touch event of the newly added smart pen for generating handwriting.
  • Further, after the matching generation unit is configured to generate multiple non-overlapping writing regions,
      • the touch detection unit is further configured to detect a fourth touch event of a smart pen, and the touch matching unit is further configured to acquire an identifier of the smart pen and match the corresponding fourth touch event;
      • the quantity monitoring unit is further configured to determine whether the fourth touch event is generated by a newly added smart pen according to the identifier of the smart pen;
      • under a condition that the fourth touch event is not generated by the newly added smart pen, the response control unit is further configured to determine whether the fourth touch event is generated in the generated writing regions according to a location where the fourth touch event is generated;
      • under a condition that the fourth touch event is not generated in the generated writing regions, the matching generation unit is further configured to tentatively expand the writing region corresponding to the smart pen according to a preset rule, so as to cover a track corresponding to the fourth touch event, and determine whether the tentatively expanded writing region overlaps with other writing regions;
      • under a condition that the tentatively expanded writing region overlaps with the other writing regions, the matching generation unit is further configured to cancel expansion of the writing region corresponding to the smart pen, and the handwriting generation unit is configured to generate no handwriting according to the fourth touch event;
      • under a condition that the tentatively expanded writing region is not overlapped with the other writing regions, the matching generation unit is further configured to expand the writing region corresponding to the smart pen, and the handwriting generation unit is configured to generate handwriting according to the fourth touch event.
  • Further, the matching generation unit includes a reference point acquisition unit, a distance determination unit and a region generation unit, wherein,
      • the reference point acquisition unit is configured to acquire a location of a leftmost coordinate point of the handwriting corresponding to each smart pen under a condition that the generated handwriting corresponds to the multiple smart pens;
      • the distance determination unit is configured to determine whether a distance between two adjacent leftmost coordinate points is greater than or equal to a first preset distance;
      • under a condition that the result of the distance determination unit is yes, the region generation unit is configured to generate the multiple non-overlapping writing regions on a screen display region.
  • Further, the region generation unit is specifically configured to:
      • under a condition that the distance between the two adjacent leftmost coordinate points is greater than or equal to the first preset distance, except for a leftmost coordinate point close to a left side of the screen display region, form the writing regions by setting vertical dividing lines referring to the other leftmost coordinate points, wherein the first preset distance is equal to a product of δ and L, L is a horizontal length of the screen display region, and δ is less than 0.5; or
      • under a condition that the distance between the two adjacent leftmost coordinate points is greater than or equal to the first preset distance, form the writing regions by using a vertical centerline of the screen display region as a dividing line, wherein the first preset distance is equal to L/2, where L is the horizontal length of the screen display region.
  • Further, the writing region generation module also includes a partition pop-up window unit and a command monitoring unit, wherein,
      • under a condition that the generated handwriting corresponds to the multiple smart pens, the partition pop-up window unit is configured to pop up an instruction window requesting a user to confirm whether to generate the multiple writing regions;
      • the command monitoring unit is configured to monitor an instruction command input by the user in the instruction window;
      • under a condition that a command of determining to generate the multiple writing regions is received from the user, the matching generation unit is further configured to generate the multiple non-overlapping writing regions, wherein the multiple writing regions are in one-to-one correspondence with the multiple smart pens, each writing region covers handwriting of a corresponding smart pen, and each writing region only responds to a touch event of the corresponding smart pen for generating handwriting,
      • under a condition that a command of cancelling generation of the multiple writing regions is received from the user, the matching generation unit is further configured to maintain original writing regions.
  • Further, each writing region includes an editing region and a menu bar region, and the editing region of each writing region covers the handwriting corresponding to the smart pen and corresponds to the touch event of the smart pen; the menu bar region includes a writing main menu, and submenus of the writing main menu include color, eraser, and stroke thickness submenus.
  • Further, the writing region generation module further includes a command monitoring unit, the command monitoring unit monitors the input command of the user in the menu bar; under a condition that the command monitoring unit detects the command of canceling the partitions, the matching generation module is further configured to restore the writing regions to the original single writing region, and delete the content in each of the writing regions.
  • For the specific description of each module and unit of the smart interactive display device involved in each embodiment of the present application, reference may be made to the specific description of the embodiments corresponding to the writing interaction method, which is not repeated here.
  • The present application also provides a smart interactive display device, including a capacitive touch screen, a processor, and a computer-readable storage medium, wherein the computer-readable storage medium stores a writing interaction program, and the writing interaction program, when being executed, implements the writing interaction method as described above. The specific steps of the writing interaction method may refer to the above embodiments. Since the smart interactive display device adopts all technical solutions of the above embodiments, it has at least all beneficial effects brought by the technical solutions of the above embodiments, will not be repeated here.
  • The present application also provides a writing interactive system, including a smart pen and the above smart interactive display device including various modules. The specific structure of the smart interactive display device may refer to the above embodiments. Since the smart interactive display device adopts all technical solutions of the above embodiments, it has at least all beneficial effects brought by the technical solutions of the above embodiments, will not be repeated here.
  • The present application also provides a writing interactive system, including an active capacitive smart pen and the above smart interactive display device including the computer storage medium. The specific structure of the smart interactive display device and the steps implemented by executing the writing interactive program may refer to the above embodiments. Since the writing interactive system adopts all technical solutions of the above embodiments, it has at least all beneficial effects brought by the technical solutions of the above embodiments, will not be repeated here.
  • It should be noted that, in the present application, the terms “comprise”, “include”, or any other variants thereof, are intended to represent a non-exclusive inclusion, such that a process, method, article or device including a series of elements includes not only those elements, but also other elements that are not explicitly listed or elements inherent to such a process, method, article or device. Without more constraints, the elements following an expression “comprise/include . . . ” do not exclude the existence of additional identical elements in the process, method, article or device that includes the elements.
  • In the present application, relational terms, such as first and second, are used merely to distinguish one entity or operation from another entity or operation, without necessarily requiring or implying any actual such relationships or orders of these entities or operations.
  • The above serial numbers of the embodiments of the present application are only for description, and do not represent the advantages or disadvantages of the embodiments.
  • According to the description of the above embodiments, those skilled in the art can clearly understand that the methods of the above embodiments can be implemented by means of software plus a necessary general hardware platform, and of course can also be implemented by hardware or software, but in many cases, the former is a better implementation. Based on this understanding, the technical solutions of the present application or the parts that make contributions to the prior art can be embodied in the form of software products. The computer software products are stored in a storage medium (such as a ROM/RAM, a magnetic disk, an optical disk) as described above, including several instructions to make a terminal device (which may be a mobile phone, a computer, a server, an air conditioner, or a network device) execute the methods described in the various embodiments of the present application.
  • Those skilled in the art can understand that, on the premise of no conflict, the above solutions can be combined and superimposed freely.
  • It should be understood that the above embodiments are only exemplary rather than restrictive, and without departing from the basic principles of the present application, those skilled in the art can make various obvious or equivalent modifications or substitutions to the above details, which will all be included within the scope of the claims of the present application.

Claims (11)

1. A writing interaction method, comprising:
detecting a first touch event of one or more smart pens and acquiring one or more corresponding identifiers of the one or more smart pens;
generating handwriting according to the first touch event, and determining whether the generated handwriting corresponds to multiple smart pens of the smart pens according to the identifiers of the smart pens;
under a condition that the handwriting corresponds to the multiple smart pens, generating multiple non-overlapping writing regions, wherein the multiple writing regions are in one-to-one correspondence with the multiple smart pens, each writing region covers handwriting of a corresponding smart pen, and each writing region only responds to a touch event of the corresponding smart pen for generating handwriting.
2. The writing interaction method according to claim 1, further comprising:
detecting a second touch event of a smart pen and acquiring a corresponding identifier of the smart pen;
determining whether the second touch event is generated by a newly added smart pen according to the identifier of the smart pen;
under a condition that the second touch event is generated by the newly added smart pen, determining whether the second touch event is generated in the generated writing regions according to a location where the second touch event is generated;
under a condition that the second touch event is generated in the generated writing regions, generating no handwriting according to the second touch event;
under a condition that the second touch event is not generated in the generated writing regions, generating handwriting according to the second touch event, and generating a new writing region corresponding to the newly added smart pen, wherein the new writing region is not overlapped with the existing multiple writing regions, the new writing region covers the handwriting corresponding to the newly added smart pen, and the new writing region only responds to a touch event of the newly added smart pen for generating handwriting.
3. The writing interaction method according to claim 1, further comprising:
detecting a third touch event of a smart pen and acquiring a corresponding identifier of the smart pen;
determining whether the third touch event is generated by a newly added smart pen according to the identifier of the smart pen;
under a condition that the third touch event is generated by the newly added smart pen, determining whether the third touch event is generated in the generated writing regions according to a location where the third touch event is generated;
under a condition that the third touch event is generated in the generated writing regions, regenerating multiple non-overlapping writing regions, wherein the regenerated multiple writing regions are in one-to-one correspondence with the multiple smart pens and the newly added smart pen, each writing region covers a touch track of a corresponding smart pen, and each writing region only responds to a touch event of the corresponding smart pen for generating handwriting;
under a condition that third touch event is not generated in the generated writing regions, generating a new writing region corresponding to the newly added smart pen, wherein the new writing region is not overlapped with the existing multiple writing regions, the new writing region covers handwriting corresponding to the newly added smart pen, and the new writing region only responds to a touch event of the newly added smart pen for generating handwriting.
4. The writing interaction method according to claim 1, further comprising:
detecting a fourth touch event of a smart pen and acquiring a corresponding identifier of the smart pen;
determining whether the fourth touch event is generated by a newly added smart pen according to the identifier of the smart pen;
under a condition that the fourth touch event is not generated by the newly added smart pen, determining whether the fourth touch event is generated in the generated writing regions according to a location where the fourth touch event is generated;
under a condition that the fourth touch event is not generated in the generated writing regions, tentatively expanding the writing region corresponding to the smart pen according to a preset rule, so as to cover a track corresponding to the fourth touch event;
determining whether the tentatively expanded writing region overlaps with other writing regions;
under a condition that the tentatively expanded writing region overlaps with the other writing regions, canceling expansion of the writing region corresponding to the smart pen and generating no handwriting according to the fourth touch event;
under a condition that the tentatively expanded writing region is not overlapped with the other writing regions, expanding the writing region corresponding to the smart pen and generating handwriting according to the fourth touch event.
5. The writing interaction method according to claim 1, wherein the step of under a condition that the handwriting corresponds to the multiple smart pens, generating multiple non-overlapping writing regions specifically comprises:
acquiring a location of a leftmost coordinate point of the handwriting corresponding to each smart pen;
determining whether a distance between two adjacent leftmost coordinate points is greater than or equal to a first preset distance;
under a condition that the distance between the two adjacent leftmost coordinate points is greater than or equal to the first preset distance, generating the multiple non-overlapping writing regions on a screen display region.
6. The writing interaction method according to claim 5, wherein under a condition that the distance between the two adjacent leftmost coordinate points is greater than or equal to the first preset distance, generating the multiple non-overlapping writing regions on a screen display region comprises:
under a condition that the distance between the two adjacent leftmost coordinate points is greater than or equal to the first preset distance, except for a leftmost coordinate point close to a left side of the screen display region, forming the writing regions by setting vertical dividing lines referring to the other leftmost coordinate points, wherein the first preset distance is equal to a product of δ and L, L is a horizontal length of the screen display region, and δ is less than 0.5; or
under a condition that the distance between the two adjacent leftmost coordinate points is greater than or equal to the first preset distance, forming the writing regions by using a vertical centerline of the screen display region as a dividing line, wherein the first preset distance is equal to L/2, where L is the horizontal length of the screen display region.
7. The writing interaction method according to claim 1, wherein the step of under a condition that the handwriting corresponds to the multiple smart pens, generating multiple non-overlapping writing regions comprises:
under a condition that the handwriting corresponds to the multiple smart pens, popping up an instruction window requesting a user to confirm whether to generate the multiple writing regions;
monitoring an instruction command input by the user in the instruction window;
under a condition that a command of determining to generate the multiple writing regions is received from the user, generating the multiple non-overlapping writing regions, wherein the multiple writing regions are in one-to-one correspondence with the multiple smart pens, each writing region covers handwriting of a corresponding smart pen, and each writing region only responds to a touch event of the corresponding smart pen for generating handwriting,
after monitoring the instruction command input by the user in the instruction window, the method further comprises:
under a condition that a command of cancelling generation of the multiple writing regions is received from the user, maintaining original writing regions.
8. A smart interactive display device for interacting with a smart pen, the smart interactive display device comprising a touch event matching module, a touch event response module, and a writing region generation module, wherein
the touch event matching module comprises a touch detection unit and a touch matching unit, the touch detection unit is configured to detect a first touch event of one or more smart pens, and the touch matching unit is configured to acquire one or more identifiers of the one or more smart pens and match the corresponding first touch event;
the touch event response module comprises a handwriting generation unit and a quantity monitoring unit, the handwriting generation unit is configured to generate handwriting corresponding to the first touch event; the quantity monitoring unit is configured to determine whether the generated handwriting corresponds to multiple smart pens of the smart pens according to the identifiers of the smart pens;
the writing region generation module comprises a matching generation unit and a response control unit, the matching generation unit is configured to generate multiple non-overlapping writing regions under a condition that the generated handwriting corresponds to the multiple smart pens, wherein the multiple writing regions are in one-to-one correspondence with the multiple smart pens, and each writing region covers handwriting of a corresponding smart pen; and the response control unit is configured to control each writing region to only respond to a touch event of the corresponding smart pen for generating handwriting.
9. A smart interactive display device, comprising a capacitive touch screen, a processor and a computer-readable storage medium, wherein the computer-readable storage medium stores a writing interaction program, and the writing interaction program, when being executed, implements the writing interaction method according to claim 1.
10. A writing interaction system, comprising a smart pen and the smart interactive display device according to claim 8.
11. A writing interaction system, comprising a smart pen and the smart interactive display device according to claim 9.
US17/913,853 2020-03-24 2020-07-06 Writing interaction method, smart interactive display device and writing interaction system Active US11861160B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
CN202010213966.9 2020-03-24
CN202010213966.9A CN111352570B (en) 2020-03-24 2020-03-24 Writing interaction method, intelligent interaction display equipment and writing interaction system
PCT/CN2020/100379 WO2021189706A1 (en) 2020-03-24 2020-07-06 Writing interaction method, smart interactive display device and writing interaction system

Publications (2)

Publication Number Publication Date
US20230315282A1 true US20230315282A1 (en) 2023-10-05
US11861160B2 US11861160B2 (en) 2024-01-02

Family

ID=71194555

Family Applications (1)

Application Number Title Priority Date Filing Date
US17/913,853 Active US11861160B2 (en) 2020-03-24 2020-07-06 Writing interaction method, smart interactive display device and writing interaction system

Country Status (4)

Country Link
US (1) US11861160B2 (en)
EP (1) EP4123439A4 (en)
CN (1) CN111352570B (en)
WO (1) WO2021189706A1 (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111352570B (en) 2020-03-24 2021-06-01 深圳市鸿合创新信息技术有限责任公司 Writing interaction method, intelligent interaction display equipment and writing interaction system
CN112578987A (en) * 2020-12-25 2021-03-30 广州壹创电子科技有限公司 Off-screen interactive touch all-in-one machine and interaction method thereof
CN113434064B (en) * 2021-07-01 2022-03-29 掌阅科技股份有限公司 Handwriting reader stroke switching method, electronic equipment and storage medium
CN113934323B (en) * 2021-10-19 2023-12-29 河北师达教育科技有限公司 Multi-point display method and device based on intelligent blackboard and terminal equipment
CN115937861B (en) * 2022-11-30 2023-09-08 广东保伦电子股份有限公司 Multi-user synchronous writing identification method and system based on touch screen
CN118414596A (en) * 2022-11-30 2024-07-30 京东方科技集团股份有限公司 Intelligent handwriting display device, intelligent handwriting display method and electronic device

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10606382B2 (en) * 2016-09-01 2020-03-31 Wacom Co., Ltd. Stylus and sensor controller for bi-directional communication using stylus identifier
US10698524B2 (en) * 2016-04-22 2020-06-30 Samsung Electronics Co., Ltd. Touch screen device, input device, and control method thereof and method thereof
US11086419B2 (en) * 2019-05-22 2021-08-10 Sharp Kabushiki Kaisha Information processing device, information processing method, and recording medium

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3994183B2 (en) * 1998-07-28 2007-10-17 キヤノン株式会社 Display control apparatus, display control method, and storage medium
JP5664303B2 (en) * 2011-02-09 2015-02-04 大日本印刷株式会社 Computer apparatus, input system, and program
TWI474221B (en) * 2011-08-24 2015-02-21 Dexin Corp Wireless transmission method for touch pen with wireless storage and forwarding capability and system thereof
KR20130138880A (en) * 2012-06-11 2013-12-20 삼성전자주식회사 Method and apparatus for controlling a input in a terminal equipment having a touch input device
CN105320390A (en) * 2014-06-20 2016-02-10 鸿合科技有限公司 Double handwriting recognition method and device based on electromagnetic white plate and electromagnetic pens
CN104571815B (en) * 2014-12-15 2019-10-29 联想(北京)有限公司 A kind of matching process and electronic equipment of display window
CN106339135A (en) 2016-08-30 2017-01-18 科盟(福州)电子科技有限公司 Infrared electronic whiteboard A/B screen splitting method capable of supporting independent operation by multiple persons
CN106547402A (en) * 2016-10-31 2017-03-29 广州华欣电子科技有限公司 A kind of touch control method, touch frame and smart pen
CN106775314A (en) 2016-12-09 2017-05-31 珠海市魅族科技有限公司 Multi-screen display method and split screen display available device
CN107515690B (en) * 2017-07-06 2021-02-05 广州视源电子科技股份有限公司 Electromagnetic screen writing operation method and electromagnetic screen
CN107491210B (en) * 2017-08-14 2020-06-09 广州视源电子科技股份有限公司 Multi-electromagnetic-pen writing distinguishing method and device and electronic equipment
CN109840046A (en) * 2017-11-29 2019-06-04 鸿合科技股份有限公司 Touch screen writes processing method and processing device
CN113093938B (en) * 2018-01-31 2022-07-12 海信视像科技股份有限公司 Method and device for realizing writing and erasing simultaneously and terminal equipment
CN108829327B (en) * 2018-05-07 2021-07-09 广州视源电子科技股份有限公司 Writing method and device of interactive intelligent equipment
CN108919983B (en) * 2018-06-15 2021-10-26 广州视源电子科技股份有限公司 Writing starting method and system based on intelligent pen operation
CN111352570B (en) * 2020-03-24 2021-06-01 深圳市鸿合创新信息技术有限责任公司 Writing interaction method, intelligent interaction display equipment and writing interaction system

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10698524B2 (en) * 2016-04-22 2020-06-30 Samsung Electronics Co., Ltd. Touch screen device, input device, and control method thereof and method thereof
US10606382B2 (en) * 2016-09-01 2020-03-31 Wacom Co., Ltd. Stylus and sensor controller for bi-directional communication using stylus identifier
US11086419B2 (en) * 2019-05-22 2021-08-10 Sharp Kabushiki Kaisha Information processing device, information processing method, and recording medium

Also Published As

Publication number Publication date
WO2021189706A1 (en) 2021-09-30
US11861160B2 (en) 2024-01-02
CN111352570B (en) 2021-06-01
EP4123439A1 (en) 2023-01-25
EP4123439A4 (en) 2024-04-24
CN111352570A (en) 2020-06-30

Similar Documents

Publication Publication Date Title
US11861160B2 (en) Writing interaction method, smart interactive display device and writing interaction system
US12093506B2 (en) Systems and methods for a touchscreen user interface for a collaborative editing tool
WO2021068366A1 (en) Writing operation method and device for intelligent interactive whiteboard, apparatus, and storage medium
CN107844953B (en) Conference file management method, device, terminal and computer readable storage medium
CN110045909B (en) Ellipse processing method, device and equipment based on writing track and storage medium
WO2009117810A1 (en) Method and tool for recognizing a hand-drawn table
CN111580903B (en) Real-time voting method, device, terminal equipment and storage medium
CN108509142B (en) Writing software interaction method and device, terminal equipment and storage medium
WO2021068412A1 (en) Interaction method and apparatus for intelligent interaction tablet, terminal device, and storage medium
US10565299B2 (en) Electronic apparatus and display control method
CN105723314B (en) Two-step content selection
US20200201519A1 (en) Information processing apparatus
CN109062491A (en) Handwriting processing method and device for interactive intelligent equipment
CN108492349B (en) Processing method, device and equipment for writing strokes and storage medium
CN114518820A (en) Icon sorting method and device and electronic equipment
KR20040043454A (en) Pen input method and apparatus in pen computing system
CN110069205A (en) Writing track adjusting method and device, terminal equipment and storage medium
WO2021068405A1 (en) Element transfer method, apparatus and device, and storage medium
US10921977B2 (en) Information processing apparatus and information processing method
CN116088744A (en) Application control method and device thereof
US20180337988A1 (en) Methods for working collaboratively and systems using the same
CN109117072A (en) Writing area control method and system, writing method and system and interactive intelligent tablet
CN112256189B (en) Whiteboard note display method, whiteboard note display device, whiteboard note display equipment and storage medium
JP2016048484A (en) Drawing control apparatus, information sharing system, drawing control method, and drawing control program
CN107679219A (en) Matching method and device, interactive intelligent panel and storage medium

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE