CN117077711A - Desktop digital interaction method and device, electronic equipment and storage medium - Google Patents

Desktop digital interaction method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN117077711A
CN117077711A CN202311068488.7A CN202311068488A CN117077711A CN 117077711 A CN117077711 A CN 117077711A CN 202311068488 A CN202311068488 A CN 202311068488A CN 117077711 A CN117077711 A CN 117077711A
Authority
CN
China
Prior art keywords
feature points
visual
image
coordinates
positioning
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311068488.7A
Other languages
Chinese (zh)
Inventor
杨宏伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to CN202311068488.7A priority Critical patent/CN117077711A/en
Publication of CN117077711A publication Critical patent/CN117077711A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K7/00Methods or arrangements for sensing record carriers, e.g. for reading patterns
    • G06K7/10Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
    • G06K7/14Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation using light without selection of wavelength, e.g. sensing reflected white light
    • G06K7/1404Methods for optical code recognition
    • G06K7/1408Methods for optical code recognition the method being specifically adapted for the type of code
    • G06K7/14172D bar codes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K7/00Methods or arrangements for sensing record carriers, e.g. for reading patterns
    • G06K7/10Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
    • G06K7/14Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation using light without selection of wavelength, e.g. sensing reflected white light
    • G06K7/1404Methods for optical code recognition
    • G06K7/1439Methods for optical code recognition including a method step for retrieval of the optical code
    • G06K7/1447Methods for optical code recognition including a method step for retrieval of the optical code extracting optical codes from image or text carrying said optical code

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Electromagnetism (AREA)
  • General Health & Medical Sciences (AREA)
  • Toxicology (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The embodiment of the specification provides a desktop digital interaction method, a desktop digital interaction device, electronic equipment and a storage medium, wherein the method comprises the following steps: identifying visual positioning marks carrying visual positioning feature points which can be identified by a computer and visual selection marks carrying visual selection feature points which can be identified by the computer in an image, and determining coordinates of the visual positioning feature points and the visual selection feature points in the image; and generating interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, and performing digital interaction.

Description

Desktop digital interaction method and device, electronic equipment and storage medium
Technical Field
The present document relates to the field of computer technologies, and in particular, to a desktop digital interaction method, a desktop digital interaction device, an electronic device, and a storage medium.
Background
Computer vision recognition technology refers to the use of computer science and artificial intelligence technology to automatically recognize content and features in images and videos through the processing and analysis of digital images or videos.
The computer vision recognition technology is widely applied to the fields of face recognition, character recognition, object recognition, scene recognition, medical diagnosis and the like, and becomes an important component of digital transformation in modern society. Specifically, computer vision recognition technology includes the following major aspects:
object identification: by analyzing and processing the characteristics of the edges, the areas, the colors, the textures and the like of the objects in the image, the objects can be automatically identified and classified.
Face recognition: by utilizing the characteristic information of the face in the image, the automatic identification and recognition of the face is realized by analyzing and processing the shape, color, texture, characteristic points and the like of the face.
Character recognition: and analyzing and processing the characters in the image, and realizing automatic recognition and conversion of the characters into text data which can be processed by a computer.
Scene recognition: and analyzing and processing the scenes and the backgrounds in the images to realize automatic recognition and classification of the scenes.
In addition, the computer vision recognition technology can be applied to the fields of medical image diagnosis, vehicle recognition, object tracking, image searching and the like, and brings great convenience and benefit for life and work of people.
At present, common computer interaction modes are realized by means of a display (comprising a touch screen), blue light of the display can cause injury to eyes and influence the health of teenagers and children, meanwhile, the main learning medium of the current teenagers and children is a written document, and at present, digital interaction operation similar to the pixel level of an electronic document cannot be realized on the written document. The above situation makes digital education for young children a dilemma.
Disclosure of Invention
The invention aims to provide a desktop digital interaction method, a desktop digital interaction device, electronic equipment and a storage medium, and aims to solve the problems in the prior art.
The invention provides a desktop digital interaction method, which comprises the following steps:
identifying visual positioning marks carrying visual positioning feature points which can be identified by a computer and visual selection marks carrying visual selection feature points which can be identified by the computer in an image, and determining coordinates of the visual positioning feature points and the visual selection feature points in the image;
and generating interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, and performing digital interaction.
The invention provides a desktop digital interaction device, which comprises:
the identification module is used for identifying visual positioning identifiers carrying visual positioning feature points which can be identified by a computer and visual selection identifiers carrying visual selection feature points which can be identified by the computer in the image, and determining coordinates of the visual positioning feature points and the visual selection feature points in the image;
and the interaction module is used for generating interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image and performing digital interaction.
The embodiment of the invention also provides electronic equipment, which comprises: the computer program comprises a memory, a processor and a computer program stored in the memory and capable of running on the processor, wherein the computer program realizes the steps of the desktop digital interaction method when being executed by the processor.
The embodiment of the invention also provides a computer readable storage medium, wherein the computer readable storage medium stores an information transmission implementation program, and the program realizes the steps of the desktop digital interaction method when being executed by a processor.
According to the embodiment of the invention, the digitalized interaction effect similar to the image interaction interface can be realized on the physical desktop (such as a desk) without using a display, the interaction operation of the visual positioning mark and the visual selection mark is more convenient for a user to carry out the interaction operation on the physical desktop by both hands, meanwhile, the introduction of the visual positioning mark greatly simplifies the realization and operation of functions such as image correction, integrated arrangement of an operation instruction area, screenshot and the like, in addition, after the screenshot is carried out by the cooperation of the visual positioning mark and the visual selection mark, the screenshot is digitalized by the cooperation of the operation instruction area or other operation instructions, the digitalized processing of the screenshot can be conveniently realized, and especially, the digitalized interaction operation of pixel-level sight of the written document can be conveniently carried out, so that the user is helped to fully utilize digitalized resources to improve the learning efficiency while the eyesight of children is protected.
Drawings
For a clearer description of one or more embodiments of the present description or of the solutions of the prior art, the drawings that are necessary for the description of the embodiments or of the prior art will be briefly described, it being apparent that the drawings in the description that follow are only some of the embodiments described in the description, from which, for a person skilled in the art, other drawings can be obtained without inventive faculty.
FIG. 1 is a flow chart of a desktop digitizing interaction method of an embodiment of the invention;
FIG. 2 is a detailed flow chart of an image-based desktop digitizing interaction method of an embodiment of the invention;
FIG. 3 is a schematic illustration of a visual location indicator and a visual selection indicator according to an embodiment of the invention;
FIG. 4 is a schematic diagram of a visual locating feature point or visual selecting feature point according to an embodiment of the present invention;
FIG. 5 is a schematic illustration of a menu area and a cut-away area disposed to the right of a visual location indicator in accordance with an embodiment of the present invention;
FIG. 6 is a schematic diagram of a partial screenshot within a rectangular box in an image in accordance with an embodiment of the invention;
FIG. 7 is a schematic diagram of a screenshot effect of an embodiment of the invention;
FIG. 8 is a schematic illustration of forming a functional menu area around a visual location indicator in accordance with an embodiment of the present invention;
FIG. 9 is a schematic diagram of a plurality of images taken in a screenshot of an embodiment of the invention;
FIG. 10 is a schematic diagram of specific steps of a screenshot of an embodiment of the invention;
FIG. 11 is a schematic diagram of specific steps for text recognition of a screenshot in accordance with an embodiment of the invention;
FIG. 12 is a schematic view of a desktop digitizing interaction apparatus according to an embodiment of the invention;
fig. 13 is a schematic diagram of an electronic device according to an embodiment of the invention.
Detailed Description
In order to enable a person skilled in the art to better understand the technical solutions in one or more embodiments of the present specification, the technical solutions in one or more embodiments of the present specification will be clearly and completely described below with reference to the drawings in one or more embodiments of the present specification, and it is obvious that the described embodiments are only some embodiments of the present specification, not all embodiments. All other embodiments, which can be made by one or more embodiments of the present disclosure without inventive faculty, are intended to be within the scope of the present disclosure.
Method embodiment
According to an embodiment of the present invention, a desktop digital interaction method is provided, and fig. 1 is a flowchart of the desktop digital interaction method according to an embodiment of the present invention, as shown in fig. 1, where the desktop digital interaction method according to an embodiment of the present invention specifically includes:
step S101, identifying visual positioning marks carrying visual positioning feature points which can be identified by a computer and visual selection marks carrying visual selection feature points which can be identified by the computer in an image, and determining coordinates of the visual positioning feature points and the visual selection feature points in the image; wherein, visual positioning feature point is one or more, visual selection feature point is one or more.
Step S102, generating interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, and performing digital interaction.
In step S102, generating the interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image specifically includes:
establishing a two-dimensional or three-dimensional positioning coordinate system according to the visual positioning characteristic points;
determining a mapping relation between an image coordinate system and a positioning coordinate system according to parameters of the visual positioning feature points in the positioning coordinate system and coordinates of the visual positioning feature points in the image;
determining the coordinates of the visual selection feature points in the positioning coordinate system according to the coordinates of the visual selection feature points in the image and the mapping relation between the image coordinate system and the positioning coordinate system;
and generating the interaction information according to the coordinates of the visual positioning feature points or the visual selection feature points in the positioning coordinate system, and performing digital interaction.
In step S101, identifying a visual positioning identifier of a visual positioning feature point which is carried in a first image and can be identified by a computer, and determining coordinates of the visual positioning feature point in the first image;
and identifying a visual selection identifier carrying visual selection feature points which can be identified by a computer in the second image, and determining coordinates of the visual selection feature points in the second image.
In step S102, a mapping relationship between the coordinate systems of the first image and the second image is established;
and generating interaction information according to the coordinates of the visual positioning feature points in the first image and the coordinates of the visual selection feature points in the second image based on the mapping relation.
In step S102, generating the interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image specifically includes:
correcting the image or the local part of the image according to the relative position relation among a plurality of the visual positioning feature points or the relative position relation between the visual positioning feature points and the visual selection feature points.
In step S102, generating the interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image specifically includes:
and intercepting local screenshot formed by the image according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, and generating the interaction information according to the screenshot.
In the implementation of the present invention, the method further includes:
setting one or more operation instruction areas in the image according to the relative positions of the operation instruction areas and the visual positioning feature points, and setting each operation instruction area to correspond to one or more instructions, wherein the positions of the operation instruction areas and the visual positioning marks are relatively fixed.
In step S102, generating the interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image specifically includes:
determining an operation instruction area in the image according to the coordinates of the visual positioning feature points;
and determining whether the visual selection feature point is in a certain operation instruction area according to the coordinates of the visual selection feature point, and generating the interaction information according to an operation instruction generation rule corresponding to the operation instruction area if the coordinates of the visual selection feature point are in the certain operation instruction area.
In step S102, generating the interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image specifically includes:
and generating the interaction information according to the coordinates of the visual positioning feature points and the visual selection feature points in the image and the stored historical interaction information or the historical coordinates of the visual positioning feature points and the visual selection feature points in the image.
The generating the interaction information specifically includes:
based on prestored, remote calling or obtaining interaction information generation rules from the visual positioning identification, generating the interaction information according to the coordinates of the visual positioning feature points or the visual selection feature points based on the interaction information generation rules.
The above technical solutions of the embodiments of the present invention are illustrated below with reference to the accompanying drawings.
FIG. 2 is a detailed flowchart of an image-based desktop digitizing interaction method according to an embodiment of the invention, as shown in FIG. 2, comprising the following steps:
identifying whether a visual positioning identifier exists in an image, wherein the visual positioning identifier carries one or more visual positioning feature points which can be identified by a computer, and if the visual positioning identifier exists, determining coordinates of the visual positioning feature points in the image;
identifying whether a visual selection identifier exists in an image, wherein the visual selection identifier carries one or more visual selection feature points which can be identified by a computer, and if the visual selection identifier exists, determining coordinates of the visual selection feature points in the image; (identifying visual positioning identification and visual selection identification, wherein the visual positioning identification can be determined first or the visual selection identification can be determined first without determining the relationship
And generating interaction information according to the coordinates of the visual positioning feature points or the visual selection feature points in the image.
Establishing a positioning coordinate system according to two or more visual positioning feature points, wherein the positioning coordinate system can be two-dimensional or three-dimensional, and generating interaction information according to the coordinates of the visual positioning feature points and the visual selection feature points in the image specifically comprises:
determining the mapping relation between an image coordinate system and a positioning coordinate system according to the parameters of the visual positioning feature points in the positioning coordinate system and the coordinates of the visual positioning feature points in the image;
according to the coordinates of the visual selection feature points in the image and the mapping relation between the image coordinate system and the positioning coordinate system, determining the coordinates of the visual selection feature points in the positioning coordinate system;
and generating the interaction information according to the coordinates of the visual positioning feature points or the visual selection feature points in the positioning coordinate system.
The interactive information is generated according to the coordinates of the visual positioning feature points or the visual selection feature points based on the interactive information generation rule which is stored in advance, called remotely or obtained from the visual positioning identification.
Furthermore, in one embodiment of the present invention, the image includes a first image and a second image, which may be the same or different, and the coordinate systems of the first image and the second image may establish a mapping relationship, and the visual positioning identifier and the visual selection identifier may be the same or different, and specifically include;
identifying whether a visual positioning identifier exists in the first image, wherein the visual positioning identifier carries one or more visual positioning feature points which can be identified by a computer, and if the visual positioning identifier exists, determining coordinates of the visual positioning feature points in the first image;
identifying whether a visual selection identifier exists in the second image, wherein the visual selection identifier carries one or more visual selection feature points which can be identified by a computer, and if the visual selection identifier exists, determining coordinates of the visual selection feature points in the second image;
and generating interaction information according to the coordinates of the visual positioning feature points in the first image and the coordinates of the visual selection feature points in the second image.
In one embodiment of the present invention, generating the interaction information according to the coordinates of the visual positioning feature point and the visual selection feature point in the image specifically includes:
and intercepting a local screenshot formed by the image according to the coordinates of the visual positioning feature points or the visual selection feature points in the image, and generating the interaction information according to the screenshot.
In one embodiment of the present invention, the capturing the partial screenshot of the image specifically includes:
correcting the local part of the intercepted image according to the relative position relation between the visual positioning characteristic points and/or the visual selection characteristic points.
In one embodiment of the present invention, one or more operation instruction areas may be determined according to a relative position to the visual positioning feature point, where each operation instruction area corresponds to one or more instructions, the operation instruction area is relatively fixed to the visual positioning identifier in reality, and generating the interaction information according to coordinates of the visual positioning feature point and the visual selection feature point in the image specifically includes:
determining an operation instruction area in the image according to the coordinates of the visual positioning feature points;
and determining whether the visual selection feature point is in a certain operation instruction area according to the coordinates of the visual selection feature point, and generating the interaction information according to an operation instruction generation rule corresponding to the operation instruction area if the coordinates of the visual selection feature point are in the certain operation instruction area.
In one embodiment of the present invention, the previously generated interaction information may be stored, and generating the interaction information according to the coordinates of the visual positioning feature point and the visual selection feature point in the image specifically includes: and generating the interaction information according to the coordinates of the visual positioning feature points and the visual selection feature points in the image and the stored historical interaction information or the historical coordinates of the visual positioning feature points and the visual selection feature points in the image.
The above technical solutions of the embodiments of the present invention are exemplified below. The embodiment of the invention realizes the pixel level interactive operation of the written text when the user sees the written text by utilizing the cooperation of the visual positioning mark and the visual selection mark. In the embodiment of fig. 3, the large arco code on the left is used as the visual positioning identifier, and the small arco code on the right, labeled '2', is used as the visual selection identifier. In the embodiment of fig. 4, the large arco code on the left side is used as the visual positioning identifier, the four corners of the large arco code can be identified and positioned by a computer to form visual positioning feature points, the visual positioning feature points can be accurately identified by human eyes, and meanwhile, the small arco code on the right side marked with '2' is used as the visual selection identifier, the upper left corner of the small arco code can be identified and positioned by the computer to form visual selection feature points, and the visual selection feature points can be accurately identified by human eyes.
In another embodiment, a point calculated from an existing visual positioning feature point or visual selection feature point may be used as the final visual positioning feature point or visual selection feature point.
In the embodiment of the invention, different interactive operations can be realized according to the relative positions of the visual positioning mark and the visual selection mark. As shown in fig. 5, in one embodiment, a menu area is set on the right side of the visual positioning identifier, each cell (operation instruction area) corresponds to a function, and in which cell the visual selection identifier is, which function is executed by sending the interaction information; a screenshot area is arranged on the lower side of the visual positioning mark, and screenshot can be performed when the visual selection mark is positioned in the screenshot area.
As shown in fig. 6, with the visual localization markers and the visual selection markers, a local screenshot within a rectangular box in an image may be taken. In one embodiment, prior to the screenshot, the image is corrected based on parameters identified in reality by visual localization and coordinates in the image. The screenshot effect is shown in fig. 7. On the premise of no need of a display, the embodiment of the invention realizes the operation of the screen capturing at the pixel level which is obtained by the view. The specific steps of the screenshot are shown in fig. 10.
As shown in fig. 8, in one embodiment, 16 cells are set on the right side of the visual positioning identifier, each cell is an operation instruction area, each operation instruction area corresponds to an operation function, a function menu area is formed around the visual positioning identifier, and the operation function can be an operation for screenshot or other operations for the intelligent device. When the visual selection indicator appears in the operation instruction area, i.e., is considered to click on this area, interactive information for executing the corresponding operation function is generated.
The screenshot can be further processed by image processing, character recognition or other processing by utilizing the result of character recognition by utilizing the operation instruction area of the menu area. In one embodiment, the screenshot may be OCR recognized and the recognized text may be word processed. The specific processing steps are shown in fig. 11.
In one embodiment, as shown in fig. 9, the visual positioning identifier and the visual selection identifier are the same object, the coordinates of the visual positioning identifier in the first image are acquired and mapped to the corresponding position in the second image, then the coordinates of the visual selection identifier are summarized by the second image, and the screenshot area is determined to perform screenshot. A function similar to a mouse circling on a screen can be implemented.
Through the combination of the visual positioning identification and the visual selection identification, the common written text can be operated on the physical desktop by matching with other operation input modes such as an operation instruction area or keys, the pixel-level interaction operation similar to the operation of the electronic document which is obtained when the user sees the electronic document is realized, the operations such as circle selection, screenshot, copying, storage, editing and the like can be accurately performed on the graphic information on the written text without a screen, and the terminal equipment can be operated, so that the eyesight of children is protected.
Device embodiment 1
According to an embodiment of the present invention, a desktop digital interaction device is provided, and fig. 12 is a schematic diagram of the desktop digital interaction device according to the embodiment of the present invention, as shown in fig. 12, where the desktop digital interaction device according to the embodiment of the present invention specifically includes:
the identifying module 120 is configured to identify a visual positioning identifier carrying a visual positioning feature point that can be identified by a computer and a visual selection identifier carrying a visual selection feature point that can be identified by a computer in an image, and determine coordinates of the visual positioning feature point and the visual selection feature point in the image;
and the interaction module 122 is configured to generate interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, so as to perform digital interaction.
The embodiment of the present invention is an embodiment of a device corresponding to the embodiment of the method, and specific operations of each module may be understood by referring to descriptions of the embodiment of the method, which are not repeated herein.
Device example two
An embodiment of the present invention provides an electronic device, as shown in fig. 13, including: a memory 130, a processor 132, and a computer program stored on the memory 130 and executable on the processor 132, which when executed by the processor 132, performs the steps as described in the method embodiments.
Device example III
Embodiments of the present invention provide a computer-readable storage medium having stored thereon a program for carrying out information transmission, which when executed by the processor 132, carries out the steps described in the method embodiments.
The computer readable storage medium of the present embodiment includes, but is not limited to: ROM, RAM, magnetic or optical disks, etc.
From the above description of embodiments, those skilled in the art will readily appreciate that the example embodiments described herein may be implemented in software, or may be implemented in software in combination with the necessary hardware. The technical solution according to the embodiment of the present invention may be embodied in the form of a software product, which may be stored in a non-volatile storage medium (may be a CD-ROM, a usb disk, a mobile hard disk, etc.) or on a network, and includes several instructions to cause a computing device (may be a personal computer, or a network device, etc.) to perform the above-described method according to the embodiment of the present invention.
The software product may employ any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. The readable storage medium can be, for example, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or a combination of any of the foregoing. More specific examples (a non-exhaustive list) of the readable storage medium would include the following: an electrical connection having one or more wires, a portable disk, a hard disk, random Access Memory (RAM), read-only memory (ROM), erasable programmable read-only memory (EPROM or flash memory), optical fiber, portable compact disk read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
The computer readable storage medium may include a data signal propagated in baseband or as part of a carrier wave, with readable program code embodied therein. Such a propagated data signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination of the foregoing. A readable storage medium may also be any readable medium that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a readable storage medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Program code for carrying out operations of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, C++ or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computing device, partly on the user's device, as a stand-alone software package, partly on the user's computing device, partly on a remote computing device, or entirely on the remote computing device. In the case of remote computing devices, the remote computing device may be connected to the user computing device through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external computing device (e.g., connected via the Internet using an Internet service provider).
Those skilled in the art will appreciate that the modules may be distributed throughout several devices as described in the embodiments, and that corresponding variations may be implemented in one or more devices that are unique to the embodiments. The modules of the above embodiments may be combined into one module, or may be further split into a plurality of sub-modules.
Finally, it should be noted that: the above embodiments are only for illustrating the technical solution of the present invention, and not for limiting the same; although the invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical scheme described in the foregoing embodiments can be modified or some or all of the technical features thereof can be replaced by equivalents; such modifications and substitutions do not depart from the spirit of the invention.

Claims (13)

1. A desktop digitizing interaction method, comprising:
identifying visual positioning marks carrying visual positioning feature points which can be identified by a computer and visual selection marks carrying visual selection feature points which can be identified by the computer in an image, and determining coordinates of the visual positioning feature points and the visual selection feature points in the image;
and generating interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, and performing digital interaction.
2. The method of claim 1, wherein the visual localization feature points are one or more and the visual selection feature points are one or more.
3. The method according to claim 1, wherein generating interaction information from coordinates of the visual localization feature points and/or the visual selection feature points in the image comprises in particular:
establishing a two-dimensional or three-dimensional positioning coordinate system according to the visual positioning characteristic points;
determining a mapping relation between an image coordinate system and a positioning coordinate system according to parameters of the visual positioning feature points in the positioning coordinate system and coordinates of the visual positioning feature points in the image;
determining the coordinates of the visual selection feature points in the positioning coordinate system according to the coordinates of the visual selection feature points in the image and the mapping relation between the image coordinate system and the positioning coordinate system;
and generating the interaction information according to the coordinates of the visual positioning feature points or the visual selection feature points in the positioning coordinate system, and performing digital interaction.
4. The method of claim 1, wherein the step of determining the position of the substrate comprises,
identifying a visual positioning identifier carrying visual positioning feature points which can be identified by a computer and a visual selection identifier carrying visual selection feature points which can be identified by the computer in an image, wherein determining coordinates of the visual positioning feature points and the visual selection feature points in the image specifically comprises:
identifying a visual positioning identifier carrying a visual positioning characteristic point which can be identified by a computer in a first image, and determining the coordinates of the visual positioning characteristic point in the first image;
identifying a visual selection identifier carrying visual selection feature points which can be identified by a computer in a second image, and determining coordinates of the visual selection feature points in the second image;
generating interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, wherein the digital interaction specifically comprises the following steps:
establishing a mapping relation between the coordinate systems of the first image and the second image;
and generating interaction information according to the coordinates of the visual positioning feature points in the first image and the coordinates of the visual selection feature points in the second image based on the mapping relation.
5. The method according to claim 1, wherein generating interaction information from coordinates of the visual localization feature points and/or the visual selection feature points in the image comprises in particular:
correcting the image or the local part of the image according to the relative position relation among a plurality of the visual positioning feature points or the relative position relation between the visual positioning feature points and the visual selection feature points.
6. The method according to claim 1 or 5, wherein generating interaction information from coordinates of the visual localization feature points and/or the visual selection feature points in the image comprises in particular:
and intercepting local screenshot formed by the image according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, and generating the interaction information according to the screenshot.
7. The method according to claim 1, wherein the method further comprises:
setting one or more operation instruction areas in the image according to the relative positions of the operation instruction areas and the visual positioning feature points, and setting each operation instruction area to correspond to one or more instructions, wherein the positions of the operation instruction areas and the visual positioning marks are relatively fixed.
8. The method according to claim 7, wherein generating interaction information from coordinates of the visual localization feature points and/or the visual selection feature points in the image comprises in particular:
determining an operation instruction area in the image according to the coordinates of the visual positioning feature points;
and determining whether the visual selection feature point is in a certain operation instruction area according to the coordinates of the visual selection feature point, and generating the interaction information according to an operation instruction generation rule corresponding to the operation instruction area if the coordinates of the visual selection feature point are in the certain operation instruction area.
9. The method according to claim 1, wherein generating interaction information from coordinates of the visual localization feature points and/or the visual selection feature points in the image comprises in particular:
and generating the interaction information according to the coordinates of the visual positioning feature points and the visual selection feature points in the image and the stored historical interaction information or the historical coordinates of the visual positioning feature points and the visual selection feature points in the image.
10. The method according to any one of claims 1-9, wherein generating interaction information comprises:
based on prestored, remote calling or obtaining interaction information generation rules from the visual positioning identification, generating the interaction information according to the coordinates of the visual positioning feature points or the visual selection feature points based on the interaction information generation rules.
11. A desktop digitizing interaction apparatus, comprising:
the identification module is used for identifying visual positioning identifiers carrying visual positioning feature points which can be identified by a computer and visual selection identifiers carrying visual selection feature points which can be identified by the computer in the image, and determining coordinates of the visual positioning feature points and the visual selection feature points in the image;
and the interaction module is used for generating interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image and performing digital interaction.
12. An electronic device, comprising: memory, a processor and a computer program stored on the memory and executable on the processor, which when executed by the processor, performs the steps of the desktop digitizing interaction method of any of claims 1 to 10.
13. A computer-readable storage medium, characterized in that it has stored thereon a program for realizing information transfer, which when executed by a processor realizes the steps of the desktop digitizing interaction method according to any of claims 1 to 10.
CN202311068488.7A 2023-08-23 2023-08-23 Desktop digital interaction method and device, electronic equipment and storage medium Pending CN117077711A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311068488.7A CN117077711A (en) 2023-08-23 2023-08-23 Desktop digital interaction method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311068488.7A CN117077711A (en) 2023-08-23 2023-08-23 Desktop digital interaction method and device, electronic equipment and storage medium

Publications (1)

Publication Number Publication Date
CN117077711A true CN117077711A (en) 2023-11-17

Family

ID=88711230

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311068488.7A Pending CN117077711A (en) 2023-08-23 2023-08-23 Desktop digital interaction method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN117077711A (en)

Similar Documents

Publication Publication Date Title
CN108446698B (en) Method, device, medium and electronic equipment for detecting text in image
CN108564082B (en) Image processing method, device, server and medium
WO2021082885A1 (en) Method and device for generating training sample for semantic segmentation model, storage medium, and electronic device
CN109766879A (en) Generation, character detection method, device, equipment and the medium of character machining model
CN109753968A (en) Generation method, device, equipment and the medium of character recognition model
CN113704531A (en) Image processing method, image processing device, electronic equipment and computer readable storage medium
CN110490959B (en) Three-dimensional image processing method and device, virtual image generating method and electronic equipment
CN104537339A (en) Information identification method and information identification system
US20230087489A1 (en) Image processing method and apparatus, device, and storage medium
US11847812B2 (en) Image generation method and apparatus, device, and storage medium
CN112949437A (en) Gesture recognition method, gesture recognition device and intelligent equipment
CN108665769B (en) Network teaching method and device based on convolutional neural network
CN116012913A (en) Model training method, face key point detection method, medium and device
CN115393872A (en) Method, device and equipment for training text classification model and storage medium
CN113837194B (en) Image processing method, image processing apparatus, electronic device, and storage medium
CN115205883A (en) Data auditing method, device, equipment and storage medium based on OCR (optical character recognition) and NLP (non-line language)
US10766143B1 (en) Voice controlled keyboard typing using computer vision
CN112348025B (en) Character detection method and device, electronic equipment and storage medium
CN113867875A (en) Method, device, equipment and storage medium for editing and displaying marked object
CN113269781A (en) Data generation method and device and electronic equipment
CN111107264A (en) Image processing method, image processing device, storage medium and terminal
CN117077711A (en) Desktop digital interaction method and device, electronic equipment and storage medium
CN115510508A (en) Page information protection method and device and electronic equipment
CN113096170B (en) Text image registration method, device, equipment and storage medium
CN113538537B (en) Image registration and model training method, device, equipment, server and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination