CN117077711A - Desktop digital interaction method and device, electronic equipment and storage medium - Google Patents
Desktop digital interaction method and device, electronic equipment and storage medium Download PDFInfo
- Publication number
- CN117077711A CN117077711A CN202311068488.7A CN202311068488A CN117077711A CN 117077711 A CN117077711 A CN 117077711A CN 202311068488 A CN202311068488 A CN 202311068488A CN 117077711 A CN117077711 A CN 117077711A
- Authority
- CN
- China
- Prior art keywords
- feature points
- visual
- image
- coordinates
- positioning
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000003993 interaction Effects 0.000 title claims abstract description 98
- 238000000034 method Methods 0.000 title claims abstract description 35
- 230000000007 visual effect Effects 0.000 claims abstract description 259
- 238000013507 mapping Methods 0.000 claims description 11
- 230000004807 localization Effects 0.000 claims description 8
- 238000004590 computer program Methods 0.000 claims description 5
- 239000000758 substrate Substances 0.000 claims 1
- 238000012545 processing Methods 0.000 description 9
- 238000010586 diagram Methods 0.000 description 8
- 230000006870 function Effects 0.000 description 8
- 238000005516 engineering process Methods 0.000 description 6
- 230000004438 eyesight Effects 0.000 description 6
- 230000002452 interceptive effect Effects 0.000 description 5
- 230000003287 optical effect Effects 0.000 description 4
- 230000000694 effects Effects 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 238000003745 diagnosis Methods 0.000 description 2
- 239000013307 optical fiber Substances 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 208000027418 Wounds and injury Diseases 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 230000006378 damage Effects 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 238000003702 image correction Methods 0.000 description 1
- 208000014674 injury Diseases 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
- G06K7/14—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation using light without selection of wavelength, e.g. sensing reflected white light
- G06K7/1404—Methods for optical code recognition
- G06K7/1408—Methods for optical code recognition the method being specifically adapted for the type of code
- G06K7/1417—2D bar codes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
- G06K7/14—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation using light without selection of wavelength, e.g. sensing reflected white light
- G06K7/1404—Methods for optical code recognition
- G06K7/1439—Methods for optical code recognition including a method step for retrieval of the optical code
- G06K7/1447—Methods for optical code recognition including a method step for retrieval of the optical code extracting optical codes from image or text carrying said optical code
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Electromagnetism (AREA)
- General Health & Medical Sciences (AREA)
- Toxicology (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Image Analysis (AREA)
Abstract
The embodiment of the specification provides a desktop digital interaction method, a desktop digital interaction device, electronic equipment and a storage medium, wherein the method comprises the following steps: identifying visual positioning marks carrying visual positioning feature points which can be identified by a computer and visual selection marks carrying visual selection feature points which can be identified by the computer in an image, and determining coordinates of the visual positioning feature points and the visual selection feature points in the image; and generating interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, and performing digital interaction.
Description
Technical Field
The present document relates to the field of computer technologies, and in particular, to a desktop digital interaction method, a desktop digital interaction device, an electronic device, and a storage medium.
Background
Computer vision recognition technology refers to the use of computer science and artificial intelligence technology to automatically recognize content and features in images and videos through the processing and analysis of digital images or videos.
The computer vision recognition technology is widely applied to the fields of face recognition, character recognition, object recognition, scene recognition, medical diagnosis and the like, and becomes an important component of digital transformation in modern society. Specifically, computer vision recognition technology includes the following major aspects:
object identification: by analyzing and processing the characteristics of the edges, the areas, the colors, the textures and the like of the objects in the image, the objects can be automatically identified and classified.
Face recognition: by utilizing the characteristic information of the face in the image, the automatic identification and recognition of the face is realized by analyzing and processing the shape, color, texture, characteristic points and the like of the face.
Character recognition: and analyzing and processing the characters in the image, and realizing automatic recognition and conversion of the characters into text data which can be processed by a computer.
Scene recognition: and analyzing and processing the scenes and the backgrounds in the images to realize automatic recognition and classification of the scenes.
In addition, the computer vision recognition technology can be applied to the fields of medical image diagnosis, vehicle recognition, object tracking, image searching and the like, and brings great convenience and benefit for life and work of people.
At present, common computer interaction modes are realized by means of a display (comprising a touch screen), blue light of the display can cause injury to eyes and influence the health of teenagers and children, meanwhile, the main learning medium of the current teenagers and children is a written document, and at present, digital interaction operation similar to the pixel level of an electronic document cannot be realized on the written document. The above situation makes digital education for young children a dilemma.
Disclosure of Invention
The invention aims to provide a desktop digital interaction method, a desktop digital interaction device, electronic equipment and a storage medium, and aims to solve the problems in the prior art.
The invention provides a desktop digital interaction method, which comprises the following steps:
identifying visual positioning marks carrying visual positioning feature points which can be identified by a computer and visual selection marks carrying visual selection feature points which can be identified by the computer in an image, and determining coordinates of the visual positioning feature points and the visual selection feature points in the image;
and generating interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, and performing digital interaction.
The invention provides a desktop digital interaction device, which comprises:
the identification module is used for identifying visual positioning identifiers carrying visual positioning feature points which can be identified by a computer and visual selection identifiers carrying visual selection feature points which can be identified by the computer in the image, and determining coordinates of the visual positioning feature points and the visual selection feature points in the image;
and the interaction module is used for generating interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image and performing digital interaction.
The embodiment of the invention also provides electronic equipment, which comprises: the computer program comprises a memory, a processor and a computer program stored in the memory and capable of running on the processor, wherein the computer program realizes the steps of the desktop digital interaction method when being executed by the processor.
The embodiment of the invention also provides a computer readable storage medium, wherein the computer readable storage medium stores an information transmission implementation program, and the program realizes the steps of the desktop digital interaction method when being executed by a processor.
According to the embodiment of the invention, the digitalized interaction effect similar to the image interaction interface can be realized on the physical desktop (such as a desk) without using a display, the interaction operation of the visual positioning mark and the visual selection mark is more convenient for a user to carry out the interaction operation on the physical desktop by both hands, meanwhile, the introduction of the visual positioning mark greatly simplifies the realization and operation of functions such as image correction, integrated arrangement of an operation instruction area, screenshot and the like, in addition, after the screenshot is carried out by the cooperation of the visual positioning mark and the visual selection mark, the screenshot is digitalized by the cooperation of the operation instruction area or other operation instructions, the digitalized processing of the screenshot can be conveniently realized, and especially, the digitalized interaction operation of pixel-level sight of the written document can be conveniently carried out, so that the user is helped to fully utilize digitalized resources to improve the learning efficiency while the eyesight of children is protected.
Drawings
For a clearer description of one or more embodiments of the present description or of the solutions of the prior art, the drawings that are necessary for the description of the embodiments or of the prior art will be briefly described, it being apparent that the drawings in the description that follow are only some of the embodiments described in the description, from which, for a person skilled in the art, other drawings can be obtained without inventive faculty.
FIG. 1 is a flow chart of a desktop digitizing interaction method of an embodiment of the invention;
FIG. 2 is a detailed flow chart of an image-based desktop digitizing interaction method of an embodiment of the invention;
FIG. 3 is a schematic illustration of a visual location indicator and a visual selection indicator according to an embodiment of the invention;
FIG. 4 is a schematic diagram of a visual locating feature point or visual selecting feature point according to an embodiment of the present invention;
FIG. 5 is a schematic illustration of a menu area and a cut-away area disposed to the right of a visual location indicator in accordance with an embodiment of the present invention;
FIG. 6 is a schematic diagram of a partial screenshot within a rectangular box in an image in accordance with an embodiment of the invention;
FIG. 7 is a schematic diagram of a screenshot effect of an embodiment of the invention;
FIG. 8 is a schematic illustration of forming a functional menu area around a visual location indicator in accordance with an embodiment of the present invention;
FIG. 9 is a schematic diagram of a plurality of images taken in a screenshot of an embodiment of the invention;
FIG. 10 is a schematic diagram of specific steps of a screenshot of an embodiment of the invention;
FIG. 11 is a schematic diagram of specific steps for text recognition of a screenshot in accordance with an embodiment of the invention;
FIG. 12 is a schematic view of a desktop digitizing interaction apparatus according to an embodiment of the invention;
fig. 13 is a schematic diagram of an electronic device according to an embodiment of the invention.
Detailed Description
In order to enable a person skilled in the art to better understand the technical solutions in one or more embodiments of the present specification, the technical solutions in one or more embodiments of the present specification will be clearly and completely described below with reference to the drawings in one or more embodiments of the present specification, and it is obvious that the described embodiments are only some embodiments of the present specification, not all embodiments. All other embodiments, which can be made by one or more embodiments of the present disclosure without inventive faculty, are intended to be within the scope of the present disclosure.
Method embodiment
According to an embodiment of the present invention, a desktop digital interaction method is provided, and fig. 1 is a flowchart of the desktop digital interaction method according to an embodiment of the present invention, as shown in fig. 1, where the desktop digital interaction method according to an embodiment of the present invention specifically includes:
step S101, identifying visual positioning marks carrying visual positioning feature points which can be identified by a computer and visual selection marks carrying visual selection feature points which can be identified by the computer in an image, and determining coordinates of the visual positioning feature points and the visual selection feature points in the image; wherein, visual positioning feature point is one or more, visual selection feature point is one or more.
Step S102, generating interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, and performing digital interaction.
In step S102, generating the interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image specifically includes:
establishing a two-dimensional or three-dimensional positioning coordinate system according to the visual positioning characteristic points;
determining a mapping relation between an image coordinate system and a positioning coordinate system according to parameters of the visual positioning feature points in the positioning coordinate system and coordinates of the visual positioning feature points in the image;
determining the coordinates of the visual selection feature points in the positioning coordinate system according to the coordinates of the visual selection feature points in the image and the mapping relation between the image coordinate system and the positioning coordinate system;
and generating the interaction information according to the coordinates of the visual positioning feature points or the visual selection feature points in the positioning coordinate system, and performing digital interaction.
In step S101, identifying a visual positioning identifier of a visual positioning feature point which is carried in a first image and can be identified by a computer, and determining coordinates of the visual positioning feature point in the first image;
and identifying a visual selection identifier carrying visual selection feature points which can be identified by a computer in the second image, and determining coordinates of the visual selection feature points in the second image.
In step S102, a mapping relationship between the coordinate systems of the first image and the second image is established;
and generating interaction information according to the coordinates of the visual positioning feature points in the first image and the coordinates of the visual selection feature points in the second image based on the mapping relation.
In step S102, generating the interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image specifically includes:
correcting the image or the local part of the image according to the relative position relation among a plurality of the visual positioning feature points or the relative position relation between the visual positioning feature points and the visual selection feature points.
In step S102, generating the interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image specifically includes:
and intercepting local screenshot formed by the image according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, and generating the interaction information according to the screenshot.
In the implementation of the present invention, the method further includes:
setting one or more operation instruction areas in the image according to the relative positions of the operation instruction areas and the visual positioning feature points, and setting each operation instruction area to correspond to one or more instructions, wherein the positions of the operation instruction areas and the visual positioning marks are relatively fixed.
In step S102, generating the interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image specifically includes:
determining an operation instruction area in the image according to the coordinates of the visual positioning feature points;
and determining whether the visual selection feature point is in a certain operation instruction area according to the coordinates of the visual selection feature point, and generating the interaction information according to an operation instruction generation rule corresponding to the operation instruction area if the coordinates of the visual selection feature point are in the certain operation instruction area.
In step S102, generating the interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image specifically includes:
and generating the interaction information according to the coordinates of the visual positioning feature points and the visual selection feature points in the image and the stored historical interaction information or the historical coordinates of the visual positioning feature points and the visual selection feature points in the image.
The generating the interaction information specifically includes:
based on prestored, remote calling or obtaining interaction information generation rules from the visual positioning identification, generating the interaction information according to the coordinates of the visual positioning feature points or the visual selection feature points based on the interaction information generation rules.
The above technical solutions of the embodiments of the present invention are illustrated below with reference to the accompanying drawings.
FIG. 2 is a detailed flowchart of an image-based desktop digitizing interaction method according to an embodiment of the invention, as shown in FIG. 2, comprising the following steps:
identifying whether a visual positioning identifier exists in an image, wherein the visual positioning identifier carries one or more visual positioning feature points which can be identified by a computer, and if the visual positioning identifier exists, determining coordinates of the visual positioning feature points in the image;
identifying whether a visual selection identifier exists in an image, wherein the visual selection identifier carries one or more visual selection feature points which can be identified by a computer, and if the visual selection identifier exists, determining coordinates of the visual selection feature points in the image; (identifying visual positioning identification and visual selection identification, wherein the visual positioning identification can be determined first or the visual selection identification can be determined first without determining the relationship
And generating interaction information according to the coordinates of the visual positioning feature points or the visual selection feature points in the image.
Establishing a positioning coordinate system according to two or more visual positioning feature points, wherein the positioning coordinate system can be two-dimensional or three-dimensional, and generating interaction information according to the coordinates of the visual positioning feature points and the visual selection feature points in the image specifically comprises:
determining the mapping relation between an image coordinate system and a positioning coordinate system according to the parameters of the visual positioning feature points in the positioning coordinate system and the coordinates of the visual positioning feature points in the image;
according to the coordinates of the visual selection feature points in the image and the mapping relation between the image coordinate system and the positioning coordinate system, determining the coordinates of the visual selection feature points in the positioning coordinate system;
and generating the interaction information according to the coordinates of the visual positioning feature points or the visual selection feature points in the positioning coordinate system.
The interactive information is generated according to the coordinates of the visual positioning feature points or the visual selection feature points based on the interactive information generation rule which is stored in advance, called remotely or obtained from the visual positioning identification.
Furthermore, in one embodiment of the present invention, the image includes a first image and a second image, which may be the same or different, and the coordinate systems of the first image and the second image may establish a mapping relationship, and the visual positioning identifier and the visual selection identifier may be the same or different, and specifically include;
identifying whether a visual positioning identifier exists in the first image, wherein the visual positioning identifier carries one or more visual positioning feature points which can be identified by a computer, and if the visual positioning identifier exists, determining coordinates of the visual positioning feature points in the first image;
identifying whether a visual selection identifier exists in the second image, wherein the visual selection identifier carries one or more visual selection feature points which can be identified by a computer, and if the visual selection identifier exists, determining coordinates of the visual selection feature points in the second image;
and generating interaction information according to the coordinates of the visual positioning feature points in the first image and the coordinates of the visual selection feature points in the second image.
In one embodiment of the present invention, generating the interaction information according to the coordinates of the visual positioning feature point and the visual selection feature point in the image specifically includes:
and intercepting a local screenshot formed by the image according to the coordinates of the visual positioning feature points or the visual selection feature points in the image, and generating the interaction information according to the screenshot.
In one embodiment of the present invention, the capturing the partial screenshot of the image specifically includes:
correcting the local part of the intercepted image according to the relative position relation between the visual positioning characteristic points and/or the visual selection characteristic points.
In one embodiment of the present invention, one or more operation instruction areas may be determined according to a relative position to the visual positioning feature point, where each operation instruction area corresponds to one or more instructions, the operation instruction area is relatively fixed to the visual positioning identifier in reality, and generating the interaction information according to coordinates of the visual positioning feature point and the visual selection feature point in the image specifically includes:
determining an operation instruction area in the image according to the coordinates of the visual positioning feature points;
and determining whether the visual selection feature point is in a certain operation instruction area according to the coordinates of the visual selection feature point, and generating the interaction information according to an operation instruction generation rule corresponding to the operation instruction area if the coordinates of the visual selection feature point are in the certain operation instruction area.
In one embodiment of the present invention, the previously generated interaction information may be stored, and generating the interaction information according to the coordinates of the visual positioning feature point and the visual selection feature point in the image specifically includes: and generating the interaction information according to the coordinates of the visual positioning feature points and the visual selection feature points in the image and the stored historical interaction information or the historical coordinates of the visual positioning feature points and the visual selection feature points in the image.
The above technical solutions of the embodiments of the present invention are exemplified below. The embodiment of the invention realizes the pixel level interactive operation of the written text when the user sees the written text by utilizing the cooperation of the visual positioning mark and the visual selection mark. In the embodiment of fig. 3, the large arco code on the left is used as the visual positioning identifier, and the small arco code on the right, labeled '2', is used as the visual selection identifier. In the embodiment of fig. 4, the large arco code on the left side is used as the visual positioning identifier, the four corners of the large arco code can be identified and positioned by a computer to form visual positioning feature points, the visual positioning feature points can be accurately identified by human eyes, and meanwhile, the small arco code on the right side marked with '2' is used as the visual selection identifier, the upper left corner of the small arco code can be identified and positioned by the computer to form visual selection feature points, and the visual selection feature points can be accurately identified by human eyes.
In another embodiment, a point calculated from an existing visual positioning feature point or visual selection feature point may be used as the final visual positioning feature point or visual selection feature point.
In the embodiment of the invention, different interactive operations can be realized according to the relative positions of the visual positioning mark and the visual selection mark. As shown in fig. 5, in one embodiment, a menu area is set on the right side of the visual positioning identifier, each cell (operation instruction area) corresponds to a function, and in which cell the visual selection identifier is, which function is executed by sending the interaction information; a screenshot area is arranged on the lower side of the visual positioning mark, and screenshot can be performed when the visual selection mark is positioned in the screenshot area.
As shown in fig. 6, with the visual localization markers and the visual selection markers, a local screenshot within a rectangular box in an image may be taken. In one embodiment, prior to the screenshot, the image is corrected based on parameters identified in reality by visual localization and coordinates in the image. The screenshot effect is shown in fig. 7. On the premise of no need of a display, the embodiment of the invention realizes the operation of the screen capturing at the pixel level which is obtained by the view. The specific steps of the screenshot are shown in fig. 10.
As shown in fig. 8, in one embodiment, 16 cells are set on the right side of the visual positioning identifier, each cell is an operation instruction area, each operation instruction area corresponds to an operation function, a function menu area is formed around the visual positioning identifier, and the operation function can be an operation for screenshot or other operations for the intelligent device. When the visual selection indicator appears in the operation instruction area, i.e., is considered to click on this area, interactive information for executing the corresponding operation function is generated.
The screenshot can be further processed by image processing, character recognition or other processing by utilizing the result of character recognition by utilizing the operation instruction area of the menu area. In one embodiment, the screenshot may be OCR recognized and the recognized text may be word processed. The specific processing steps are shown in fig. 11.
In one embodiment, as shown in fig. 9, the visual positioning identifier and the visual selection identifier are the same object, the coordinates of the visual positioning identifier in the first image are acquired and mapped to the corresponding position in the second image, then the coordinates of the visual selection identifier are summarized by the second image, and the screenshot area is determined to perform screenshot. A function similar to a mouse circling on a screen can be implemented.
Through the combination of the visual positioning identification and the visual selection identification, the common written text can be operated on the physical desktop by matching with other operation input modes such as an operation instruction area or keys, the pixel-level interaction operation similar to the operation of the electronic document which is obtained when the user sees the electronic document is realized, the operations such as circle selection, screenshot, copying, storage, editing and the like can be accurately performed on the graphic information on the written text without a screen, and the terminal equipment can be operated, so that the eyesight of children is protected.
Device embodiment 1
According to an embodiment of the present invention, a desktop digital interaction device is provided, and fig. 12 is a schematic diagram of the desktop digital interaction device according to the embodiment of the present invention, as shown in fig. 12, where the desktop digital interaction device according to the embodiment of the present invention specifically includes:
the identifying module 120 is configured to identify a visual positioning identifier carrying a visual positioning feature point that can be identified by a computer and a visual selection identifier carrying a visual selection feature point that can be identified by a computer in an image, and determine coordinates of the visual positioning feature point and the visual selection feature point in the image;
and the interaction module 122 is configured to generate interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, so as to perform digital interaction.
The embodiment of the present invention is an embodiment of a device corresponding to the embodiment of the method, and specific operations of each module may be understood by referring to descriptions of the embodiment of the method, which are not repeated herein.
Device example two
An embodiment of the present invention provides an electronic device, as shown in fig. 13, including: a memory 130, a processor 132, and a computer program stored on the memory 130 and executable on the processor 132, which when executed by the processor 132, performs the steps as described in the method embodiments.
Device example III
Embodiments of the present invention provide a computer-readable storage medium having stored thereon a program for carrying out information transmission, which when executed by the processor 132, carries out the steps described in the method embodiments.
The computer readable storage medium of the present embodiment includes, but is not limited to: ROM, RAM, magnetic or optical disks, etc.
From the above description of embodiments, those skilled in the art will readily appreciate that the example embodiments described herein may be implemented in software, or may be implemented in software in combination with the necessary hardware. The technical solution according to the embodiment of the present invention may be embodied in the form of a software product, which may be stored in a non-volatile storage medium (may be a CD-ROM, a usb disk, a mobile hard disk, etc.) or on a network, and includes several instructions to cause a computing device (may be a personal computer, or a network device, etc.) to perform the above-described method according to the embodiment of the present invention.
The software product may employ any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. The readable storage medium can be, for example, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or a combination of any of the foregoing. More specific examples (a non-exhaustive list) of the readable storage medium would include the following: an electrical connection having one or more wires, a portable disk, a hard disk, random Access Memory (RAM), read-only memory (ROM), erasable programmable read-only memory (EPROM or flash memory), optical fiber, portable compact disk read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
The computer readable storage medium may include a data signal propagated in baseband or as part of a carrier wave, with readable program code embodied therein. Such a propagated data signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination of the foregoing. A readable storage medium may also be any readable medium that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a readable storage medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Program code for carrying out operations of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, C++ or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computing device, partly on the user's device, as a stand-alone software package, partly on the user's computing device, partly on a remote computing device, or entirely on the remote computing device. In the case of remote computing devices, the remote computing device may be connected to the user computing device through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external computing device (e.g., connected via the Internet using an Internet service provider).
Those skilled in the art will appreciate that the modules may be distributed throughout several devices as described in the embodiments, and that corresponding variations may be implemented in one or more devices that are unique to the embodiments. The modules of the above embodiments may be combined into one module, or may be further split into a plurality of sub-modules.
Finally, it should be noted that: the above embodiments are only for illustrating the technical solution of the present invention, and not for limiting the same; although the invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical scheme described in the foregoing embodiments can be modified or some or all of the technical features thereof can be replaced by equivalents; such modifications and substitutions do not depart from the spirit of the invention.
Claims (13)
1. A desktop digitizing interaction method, comprising:
identifying visual positioning marks carrying visual positioning feature points which can be identified by a computer and visual selection marks carrying visual selection feature points which can be identified by the computer in an image, and determining coordinates of the visual positioning feature points and the visual selection feature points in the image;
and generating interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, and performing digital interaction.
2. The method of claim 1, wherein the visual localization feature points are one or more and the visual selection feature points are one or more.
3. The method according to claim 1, wherein generating interaction information from coordinates of the visual localization feature points and/or the visual selection feature points in the image comprises in particular:
establishing a two-dimensional or three-dimensional positioning coordinate system according to the visual positioning characteristic points;
determining a mapping relation between an image coordinate system and a positioning coordinate system according to parameters of the visual positioning feature points in the positioning coordinate system and coordinates of the visual positioning feature points in the image;
determining the coordinates of the visual selection feature points in the positioning coordinate system according to the coordinates of the visual selection feature points in the image and the mapping relation between the image coordinate system and the positioning coordinate system;
and generating the interaction information according to the coordinates of the visual positioning feature points or the visual selection feature points in the positioning coordinate system, and performing digital interaction.
4. The method of claim 1, wherein the step of determining the position of the substrate comprises,
identifying a visual positioning identifier carrying visual positioning feature points which can be identified by a computer and a visual selection identifier carrying visual selection feature points which can be identified by the computer in an image, wherein determining coordinates of the visual positioning feature points and the visual selection feature points in the image specifically comprises:
identifying a visual positioning identifier carrying a visual positioning characteristic point which can be identified by a computer in a first image, and determining the coordinates of the visual positioning characteristic point in the first image;
identifying a visual selection identifier carrying visual selection feature points which can be identified by a computer in a second image, and determining coordinates of the visual selection feature points in the second image;
generating interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, wherein the digital interaction specifically comprises the following steps:
establishing a mapping relation between the coordinate systems of the first image and the second image;
and generating interaction information according to the coordinates of the visual positioning feature points in the first image and the coordinates of the visual selection feature points in the second image based on the mapping relation.
5. The method according to claim 1, wherein generating interaction information from coordinates of the visual localization feature points and/or the visual selection feature points in the image comprises in particular:
correcting the image or the local part of the image according to the relative position relation among a plurality of the visual positioning feature points or the relative position relation between the visual positioning feature points and the visual selection feature points.
6. The method according to claim 1 or 5, wherein generating interaction information from coordinates of the visual localization feature points and/or the visual selection feature points in the image comprises in particular:
and intercepting local screenshot formed by the image according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image, and generating the interaction information according to the screenshot.
7. The method according to claim 1, wherein the method further comprises:
setting one or more operation instruction areas in the image according to the relative positions of the operation instruction areas and the visual positioning feature points, and setting each operation instruction area to correspond to one or more instructions, wherein the positions of the operation instruction areas and the visual positioning marks are relatively fixed.
8. The method according to claim 7, wherein generating interaction information from coordinates of the visual localization feature points and/or the visual selection feature points in the image comprises in particular:
determining an operation instruction area in the image according to the coordinates of the visual positioning feature points;
and determining whether the visual selection feature point is in a certain operation instruction area according to the coordinates of the visual selection feature point, and generating the interaction information according to an operation instruction generation rule corresponding to the operation instruction area if the coordinates of the visual selection feature point are in the certain operation instruction area.
9. The method according to claim 1, wherein generating interaction information from coordinates of the visual localization feature points and/or the visual selection feature points in the image comprises in particular:
and generating the interaction information according to the coordinates of the visual positioning feature points and the visual selection feature points in the image and the stored historical interaction information or the historical coordinates of the visual positioning feature points and the visual selection feature points in the image.
10. The method according to any one of claims 1-9, wherein generating interaction information comprises:
based on prestored, remote calling or obtaining interaction information generation rules from the visual positioning identification, generating the interaction information according to the coordinates of the visual positioning feature points or the visual selection feature points based on the interaction information generation rules.
11. A desktop digitizing interaction apparatus, comprising:
the identification module is used for identifying visual positioning identifiers carrying visual positioning feature points which can be identified by a computer and visual selection identifiers carrying visual selection feature points which can be identified by the computer in the image, and determining coordinates of the visual positioning feature points and the visual selection feature points in the image;
and the interaction module is used for generating interaction information according to the coordinates of the visual positioning feature points and/or the visual selection feature points in the image and performing digital interaction.
12. An electronic device, comprising: memory, a processor and a computer program stored on the memory and executable on the processor, which when executed by the processor, performs the steps of the desktop digitizing interaction method of any of claims 1 to 10.
13. A computer-readable storage medium, characterized in that it has stored thereon a program for realizing information transfer, which when executed by a processor realizes the steps of the desktop digitizing interaction method according to any of claims 1 to 10.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311068488.7A CN117077711A (en) | 2023-08-23 | 2023-08-23 | Desktop digital interaction method and device, electronic equipment and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311068488.7A CN117077711A (en) | 2023-08-23 | 2023-08-23 | Desktop digital interaction method and device, electronic equipment and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN117077711A true CN117077711A (en) | 2023-11-17 |
Family
ID=88711230
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202311068488.7A Pending CN117077711A (en) | 2023-08-23 | 2023-08-23 | Desktop digital interaction method and device, electronic equipment and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117077711A (en) |
-
2023
- 2023-08-23 CN CN202311068488.7A patent/CN117077711A/en active Pending
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108446698B (en) | Method, device, medium and electronic equipment for detecting text in image | |
CN108564082B (en) | Image processing method, device, server and medium | |
WO2021082885A1 (en) | Method and device for generating training sample for semantic segmentation model, storage medium, and electronic device | |
CN109766879A (en) | Generation, character detection method, device, equipment and the medium of character machining model | |
CN109753968A (en) | Generation method, device, equipment and the medium of character recognition model | |
CN113704531A (en) | Image processing method, image processing device, electronic equipment and computer readable storage medium | |
CN110490959B (en) | Three-dimensional image processing method and device, virtual image generating method and electronic equipment | |
CN104537339A (en) | Information identification method and information identification system | |
US20230087489A1 (en) | Image processing method and apparatus, device, and storage medium | |
US11847812B2 (en) | Image generation method and apparatus, device, and storage medium | |
CN112949437A (en) | Gesture recognition method, gesture recognition device and intelligent equipment | |
CN108665769B (en) | Network teaching method and device based on convolutional neural network | |
CN116012913A (en) | Model training method, face key point detection method, medium and device | |
CN115393872A (en) | Method, device and equipment for training text classification model and storage medium | |
CN113837194B (en) | Image processing method, image processing apparatus, electronic device, and storage medium | |
CN115205883A (en) | Data auditing method, device, equipment and storage medium based on OCR (optical character recognition) and NLP (non-line language) | |
US10766143B1 (en) | Voice controlled keyboard typing using computer vision | |
CN112348025B (en) | Character detection method and device, electronic equipment and storage medium | |
CN113867875A (en) | Method, device, equipment and storage medium for editing and displaying marked object | |
CN113269781A (en) | Data generation method and device and electronic equipment | |
CN111107264A (en) | Image processing method, image processing device, storage medium and terminal | |
CN117077711A (en) | Desktop digital interaction method and device, electronic equipment and storage medium | |
CN115510508A (en) | Page information protection method and device and electronic equipment | |
CN113096170B (en) | Text image registration method, device, equipment and storage medium | |
CN113538537B (en) | Image registration and model training method, device, equipment, server and medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |