WO2019039799A1 - Traitement d'entrée visuelle - Google Patents

Traitement d'entrée visuelle Download PDF

Info

Publication number
WO2019039799A1
WO2019039799A1 PCT/KR2018/009470 KR2018009470W WO2019039799A1 WO 2019039799 A1 WO2019039799 A1 WO 2019039799A1 KR 2018009470 W KR2018009470 W KR 2018009470W WO 2019039799 A1 WO2019039799 A1 WO 2019039799A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
text
electronic device
animation
screen
Prior art date
Application number
PCT/KR2018/009470
Other languages
English (en)
Korean (ko)
Inventor
강유훈
김종택
Original Assignee
네이버 주식회사
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020180035734A external-priority patent/KR102026475B1/ko
Application filed by 네이버 주식회사 filed Critical 네이버 주식회사
Priority to JP2020509490A priority Critical patent/JP7027524B2/ja
Publication of WO2019039799A1 publication Critical patent/WO2019039799A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K19/00Record carriers for use with machines and with at least a part designed to carry digital markings
    • G06K19/06Record carriers for use with machines and with at least a part designed to carry digital markings characterised by the kind of the digital marking, e.g. shape, nature, code
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T1/00General purpose image data processing
    • G06T1/20Processor architectures; Processor configuration, e.g. pipelining
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • G06T13/802D [Two Dimensional] animation, e.g. using sprites
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation

Definitions

  • the following description relates to techniques for processing visual input, and more particularly to techniques for processing visual input that can provide a user interface associated with meaningful and / or meaningful information associated with at least one frame of a video stream that is continuously captured via a camera
  • Korean Patent Laid-Open No. 10-2003-0024786 analyzes an entire image taken by a digital camera in relation to text information, recognizes and interprets the information by an OCR (Optical Character Reader) technique, And / or < / RTI > compressed text codes for transmission.
  • OCR Optical Character Reader
  • An image processing method and system that can provide a user interface associated with meaningful and / or meaningful information associated with at least one frame of a video stream that is continuously captured via a camera, and in combination with a computer,
  • a computer program stored in a computer readable recording medium for causing a computer to execute an image processing method according to the present invention, and a recording medium therefor.
  • An image processing method comprising: driving a camera module in response to entering an image processing mode; Sequentially receiving and sequentially displaying a stream of images through the camera module; Transmitting a stream of the received images to a recognition engine; And displaying the stream of the received images on a screen being displayed sequentially when the recognition result recognized by the recognition engine exists for the stream of the received images. As shown in FIG.
  • An image processing method comprising: receiving an image captured via a camera module of an electronic device through a network; Generating a detaching animation for an object included in the received image; Transmitting the generated detecming animation to the electronic device; Generating an image search result for the received image; And transmitting the generated image search result to the electronic device.
  • a computer program stored in a computer-readable medium for causing a computer to execute the image processing method in combination with a computer.
  • a computer-readable recording medium having recorded thereon a program for causing a computer to execute the image processing method.
  • the computer system comprising: at least one processor configured to execute computer readable instructions, wherein the at least one processor drives a camera module in response to entering an image processing mode, And sequentially transmits the streams of the input images to a recognition engine. When there is a recognition result recognized by the recognition engine for the streams of the received images, And sequentially displaying the streams of the input images on a screen being displayed.
  • a computer-readable medium having computer-executable instructions for performing the steps of: receiving, via the network, an image captured via a camera module of an electronic device by the at least one processor; Generating a detection result for the object, transmitting the generated detec- tion animation to the electronic device, generating an image search result for the received image, and transmitting the generated image search result to the electronic device And performing operations of the computer system.
  • And may provide a user interface associated with meaningful and / or meaningful information associated with at least one frame of the video stream being continuously captured through the camera.
  • FIG. 1 is a diagram illustrating an example of a network environment according to an embodiment of the present invention.
  • FIG. 2 is a block diagram illustrating an internal configuration of an electronic device and a server according to an embodiment of the present invention.
  • FIG. 3 is a diagram showing an example of an execution environment of an image processing system according to an embodiment of the present invention.
  • FIGS. 4 and 5 illustrate examples of providing additional information or additional functions according to OCR recognition in an embodiment of the present invention.
  • 6 to 8 are diagrams illustrating an example of adjusting the font size according to the number of recognized characters in an embodiment of the present invention.
  • FIG. 9 is a diagram illustrating an example of providing an additional function according to bar code recognition in an embodiment of the present invention.
  • FIG. 10 is a diagram illustrating an example of providing an additional function according to QR code recognition in an embodiment of the present invention.
  • 11 to 15 are views showing examples of providing an image search result in an embodiment of the present invention.
  • FIG. 16 is a diagram illustrating examples of providing an image search result in an embodiment of the present invention.
  • FIG. 17 is a diagram illustrating an example of limiting the saturation by extracting a main color in an embodiment of the present invention to be.
  • FIGS. 18 to 20 are views showing examples of providing image search results through a template designed according to a type, in an embodiment of the present invention.
  • FIGS. 21 to 24 illustrate examples of displaying a detecting animation in an embodiment of the present invention.
  • 25 is a diagram showing an example of a detection animation in an embodiment of the present invention.
  • 26 is a diagram showing an example of an image processing method according to an embodiment of the present invention.
  • FIG. 27 is a view showing another example of the image processing method according to the embodiment of the present invention.
  • FIG. 28 is a diagram showing another example of the image processing method according to the embodiment of the present invention.
  • 29 is a flowchart illustrating an example of a method of generating a detaching animation in an embodiment of the present invention.
  • 30 is a diagram showing an example of providing additional information according to place recognition in an embodiment of the present invention.
  • 31 is a diagram showing an example of providing additional information according to recognition of an image code in an embodiment of the present invention.
  • the image processing method according to embodiments of the present invention can be performed through a computer apparatus such as an electronic apparatus and / or a server to be described later.
  • a computer program according to an embodiment of the present invention can be installed and driven in the computer apparatus, and the computer apparatus can perform the image processing method according to an embodiment of the present invention under the control of the computer program that is driven .
  • the above-described computer program may be stored in a computer-readable recording medium for execution by a computer in combination with a computer apparatus to perform an image processing method.
  • 1 is a diagram illustrating an example of a network environment according to an embodiment of the present invention. 1 shows an example in which a plurality of electronic devices 110, 120, 130, 140, a plurality of servers 150, 160, and a network 170 are included. 1, the number of electronic devices and the number of servers are not limited to those shown in FIG.
  • the plurality of electronic devices 110, 120, 130, 140 may be a fixed terminal implemented as a computer device or a mobile terminal.
  • Examples of the plurality of electronic devices 110, 120, 130 and 140 include a smart phone, a mobile phone, a navigation device, a computer, a notebook, a digital broadcast terminal, a PDA (Personal Digital Assistants) ), And tablet PCs.
  • FIG. 1 illustrates the shape of a smartphone as an example of the first electronic device 110, but in the embodiments of the present invention, the first electronic device 110 transmits the network 170 using a wireless or wired communication method.
  • any of a variety of physical computer devices capable of communicating with other electronic devices 120, 130, 140 and / or servers 150 any of a variety of physical computer devices capable of communicating with other electronic devices 120, 130, 140 and / or servers 150,
  • the communication method is not limited, and may include a communication method using a communication network (for example, a mobile communication network, a wired Internet, a wireless Internet, a broadcasting network) that the network 170 may include, as well as a short-range wireless communication between the devices.
  • the network 170 may be a personal area network (LAN), a local area network (LAN), a campus area network (CAN), a metropolitan area network (MAN), a wide area network (WAN) , A network such as the Internet, and the like.
  • the network 170 may also include any one or more of a network topology including a bus network, a star network, a ring network, a mesh network, a star-bus network, a tree or a hierarchical network, It is not limited.
  • Each of the servers 150 and 160 is a computer device or a plurality of computers that communicate with a plurality of electronic devices 110, 120, 130 and 140 through a network 170 to provide commands, codes, files, Lt; / RTI > devices.
  • the server 150 may be a system that provides a first service to a plurality of electronic devices 110, 120, 130, 140 connected through a network 170, 170, and 140 to the first and second electronic devices 110, 120, 130, and 140, respectively.
  • the server 150 may be a computer program that is installed in a plurality of electronic apparatuses 110, 120, 130, and 140, Mail service, content transfer service, and the like) as a first service to a plurality of electronic devices 110, 120, 130, 140 in addition to the service for image processing, the information providing service, the messaging service,
  • the server 160 may provide a service for distributing a file for installing and running the application to the plurality of electronic devices 110, 120, 130, and 140 as a second service.
  • 2 is a block diagram illustrating an internal configuration of an electronic device and a server according to an embodiment of the present invention. 2 illustrates an internal configuration of the electronic device 1 (110) and the server 150 as an example of the electronic device. Other electronic devices 120, 130, 140 and server 160 may also have the same or similar internal configuration as electronic device 1 110 or server 150 described above.
  • the electronic device 1 110 and the server 150 may include memories 211 and 221, processors 212 and 222, communication modules 213 and 223 and input / output interfaces 214 and 224.
  • the memories 211 and 221 may be a computer-readable recording medium and may include a permanent mass storage device such as a random access memory (RAM), a read only memory (ROM), and a disk drive.
  • the non-decaying mass storage device such as a ROM and a disk drive may be included in the electronic device 110 or the server 150 as a separate persistent storage device different from the memory 211 or 221.
  • the memory 211 and the memory 221 are provided with an operating system and at least one program code (for example, a program installed in the electronic device 1 (110) and used for a browser or an application installed in the electronic device 1 Code) can be stored.
  • These software components may be loaded from a computer readable recording medium separate from the memories 211 and 221.
  • a computer-readable recording medium may include a computer-readable recording medium such as a floppy drive, a disk, a tape, a DVD / CD-ROM drive, and a memory card.
  • the software components may be loaded into memory 211, 221 via communication modules 213, 223 rather than a computer readable recording medium.
  • At least one program may be a computer program installed by files provided by a file distribution system (e.g., the server 160 described above) that distributes installation files of developers or applications, May be loaded into the memory 211, 221 based on the application (e.g., the application described above).
  • a file distribution system e.g., the server 160 described above
  • the application e.g., the application described above.
  • Processors 212 and 222 may be configured to process instructions of a computer program by performing basic arithmetic, logic, and input / output operations.
  • the instructions may be provided to the processors 212 and 222 by the memories 211 and 221 or the communication modules 213 and 223.
  • the processor 212, 222 may be configured to execute a command received in accordance with a program code stored in a recording device, such as the memory 211, 221.
  • the communication modules 213 and 223 may provide functions for the electronic device 1 110 and the server 150 to communicate with each other through the network 170 and may be provided to the electronic device 1 110 and / May provide a function for communicating with another electronic device (e.g., electronic device 2 120) or another server (e.g., server 160).
  • another electronic device e.g., electronic device 2 120
  • another server e.g., server 160
  • the request generated by the processor 212 of the electronic device 1 110 according to the program code stored in the recording device such as the memory 211 is transmitted to the server 170 via the network 170 under the control of the communication module 213 150 < / RTI > Conversely, control signals, commands, contents, files, and the like provided under the control of the processor 222 of the server 150 are transmitted to the communication module 223 of the electronic device 110 via the communication module 223 and the network 170 213 to the electronic device 1 (110).
  • control signals, commands, contents, files, and the like of the server 150 received through the communication module 213 can be transmitted to the processor 212 or the memory 211, (The above-mentioned persistent storage device), which may further include a storage medium 110.
  • the input / output interface 214 may be a means for interfacing with the input / output device 215.
  • the input device may include a device such as a keyboard or a mouse, and the output device may include a device such as a display, a speaker, and the like.
  • the input / output interface 214 may be a means for interfacing with a device having integrated functions for input and output, such as a touch screen.
  • the input / output device 215 may be composed of the electronic device 1 (110) and one device.
  • the input / output interface 224 of the server 150 may be a means for interfacing with the server 150 or an interface with a device (not shown) for input or output that the server 150 may include.
  • the configuration is performed using the data provided by the server 150 or the electronic device 2 (120)
  • a service screen or contents can be displayed on the display through the input / output interface 214.
  • electronic device 1 110 and server 150 may include more components than the components of FIG. However, there is no need to clearly illustrate most prior art components.
  • electronic device 1 110 may be implemented to include at least a portion of input / output devices 215 described above, or may be implemented with other components such as a transceiver, Global Positioning System (GPS) module, camera, Elements. More specifically, when the electronic device 1 (110) is a smart phone, the acceleration sensor, the gyro sensor, the camera module, various physical buttons, buttons using a touch panel, input / output ports, A vibrator, and the like may be further included in the electronic device 1 (110).
  • GPS Global Positioning System
  • 3 is a diagram showing an example of an execution environment of an image processing system according to an embodiment of the present invention.
  • electronic device 1 110 includes a camera module 310, an OCR recognition engine 320, a barcode recognition engine 330 and a QR code recognition engine 340 and the server 150 includes an image search engine 350 and a detec- tion animation generation engine 360 are shown.
  • the detec- tion animation generation engine 360 may be included in the first electronic device 110 according to the embodiment.
  • Each of the engines 320 to 360 may be implemented in the form of a software module.
  • the OCR recognition engine 320, the barcode recognition engine 330, and the QR code recognition engine 340 included in the electronic device 1 (110) are provided by an application installed and driven in the electronic device 1 (110) It can be a functional expression.
  • the processor 212 of the electronic device 110 may perform operations according to the OCR recognition engine 320, the barcode recognition engine 330, and the QR code recognition engine 340 according to an application code.
  • the image search engine 350 and the detection animation generation engine 360 included in the server 150 may be implemented in the form of a software module, and may be a functional expression provided by a computer program running on the server 150 .
  • the processor 222 of the server 150 may perform operations according to the image search engine 350 and the detection animation generation engine 360 according to the code of the computer program.
  • the OCR recognition engine 320 may recognize one or more characters and / or numbers within the image.
  • the barcode recognition engine 330 can recognize the barcode in the image.
  • the QR code recognition engine 340 can recognize the QR code in the image.
  • the image search engine 350 may receive an image as input and return various search results (image, text, etc.) associated with the image.
  • Detecting animation generation engine 360 may generate and provide a detaching animation for visually expressing the process of searching for an object in the image. Such detec- ting animation can be utilized as an effect to induce a user's interest while waiting for a search result and give a feeling that the search result is not delayed.
  • the electronic device 1 110 can drive the camera module 310 and the camera module 310 can be operated without any additional input from the user.
  • the barcode recognition engine 330 and the QR code recognition engine 340 To the inputs of the OCR recognition engine 320, the barcode recognition engine 330 and the QR code recognition engine 340.
  • Each of the OCR recognition engine 320, the bar code recognition engine 330 and the QR code recognition engine 340 sequentially analyzes the images (frames) of the input image stream and generates a corresponding object , A number, a bar code, a QR code, etc.).
  • the OCR recognition engine 320 may sequentially analyze images included in the image stream to attempt recognition of characters and / or numbers contained within the image, and may return recognized characters and / or numbers .
  • the electronic device 1 (110) may display the returned characters and / or numbers on the screen, and may provide additional information or additional functions associated with the displayed characters and / or numbers.
  • the electronic device 110 provides a user interface for accessing a translation function capable of translating the recognized first language character into a character of another language .
  • the electronic device 110 may provide a user interface for accessing a search function that uses the returned characters and / or numbers as a keyword.
  • the electronic device 1 (110) may automatically perform a search using a character and / or a number, which are returned, and provide the search result.
  • the barcode recognition engine 330 can sequentially analyze the images included in the image stream to attempt to recognize the barcode included in the image, and to return information on the recognized barcode.
  • the electronic device 1 (110) may provide additional information or additional functions associated with the returned information. For example, the electronic device 1 (110) searches the information on the returned barcode (for example, information on a book or wine corresponding to the barcode) corresponding to the barcode and returns the search result as additional information .
  • the electronic device 1 (110) may provide a user interface for accessing information corresponding to the bar code.
  • the QR code recognition engine 340 can sequentially analyze the images included in the image stream to attempt to recognize the QR code included in the image, and to return information on the recognized QR code.
  • the electronic device 1 (110) may provide additional information or additional functions associated with the returned information.
  • the electronic device 110 provides information (e.g., information corresponding to the URL included in the QR code) corresponding to the recognized QR code as additional information, or provides information corresponding to the recognized QR code And may provide a user interface for accessing information.
  • the electronic device 1 110 drives the camera module 310 without any additional input from the user, and the image provided through the camera module 310
  • the search for objects such as letters, numbers, barcodes, QR codes, etc. can be performed automatically, and additional information or additional functions related to the searched objects can be automatically provided.
  • the electronic device 1 (110) generates a user input through a predetermined user interface such as a selection of a user's photographing button (for example, a user touches an area of a photographing button displayed on the touch screen in a touch screen environment) Can be monitored.
  • a predetermined user interface such as a selection of a user's photographing button (for example, a user touches an area of a photographing button displayed on the touch screen in a touch screen environment) Can be monitored.
  • the process 370 monitors whether a user input is generated and, when a user input occurs, the captured image is transmitted to the server 150 according to user input. 2
  • the electronic device 1 110 may transmit the captured image via the network 170 to the server 150 using the communication module 213, and the server 150 may transmit the captured image via the network 170 to the server 150, May receive the captured image transmitted via the network 170 via the communication module 223.
  • the server 150 may provide the delivered image to the image search engine 350 and the detection animation generation engine 360, respectively.
  • the image search engine 350 may receive the captured image captured by the first electronic device 110, and may return various search results associated with the image. For example, the image search engine 350 may recognize an object included in the image and search for and return an image, document, text, or the like related to the recognized object. As a more specific example, when a puppy included in the image is recognized and the puppy of the recognized dog is analyzed as a 'retriever', a search result such as an image or a document related to the 'retriever' can be generated and returned.
  • the server 150 may transmit the retrieved search result to the electronic device 110 through the network 170 and the electronic device 110 may provide the search result to the user.
  • the electronic device 1 110 not only receives the image but also the time when capturing the image, the current position of the electronic device 110, information of the user of the electronic device 110, and the like to the server 150 It may be further transmitted.
  • the server 150 may provide search results based on at least one of location, time, and user information. For example, a search result related to a user's current position or a search result related to time among various search results associated with an image can obtain a priority for exposure of the search result.
  • the detec- tion animation generation engine 360 may generate a detec- tion animation for visually expressing a process of capturing an image captured by the electronic device 110 and receiving the received image as input and searching for an object in the image.
  • the server 150 may transmit the generated detecting animation to the first electronic device 110 via the network 170 and may associate the detected animation with the corresponding image for a predetermined time
  • the user of the electronic device 1 110 waits for the search result (the search result returned through the image search engine 350 described above and provided to the electronic device 1 110 at the server 150) And can be utilized as an effect to give an impression that the search result is not delayed.
  • Such a detecting animation may basically consist of a plurality of points of a position related to an object to be searched in the image and a line connecting these points and a representation of the lines connecting the points is displayed as an animation effect .
  • the thickness, size, brightness, color, etc. of dots and lines may be changed to provide an additional animation effect.
  • embodiments may be considered in which the lines formed by the lines connecting the points and the points are displayed in different colors to give a three-dimensional effect or curved lines connecting the points. This detecting animation will be described in more detail later.
  • FIGS. 4 and 5 illustrate examples of providing additional information or additional functions according to OCR recognition in an embodiment of the present invention.
  • FIG. 4 shows screen examples 410 to 440 of the electronic device 1 (110). 3, an image stream captured through the camera module 310 is automatically transmitted to the OCR recognition engine 320, and the OCR recognition engine 320 real-time recognizes characters As shown in FIG.
  • the color of the recognized character is changed and displayed on the image, and the T (TEXT) character recognition button is exposed.
  • the third screen example 430 shows an example in which functions related to recognized characters such as copying, translation, and reading of recognized characters are provided by selecting the T character recognition button.
  • the copy function may be a function for copying recognized characters to the clipboard.
  • the translation function may be a function for translating a recognized first language character into a second language character.
  • the reading function may be a function of reading the recognized first character, or may be a function of generating and outputting audio corresponding to the first character.
  • the fourth screen example 440 shows an example in which characters of the first language recognized as the user selects the translation function are translated and displayed as characters of the second language.
  • the detail view function displayed in the fourth screen example 440 may provide additional contents such as a language dictionary search result for the recognized first language character, a search result using the recognized first language character as a keyword, or the like It may be a function for landing on a separate translation result page. Further, if the recognition is not performed correctly or the translation result is not the desired result, a handwriting search function for proceeding with the search using the handwriting recognition function button may be further provided.
  • the first screen example 510 shows an example in which an image is displayed on the screen before recognition of text (characters and / or numbers).
  • the second screen example 520 shows an example in which recognized text is displayed at a position similar to the position of the text in the image as the text is recognized.
  • the third screen example 530 shows an example in which the recognized text is reconstructed and displayed so that the user can easily understand it. At this time, the third screen example 530 shows an example in which the image is darkened so that the recognized text can be expressed better. In the third screen example 530, an example of providing additional functions such as a copying function, a translation function, and a reading function in association with the recognized text is described.
  • the texts displayed in the second screen example 520 and the third screen example 530 may have an animation effect such as an animation effect such as a color change, a three-dimensional change in a two-dimensional frame in which texts are displayed, It may be displayed together.
  • an animation effect such as a color change, a three-dimensional change in a two-dimensional frame in which texts are displayed, It may be displayed together.
  • 6 to 8 are diagrams illustrating an example of adjusting the font size according to the number of recognized characters in an embodiment of the present invention.
  • FIG. 6 shows an example of displaying text recognized as 80 px size when the recognized number of characters is 1 to 6 characters.
  • FIG. 7 shows an example of displaying a text recognized as a 60px size when the recognized number of characters is 7 to 40 characters.
  • FIG. 8 shows an example of displaying a text recognized as a 40px size when the recognized number of characters is 41 or more.
  • the electronic device 1 (110) automatically adjusts the size of the displayed character.
  • FIG. 9 is a diagram illustrating an example of providing an additional function according to bar code recognition in an embodiment of the present invention.
  • the first screen example 910 shows an example in which an image including a bar code is displayed and the second screen example 920 shows an example in which a bar code button is exposed as a bar code is recognized by the bar code recognition engine 330 .
  • product information for example, book information, wine information, and the like
  • product information for example, book information, wine information, and the like
  • FIG. 10 is a diagram illustrating an example of providing an additional function according to QR code recognition in an embodiment of the present invention.
  • FIG. 10 shows screen examples 1010 and 1020 of the electronic device 1 (110).
  • an image including a QR code is displayed.
  • a QR code button is exposed as the QR code is recognized by the QR code recognition engine 340 For example.
  • the page of the URL included in the QR code can be landed.
  • 11 to 15 are views showing examples of providing an image search result in an embodiment of the present invention.
  • the first screen example 1110 shows an example in which an image is displayed
  • the second screen example 1120 shows an example in which an object is searched in an image.
  • the image can be transmitted to the server 150 as the user presses the photographing button.
  • the image retrieval engine 350 and the detec- tion animation generation engine 360 can perform image retrieval and detection animation Can be generated.
  • the second screen example 1120 receives a detecting animation composed of lines connecting the plurality of dots and points from the server 150 and displays it on the screen to visually detect that the dog is searching for a face As shown in FIG.
  • the third screen example 1130 shows an example in which "dog" and "labrador retriever” are displayed as text information (associated keywords) related to the searched object, and image search results for the image are further displayed.
  • the text information and the image search result may be provided through the server 150.
  • a text search result using the corresponding text information as a keyword may be provided to the user.
  • a page corresponding to the search result may be landed.
  • the first screen example 1210 shows an example in which an image is displayed and the second screen example 1220 and the third screen example 1230 show a detaching animation composed of lines connecting a plurality of dots and points, And an example of a process of visually informing the user that the fish is being searched.
  • the third screen example 1240 shows an example in which "aquarium fish" and "Asian arowana” are displayed as text information (associated keywords) related to the searched object, and image search results for the image are further displayed.
  • FIGS. 13 to 15 also show examples of processes for providing image search results as shown in FIGS. 11 and 12.
  • FIG. 13 illustrates a process of recognizing an object through detec- tion along a contour line of a flower, providing a detec- tion animation and an image search result
  • FIG. 14 illustrates a process of recognizing an object by detecting a cat face along an outline
  • FIG. 15 illustrates the process of providing an animation and an image search result
  • FIG. 15 illustrates a process of recognizing an object by detecting an Eiffel tower along an outline line, and providing a detecting animation and an image search result.
  • FIG. 16 is a diagram illustrating examples of providing an image search result in an embodiment of the present invention.
  • FIG. 17 is a diagram illustrating an example of limiting the saturation by extracting a main color in an embodiment of the present invention to be.
  • FIG. 16 shows an example of extracting a main color of an image using an auto technique and utilizing the extracted main color for exposure of a related keyword or an image search result.
  • the saturation (S) and the brightness (B) values may be limited to a range of 50 to 70% in the hue-saturation-brightness (HSB) value in consideration of the visibility of the text.
  • FIG. 17 shows an example in which the main color is extracted using a color picker to limit the saturation value to 50%.
  • FIGS. 18 to 20 are views showing examples of providing an image search result in an embodiment of the present invention.
  • Figures 18 to 20 illustrate the results of a method of determining whether a person is a member of a group of persons, such as "person_domain”, “person_group”, “person_outside”, “ And “ Domestic Place “, and the like, the specific search result of the image search result is implemented in the form of the correct answer card through the template designed in advance according to various types.
  • the saturation or brightness value of the HSV (Hue-Saturation-Brightness) value may be limited to 50 to 70% in view of the visibility of the text.
  • FIGS. 21 to 24 illustrate examples of displaying a detecting animation in an embodiment of the present invention.
  • the method of generating the detection animation is as follows.
  • a detection animation generation method may be performed by the detection animation generation engine 360 described above.
  • Preprocessing process Performing the pre-processing required for image search such as grayscale, blurring, and edge detection.
  • (2) Object Detection The process of searching for an object in the input image and creating a bounding box containing the object.
  • a well-known object search technique can be used to search for an object, and the following process can be performed on the generated bounding box area.
  • a convex polygon including all of the feature points may be generated as a convex hole, and the generation of the convex hole may be generated through a known algorithm.
  • the point constituting the convex hole does not reach a predetermined number (for example, 6), proceeding to an additional step to add points until the necessary number is reached.
  • a predetermined number for example, 6
  • it can be used as a point for constructing a convex hole by selecting a feature point closest to the middle between two points constituting a convex hole.
  • the process of generating the detecting animation according to the processes (1) to (9) may be modified into various forms according to the design as one embodiment.
  • FIG. 21 shows a detecting animation having a triangle structure by connecting points of a convex hole and a depth 3, which are composed of five points. More specifically, FIG. 21 shows a case in which pentagonal dots are formed and connected to an outline of an object which is a convex hole, two dots of depth are formed and connected to pentagonal dots, a left and a right of two dots of depth are connected, An example of generating a detecting animation form by vertically connecting two dots of depth and three dots of depth is shown. At this time, an example in which the depth 2 dots are formed at one-third of the line connecting the center of the connecting line between the pentagonal dots and the center point (depth 3 dots) is shown.
  • FIG. 22 shows a detecting animation having a triangle and a quadrilateral structure by connecting points of a convex hole and a depth 3, which are composed of nine dots.
  • the color difference between the contour dot (nine points constituting the convex hole) and the center dot (center point) is not less than a predetermined value (for example, 20 or more from the B value of the starting point RGB value of the contour dot
  • a dot having the closest distance between the dots of the upper and lower depths is connected to generate a detecting animation form.
  • FIGS. 23 and 24 examples of varying the color of the triangular structure to give a three-dimensional effect and simultaneously controlling the transparency are shown.
  • FIG. 23 shows an example in which the transparency in the triangular structure is higher in FIG.
  • FIG. 25 is a diagram showing an example of a detection animation in an embodiment of the present invention.
  • FIG. 25 is a view showing a result of the detection of the animation of the Epithelial Top described with reference to FIG. 15, as a result of repeating display of lines connecting points and dots as shown in FIG. 25, It can be utilized as an effect for inducing the user's interest while waiting for the search result and giving the user the feeling that the search result is not delayed.
  • the detecting animation may change the thickness, size, brightness, color, etc. of dots and lines to give additional animation effects.
  • FIGS. 23 and 24 By applying differently, it is possible to determine the extent to which the image is displayed by giving a three-dimensional effect or adjusting the transparency of the color.
  • FIG. 26 is a flowchart showing an example of an image processing method according to an embodiment of the present invention.
  • the image processing method according to the present embodiment can be performed by a computer apparatus such as the above-described electronic apparatus 1 (110).
  • the processor 212 of the electronic device 1 110 may be implemented to execute a control instruction according to code of an operating system included in the memory 211 or code of at least one computer program.
  • the processor 212 controls the electronic device 1 (110) to perform the steps 2610 to 2640 included in the image processing method of FIG. 26 according to the control command provided by the code stored in the electronic device 1 (110) It is possible to control the device 1 (110).
  • the computer device may drive the camera module in response to entering the image processing mode.
  • an application installed on a computer device may provide a user interface for entering the image processing mode to the user.
  • the computer device can drive the camera module according to the entering image processing mode according to the generated input.
  • the computer device may sequentially receive and sequentially display a stream of images through the driven camera module. For example, when a camera is operated in a smartphone, an image input through a camera and displayed in real time on a screen of a smartphone may correspond to a stream of such images.
  • the computer device may deliver a stream of received images to the recognition engine.
  • the camera module can deliver a continuous stream of images to the recognition engine on a continuous and real-time basis, and the recognition engine can analyze the stream of images to generate the recognition results desired by the recognition engine.
  • electronic device 1 110 includes a text recognition engine, such as OCR recognition engine 320, and image code recognition engines, such as bar code recognition engine 330 and QR code recognition engine 340, An example of including this has been described.
  • the computer device may further display the stream of the received images on the screen being displayed sequentially.
  • the recognition engine may include a text recognition engine that recognizes text included in the images received as recognition results.
  • the computer device can dynamically adjust the display position of the recognized text based on the position of the text area recognized in the input images in step 2640 have.
  • the first screen example 410 of FIG. 4 shows an example in which recognized text is displayed at a location of an area containing text in an image.
  • the second screen example 520 of FIG. 5 an example in which recognized text is displayed at a position similar to a position in the image of the text as the text is recognized is described.
  • the computer device such as a smart phone shakes in the user's hand, the position of the same text area in the received images may be continuously changed.
  • the computer device can dynamically adjust the display position of the text so that the recognized text can be displayed at the tracked position by tracking the position of the text area as the position of the text area is changed.
  • the computer device displays the recognized text on the screen sequentially displaying the streams of the input images, wherein the text of the input images is based on the position of the text area in the recognized image
  • the display position of the recognized text can be determined.
  • the computer device dynamically adjusts the display position of the recognized text as the position of the text area is changed, and when the recognition of the text is finally completed, The text can be displayed.
  • the display position of the recognized text may be fixed even if the position of the text area continuously changes in the received images or other images without text are continuously input.
  • the computer device may further display a stream of input images on a screen being displayed sequentially, through a user interface for user confirmation of the recognized text with respect to a stream of inputted images.
  • the computer device can finally process the process of receiving the user's confirmation of the recognized text.
  • the second screen 420 shown in FIG. 4 shows an example of displaying a user interface for confirming a recognized text such as a character recognition button. At this time, a stream of images may still be displayed in real time on the screen of the computer device.
  • the computer device can display an image displaying the text of the received images on the screen. For example, if the text recognized by the user is recognized through the first image of the input images, the computer device may display the first image that has already been displayed on the screen instead of displaying the stream of images on the screen.
  • the computer device may further display at least one of a user interface for copying the recognized text on the image displayed on the screen and a user interface for the translation of the recognized text.
  • FIGS. 4 and 5 show examples of user interfaces for copying and translating recognized text, such as a 'Copy' button and a 'Translation' button.
  • the computer device may dynamically change the size of the recognized text according to the number of characters of the recognized text, and display the changed size on the screen. For example, in FIGS. 6 to 8, an example has been described in which the recognized text is dynamically reduced and displayed on the screen as the number of characters in the recognized text increases.
  • the recognition engine may include an image code recognition engine for recognizing the image code included in the received images.
  • the computer device displays a stream of images sequentially received as a recognition result of the link to the page corresponding to the recognized image code in the images sequentially displayed on the screen in step 2640, can do.
  • FIG. 9 shows an example of displaying a link to a page corresponding to a recognized bar code through a 'barcode' button.
  • a link to a page corresponding to the recognized QR code is referred to as a 'QR code' Button is displayed.
  • the recognition engine may comprise a plurality of recognition engines, such as a text recognition engine and an image code recognition engine, in which case a stream of images input via the camera module may be input to each of a plurality of recognition engines .
  • the steps 2610 to 2640 described above may be performed in such a manner that the computer device automatically recognizes text, image codes, and the like through a stream of images input through the camera before the user's photographing through the camera is performed An embodiment is described. In other words, in response to entering the image processing mode, even if the user does not select an image at a specific point in time by pressing the photographing button, the recognition result for the text or image code can be automatically provided have.
  • the computer device can provide the user with a function different from the previous embodiment.
  • FIG. 27 is a view showing another example of the image processing method according to the embodiment of the present invention.
  • the image processing method according to the present embodiment can also be performed by a computer apparatus such as the electronic apparatus 1 (110) described above. Steps 2710 through 2750 of FIG. 27 may be performed after the step 2620 of FIG. 26, when a shooting input occurs, and when steps of FIG. 27 are performed, steps 2630 and 2640 ) May be omitted.
  • step 2710 when a shooting input occurs during the sequential display of a stream of input images, the computer device may capture an image associated with the shooting input occurrence time and display the captured image on the screen.
  • This step 2710 may refer to a process of taking a picture according to the occurrence of a photographing input by a user.
  • the computer device may send the captured image to the server.
  • the server may correspond to a computer device such as the server 150 described above, and the captured image may be transmitted to the server via the network 170.
  • the computer device may receive a detaching animation for an object included in the image transmitted from the server.
  • the detection animation may include an animation that displays a plurality of feature points extracted from an outline of an object at a position on an extracted image of a plurality of feature points, and connects at least some feature points of the displayed feature points with lines .
  • the computing device may display the detection animation in association with the object. Examples of displaying the detection animation in association with the object on the screen have been described with reference to Figs. 11 to 15, and Figs. 21 to 25. Fig.
  • the computer device receives the image analysis result of the image transmitted from the server and displays the image analysis result in association with the image displayed on the screen.
  • the image analysis result may include a kind and / or name of an object included in the image, and may further include at least one of images, documents, and text retrieved from the server in association with the object.
  • FIG. 28 is a diagram showing another example of the image processing method according to the embodiment of the present invention.
  • the image processing method according to the present embodiment can be performed by a computer apparatus such as the server 150 described above.
  • the processor 222 of the server 150 may be implemented to execute control instructions in accordance with the code of the operating system or the code of at least one computer program that the memory 221 contains.
  • the processor 222 determines whether the server 150 performs the steps 2810 through 2850 included in the image processing method of FIG. 28 in accordance with the control command provided by the code stored in the server 150 Can be controlled.
  • the computer device may receive the captured image via the camera module of the electronic device over the network.
  • the electronic device may correspond to the electronic device 1 (110) described above, and the received image may correspond to the image transmitted in step 2720 of FIG.
  • the server 150 can receive the image in step 2810.
  • the computing device may generate a detaching animation for the object that the received image contains.
  • the detection animation may include an animation that displays a plurality of feature points extracted from an outline of an object at a position on the extracted image of the plurality of feature points, and connects at least some of the feature points of the displayed feature by lines. have.
  • the computer device may transmit the generated detecting animation to the electronic device.
  • the electronic device may receive the detection animation and display the received detection animation in association with the object on the screen, as in step 2740.
  • the computer device may generate an image search result for the received image.
  • the image search result may include the type and / or name of an object included in the image, and may further include an image, document and / or text retrieved in association with the object.
  • the computer device may send the generated image search results to the electronic device.
  • the type and / or the name of the object may overlap with the image input through the camera of the electronic device and displayed on the screen of the electronic device.
  • the retrieved image, document and / or text may be further displayed on the screen of the electronic device in association with the image input through the camera module of the electronic device, including a link to the corresponding page.
  • Figure 12 shows the types and names of recognized objects, such as 'ornamental language' and 'Asian arowana', and further associates images, documents and / or text with images as a result of an Internet search through images Is displayed.
  • the computer device may implement the information retrieved through the template designed in advance for each type according to the type of the information retrieved in association with the object included in the received image, in the form of a card, and provide the retrieved information to the electronic device.
  • Figs. 18 to 20 illustrate examples of the types of information (person_domain, person_group, person_outside, bag_ animal, bag_furniture, bag_wine, spot, area, domestic_place, etc.)
  • There is a pre-designed template and information retrieved through a template of the type according to the type of the retrieved information is implemented and provided in the form of a card.
  • the type and / or the name of the object displayed on the electronic device is selected (for example, the user touches the area where the type and / or the name of the object is displayed in the touch screen environment with the finger)
  • the generated signal can be transmitted to the server via the network.
  • the computer device can receive the corresponding signal, and in response to receiving the signal, the text search result can be generated by using the type or name of the object as a keyword.
  • the computer device can provide the generated text search result to the electronic device.
  • the user of the electronic device can receive sequentially the text search result on the text obtained through the image, in addition to the image search result on the image.
  • 29 is a flowchart illustrating an example of a method of generating a detaching animation in an embodiment of the present invention.
  • the steps 2910 to 2950 included in the method of the present embodiment may be performed in step 2820 of Fig.
  • the computer device may search for an object that the received image contains. For example, the computer device performs a preprocessing operation required for image search such as grayscale, blurring, and edge detection on a received image, and then searches for an object in the image and creates a bounding box including the object Can be generated. Searching for these objects can generally be done using well-known object search techniques.
  • the computing device may extract a plurality of feature points from the contour of the object. For example, a predetermined number of feature points (e.g., several hundreds) can be extracted using a Feature from Accelerated Segment Test (FAST) algorithm.
  • FAST Accelerated Segment Test
  • the computer device may generate a convex hull for the extracted feature points.
  • a convex hole can be generated from minutiae extracted through the convex hole algorithm. If the point for constructing the convex hole does not reach the predefined number, the feature points may be further extracted.
  • the computer device may calculate a center point based on the center coordinates of a predetermined number of outline points constituting the convex hole among the extracted minutiae points. For example, the points of the outer points constituting the convex hole (the average of the X coordinate value and the average of the Y coordinate value) can be calculated centrally.
  • the computer device may select feature points that are closest to the midpoint between each of the outline points and the midpoint among the feature points of the object.
  • the selected minutiae can be used again as outer minus points to obtain the center point, and then the inner minus points can be added by selecting the minutiae points between the outer points and the center point.
  • FIG. 21 shows an example in which a detecting animation having a triangular structure is formed by connecting points of a convex hole and a depth 3 which are composed of five points.
  • the generated detection animation may be transmitted to the electronic device as in step 2830 of FIG. 28 and may be displayed on the screen of the electronic device in association with the object in the electronic device, such as in step 2740 of FIG. 27 .
  • the computer device may transmit information on the coordinates of the selection points including the outer points, the center point and the closest minutiae points, and the order of connecting the selection points by lines to the electronic device as a detecting animation.
  • the electronic device may display a line through the information of the coordinates of the selected points, and may display an animation connecting the selected points with lines according to the information on the sequence.
  • the computer device may also transmit the animation itself connecting the selection points in this order to the electronic device as a detection animation. In this case, the electronic device can display the detection animation by reproducing the animation in association with the object.
  • 30 is a diagram showing an example of providing additional information according to place recognition in an embodiment of the present invention.
  • 30 shows screen examples 3010 and 3020 of the electronic device 1 (110).
  • the first screen example 3010 shows an example in which information (for example, a text such as mutual or a picture corresponding to a specific name, etc.) capable of identifying a specific place such as a signboard of a shop is displayed on the image.
  • information for example, a text such as mutual or a picture corresponding to a specific name, etc.
  • an image stream captured through the camera module 310 may be automatically delivered to the OCR recognition engine 320 and recognized by the OCR recognition engine 320 in real time.
  • the electronic device 1 (110) can determine whether the recognized character is information for identifying a specific place according to the control of the application.
  • the electronic device 110 may transmit the recognized character or the image shown in the first screen example 3010 to the server 150.
  • the server 150 can obtain a more accurate place identifier, and can extract metadata of the place (for example, mutual, business type, description, etc., in the case of a store) and transmit it to the electronic device 1 (110).
  • the second screen example 3020 shows an example where the metadata of the place provided from the server 150 is displayed on the screen in the form of an upper notification bar 3021. At this time, a link to a URL related to the place may be set in the upper notification bar 3021.
  • the electronic device 1 (110) may analyze the image displayed in the first screen example (3010) according to the control of the application to determine whether the image is an image for a specific place.
  • the electronic device 1 110 roughly analyzes the image to determine whether or not the image is a specific place, and then transmits the image to the server 150 so that an identifier of a more accurate place can be extracted through the server 150 Lt; / RTI >
  • FIG. 31 is a diagram showing an example of providing additional information according to recognition of an image code in an embodiment of the present invention.
  • the QR code button is exposed as the electronic device 1 (110) recognizes the QR code and the user selects the exposed QR code button, an example in which the page of the URL included in the QR code is landed .
  • the first screen example 3110 shows an example in which an image including a QR code is displayed.
  • the second screen example 3120 obtains the metadata of the URL included in the recognized QR code as the QR code is recognized by the QR code recognition engine 340 from the server 150, As shown in Fig.
  • a link to a URL included in the QR code may be set in the upper notification bar 3121.
  • the metadata of the related product can be obtained from the server 150 in the form of an upper notification bar in addition to the QR code as well as the bar code.
  • the link set in the upper notification bar may be a page associated with the purchase of the related product.
  • the metadata displayed in the upper notification bar is information included in a page provided through a URL included in the image code, and may include various information such as a URL, a moving image, an image, a description, and the like.
  • embodiments of the present invention may provide a user interface associated with meaningful and / or meaningful information associated with at least one frame of a video stream that is continuously captured through a camera.
  • the system or apparatus described above may be implemented as a hardware component, a software component or a combination of hardware components and software components.
  • the apparatus and components described in the embodiments may be implemented within a computer system, such as, for example, a processor, a controller, an arithmetic logic unit (ALU), a digital signal processor, a microcomputer, a field programmable gate array (FPGA) , A programmable logic unit (PLU), a microprocessor, or any other device capable of executing and responding to instructions.
  • the processing device may execute an operating system (OS) and one or more software applications running on the operating system.
  • the processing device may also access, store, manipulate, process, and generate data in response to execution of the software.
  • OS operating system
  • the processing device may also access, store, manipulate, process, and generate data in response to execution of the software.
  • the processing apparatus may be described as being used singly, but those skilled in the art will recognize that the processing apparatus may have a plurality of processing elements and / As shown in FIG.
  • the processing unit may comprise a plurality of processors or one processor and one controller.
  • Other processing configurations are also possible, such as a parallel processor.
  • the software may include a computer program, code, instructions, or a combination of one or more of the foregoing, and may be configured to configure the processing device to operate as desired or to process it collectively or collectively Device can be commanded.
  • the software and / or data may be in the form of any type of machine, component, physical device, virtual equipment, computer storage media, or device As shown in FIG.
  • the software may be distributed over a networked computer system and stored or executed in a distributed manner.
  • the software and data may be stored on one or more computer readable recording media.
  • the method according to an embodiment may be implemented in the form of a program command that can be executed through various computer means and recorded in a computer-readable medium.
  • the computer-readable medium may include program instructions, data files, data structures, and the like, alone or in combination.
  • the medium may be one that continues to store computer executable programs, or temporarily store them for execution or download.
  • the medium may be a variety of recording means or storage means in the form of a combination of a single hardware or a plurality of hardware, but is not limited to a medium directly connected to a computer system, but may be dispersed on a network.
  • Examples of the medium include a magnetic medium such as a hard disk, a floppy disk and a magnetic tape, an optical recording medium such as CD-ROM and DVD, a magneto-optical medium such as a floptical disk, And program instructions including ROM, RAM, flash memory, and the like.
  • a recording medium or a storage medium managed by a site or a server that supplies or distributes an application store or various other software to distribute the application may be mentioned.
  • Examples of program instructions include machine language code such as those produced by a compiler, as well as high-level language code that can be executed by a computer using an interpreter or the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Processing Or Creating Images (AREA)

Abstract

La présente invention a trait à une technologie de traitement d'une entrée visuelle. Un procédé de traitement d'images selon un mode de réalisation peut fournir des informations significatives associées à au moins une trame d'un flux d'images continuellement capturées par une caméra, et/ou une interface utilisateur associée aux informations significatives.
PCT/KR2018/009470 2017-08-20 2018-08-17 Traitement d'entrée visuelle WO2019039799A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2020509490A JP7027524B2 (ja) 2017-08-20 2018-08-17 視覚的入力の処理

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
KR10-2017-0105156 2017-08-20
KR20170105156 2017-08-20
KR10-2018-0035734 2018-03-28
KR1020180035734A KR102026475B1 (ko) 2017-08-20 2018-03-28 시각적 입력의 처리

Publications (1)

Publication Number Publication Date
WO2019039799A1 true WO2019039799A1 (fr) 2019-02-28

Family

ID=65439151

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2018/009470 WO2019039799A1 (fr) 2017-08-20 2018-08-17 Traitement d'entrée visuelle

Country Status (1)

Country Link
WO (1) WO2019039799A1 (fr)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010103694A (ja) * 2008-10-22 2010-05-06 Seiko Precision Inc 翻訳機能付きカメラ、及びテキストの表示方法
KR20130057584A (ko) * 2011-11-24 2013-06-03 에스케이플래닛 주식회사 특징점 설정 장치 및 방법과 이를 이용한 객체 추적 장치 및 방법
JP5852093B2 (ja) * 2011-03-01 2016-02-03 パナソニック インテレクチュアル プロパティ コーポレーション オブアメリカPanasonic Intellectual Property Corporation of America 映像処理装置、映像処理方法、プログラム
KR20160149185A (ko) * 2016-12-20 2016-12-27 에스케이플래닛 주식회사 통합 영상 검색 시스템 및 그 서비스 방법
KR20170091878A (ko) * 2016-02-02 2017-08-10 주식회사 얍컴퍼니 시각적 콘텐츠의 객체 인식 장치 및 방법

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010103694A (ja) * 2008-10-22 2010-05-06 Seiko Precision Inc 翻訳機能付きカメラ、及びテキストの表示方法
JP5852093B2 (ja) * 2011-03-01 2016-02-03 パナソニック インテレクチュアル プロパティ コーポレーション オブアメリカPanasonic Intellectual Property Corporation of America 映像処理装置、映像処理方法、プログラム
KR20130057584A (ko) * 2011-11-24 2013-06-03 에스케이플래닛 주식회사 특징점 설정 장치 및 방법과 이를 이용한 객체 추적 장치 및 방법
KR20170091878A (ko) * 2016-02-02 2017-08-10 주식회사 얍컴퍼니 시각적 콘텐츠의 객체 인식 장치 및 방법
KR20160149185A (ko) * 2016-12-20 2016-12-27 에스케이플래닛 주식회사 통합 영상 검색 시스템 및 그 서비스 방법

Similar Documents

Publication Publication Date Title
CN107223246B (zh) 图像标注方法、装置及电子设备
CN110446063B (zh) 视频封面的生成方法、装置及电子设备
CN110225387A (zh) 一种信息搜索方法、装置及电子设备
WO2020029466A1 (fr) Procédé et appareil de traitement d'image
WO2018174314A1 (fr) Procédé et système de production d'une séquence vidéo d'histoire
WO2021141419A1 (fr) Procédé et appareil pour générer un contenu personnalisé en fonction de l'intention de l'utilisateur
WO2020060019A1 (fr) Dispositif, procédé et système de détection de caractère
WO2019017582A1 (fr) Procédé et système de collecte de modèles de contenu de réalité augmentée en fonction d'une source en nuage et de génération automatique d'un contenu de réalité augmentée
WO2018131825A1 (fr) Procédé de fourniture de service de livre électronique et programme informatique associé
CN115061679B (zh) 离线rpa元素拾取方法及系统
WO2018052269A1 (fr) Procédé d'affichage d'images dans un mode multidimensionnel en fonction de sujets personnalisés
WO2020116768A1 (fr) Dispositif de traitement d'image et son procédé de fonctionnement
WO2019093599A1 (fr) Appareil permettant de générer des informations d'intérêt d'un utilisateur et procédé correspondant
JP2019075130A (ja) 情報処理装置、制御方法、プログラム
KR102422221B1 (ko) 이미지 내 텍스트 색상과 배경 색상을 추출하여 제공하는 방법, 시스템, 및 컴퓨터 프로그램
WO2019066231A1 (fr) Génération d'image représentative
CN105955683B (zh) 系统和控制方法
WO2014148691A1 (fr) Dispositif mobile et son procédé de commande
EP3896607A2 (fr) Procédé et appareil de collocation de vêtements et dispositif informatique et support
CN111198962A (zh) 信息处理装置、系统、方法、类似与否判断方法以及介质
WO2018164435A1 (fr) Appareil électronique, son procédé de commande, et support de stockage non transitoire lisible par ordinateur
WO2019039799A1 (fr) Traitement d'entrée visuelle
KR102026475B1 (ko) 시각적 입력의 처리
CN115546824B (zh) 禁忌图片识别方法、设备及存储介质
EP3577583A1 (fr) Appareil électronique, son procédé de commande, et support de stockage non transitoire lisible par ordinateur

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18848367

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2020509490

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18848367

Country of ref document: EP

Kind code of ref document: A1