WO2005096760A2 - Dispositif de lecture portatif avec traitement modal - Google Patents

Dispositif de lecture portatif avec traitement modal Download PDF

Info

Publication number
WO2005096760A2
WO2005096760A2 PCT/US2005/011180 US2005011180W WO2005096760A2 WO 2005096760 A2 WO2005096760 A2 WO 2005096760A2 US 2005011180 W US2005011180 W US 2005011180W WO 2005096760 A2 WO2005096760 A2 WO 2005096760A2
Authority
WO
WIPO (PCT)
Prior art keywords
user
image
program product
computer program
machine
Prior art date
Application number
PCT/US2005/011180
Other languages
English (en)
Other versions
WO2005096760A3 (fr
Inventor
Raymond C. Kurzweil
Paul Albrecht
James Gashel
Lev Lvovsky
Kenneth Linde
Lucy Gibson
Original Assignee
Kurzweil Technologies, Inc.
National Federation Of The Blind
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US11/096,714 external-priority patent/US7641108B2/en
Priority claimed from US11/097,956 external-priority patent/US7659915B2/en
Priority claimed from US11/097,094 external-priority patent/US7840033B2/en
Priority claimed from US11/097,959 external-priority patent/US20060020486A1/en
Priority claimed from US11/097,448 external-priority patent/US7629989B2/en
Priority claimed from US11/096,711 external-priority patent/US7325735B2/en
Priority claimed from US11/096,703 external-priority patent/US8249309B2/en
Priority claimed from US11/097,944 external-priority patent/US7505056B2/en
Priority claimed from US11/097,958 external-priority patent/US8036895B2/en
Priority claimed from US11/097,978 external-priority patent/US8320708B2/en
Priority claimed from US11/096,712 external-priority patent/US7627142B2/en
Priority claimed from US11/097,986 external-priority patent/US8873890B2/en
Priority claimed from US11/097,447 external-priority patent/US9236043B2/en
Priority to EP05733206A priority Critical patent/EP1756802A2/fr
Application filed by Kurzweil Technologies, Inc., National Federation Of The Blind filed Critical Kurzweil Technologies, Inc.
Publication of WO2005096760A2 publication Critical patent/WO2005096760A2/fr
Publication of WO2005096760A3 publication Critical patent/WO2005096760A3/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/14Image acquisition
    • G06V30/142Image acquisition using hand-held instruments; Constructional details of the instruments

Definitions

  • the computer program product further includes instructions for causing the reading device to process low resolution images for detecting gesture or user defined commands in gesture or directed reading modes and process,- using- optical character recognition, on the high resolution image for recognition of text on the high resolution image.
  • the reading device generates a multi-channel auditory signal and the perceived directionality of the signal conveys information about what is being read to the user.
  • the reading device includes a member configured to fit about a portion of a user' s head, with the image input device supported by the member and with the image input device having a communications link to feed data from the image input device to the computing device.
  • the link is wireless or cable.
  • the member further supports an integrated speaker or earphone to allow the user to hear speech that is produced from the reading device and uploaded to the integrated speaker or earphone through the link.
  • the computer program product also includes instructions to process the one or several images to identify one or more of the target elements in the image using information obtained from the knowledge base.
  • a method of operating a reading machine includes receiving data that specifies a mode to use for processing an image, accessing a knowledge base to provide data to configure the computer program product for the specified mode, with the data including specific target visual elements that are expected to be encountered in processing images for the specified mode and capturing one or several images.
  • the method also includes processing the one or several images to identify one or more of the target elements in the image using information obtained from the knowledge base.
  • One or more aspects of the invention may provide one or more of the following advantages.
  • the reading device is configured to operate in different modes.
  • a method of operating a device includes receiving data by the device that specifies a transaction mode to use for operating the device and accessing a knowledge base to provide data to configure the device for the transaction mode, the data including data specific to the transaction mode.
  • the method also includes receiving an image or images of a portion of a machine that the user will use to perform the transaction, processing the image or images to identify a pattern of controls on the machine and to detect the presence of a user-controlled pointing item over controls on the machine, and announcing to the user the name or function of the control closest to an end of the user-controlled pointing item.
  • an apparatus includes a computing device, an image input device and a computer readable medium storing a computer program product assisting a user of the apparatus to use the apparatus to perform a transaction with a machine.
  • the computer program product includes instructions for causing the computing device to receive data that specifies a transaction mode, access a knowledge base to provide data to configure the apparatus for the transaction mode, the data including data specific to the transaction mode and receive an image or images from the image input device the image or images being of a portion of a machine that the user will use to perform the transaction.
  • the computer program product also includes instructions to process the image or images to identify a pattern of controls on the machine and to detect the presence of a user-controlled pointing item over controls on the machine and announce to the user the name or function of the control closest to an end of the user-controlled pointing item.
  • One or more aspects of the invention may provide one or more of the following advantages.
  • the reading device determines various characteristics of the captured image to inform the user of identified controls that were recognized by the device. This allows the user to use the device as an aid to help the user perform a transaction with a machine such as an automatic teller machine (ATM) , e-ticket machine, electronic voting machine, credit/debit point of sale machine, and so forth. This device gives users such as blind or visually impaired persons additional autonomy.
  • ATM automatic teller machine
  • the instructions determine an aesthetically valid match include instructions to provide a description of colors and patterns for the first and subsequent articles of clothing, retrieve data from a knowledge base containing rules on how to match clothes, and process the information to determine whether the colors and patterns of the first and subsequent articles of clothing provide the aesthetically valid match according to the rules.
  • the computer program product includes instructions to report to the user whether an aesthetically valid match was determined.
  • the computer program product includes instructions to receive data that specifies a clothing mode to use for processing an image, access a knowledge base to provide data to configure the computer program product for the clothing mode, the data including data specific to the clothing mode and receive an image or images of an article of clothing.
  • the computer program product includes instructions to process the image or images to identify patterns in the image corresponding to items of clothing based on information obtained from the knowledge base .
  • One or more aspects of the invention may provide one or more of the following advantages.
  • the "clothing" mode can help the user, e.g., to get dressed by matching clothing based on color and pattern.
  • the clothing mode is helpful for those who are visually impaired, including those who are colorblind but otherwise have normal vision.
  • the computer program product classifies the document according to key words or phrases in the document and retrieves data from the knowledge base that is associated with the classification of the document .
  • the instructions are configured to permit a user to select sections of documents that are books, memos, newspapers or journals.
  • the computer program product includes instructions to save data associated with reading a document whose sections were identified, the associated data being data retrieved from a knowledge base for the document and data produced or used while processing the document.
  • the user commands include read text, spell word, speak faster/slower, skip sentence or paragraph, and read named part of template. Processing user commands includes detecting that the user command did not provide a user command; and selecting a default for the reading machine to select the document section.
  • a document mode 130 is provided to read books, magazines and paper copy.
  • the document mode 130 supports various layout variations found in memos, journals and books. Data regarding the document mode is retrieved 132 from the knowledge base.
  • the document mode 130 accommodates different types of formats for documents.
  • the contents of received 134 image (s) are compared 136 against different document models retrieved from the knowledge base to determine which model (s) match best to the contents of the image.
  • the document mode supports multi-page documents in which the portable reading machine combines 138 information from multiple pages into one composite internal representation of the document that is used in the reading machine to convey information to the user.
  • the portable reading machine processes pages, looking for page numbers, section headings, figures captions and any other elements typically found in the particular document.
  • the portable reading machine may identify the standard sections of the patent, including the title, inventors, abstract, claims, etc.
  • the document mode allows a user to navigate 140 the document contents, stepping forward or backward by a paragraph or section, or skipping to a specific section of the document or to a key phrase .
  • the portable reading machine reads 142 the document to a user using text-to-speech synthesis software. Using such an internal representation allows the reading machine to read the document more like a sighted person would read such a document.
  • Each element may include several parameters besides color, such as width (for stripes) , or orientation (e.g. vertical stripes). For example, slacks may be described by the device as "gray vertical stripes on a black background", or a jacket as "Kelly green, deep red and light blue plaid” .
  • the portable reading machine receives 156 input data corresponding to the scanned clothing and identifies 158 various attributes of the clothing by processing the input data corresponding to the captured images in accordance with parameters received from the knowledge base.
  • the portable reading machine reports 160 the various attributes of the identified clothing item such as the color (s) of the scanned garment, patterns, etc.
  • the clothing attributes have associated descriptions that are sent to speech synthesis software to announce the report to the user.
  • the portable reading machine recognizes the presence of patterns such as stripes or check by comparisons to stored patterns or using other pattern recognition techniques.
  • the clothing mode may "learn" 162 the wardrobe elements (e.g. shirts, pants, socks) that have characteristic patterns, allowing a user to associate specific names or descriptions with individual articles of clothing, making identification of such items easier in future uses.
  • the machine may have a mode that matches a given article of clothing to another article of clothing (or rejects the match as incongruous) .
  • This automatic clothing matching mode makes use of two references: one is a database of the current clothes in the user's possession, containing a description of the clothes ' colors and patterns as described above .
  • the transaction mode 170 applies to transaction-oriented devices that have a layout of controls, e.g. buttons, such as automatic teller machines (ATM) , e-ticket devices, electronic voting machines, credit/debit devices at the supermarket, and so forth.
  • the portable reading machine 10 can examines a layout of controls, e.g., buttons, and recognize the buttons in the layout of the transaction-oriented device.
  • the portable reading machine 10 can tell the user how to operate the device based on the layout of recognized controls or buttons.
  • many of these devices have standardized layouts of buttons for which the portable reading machine 10 can have stored templates to more easily recognize the layouts and navigate the user through use of the transaction-oriented device.
  • data can be retrieved from a database on the reading machine, from the transaction device or via another device.
  • Data retrieval to make the transaction mode more robust and accurate can involve a layout of the device, e.g., an automatic teller machine (ATM) , which is pre-programmed or learned as a customized mode by the reading machine. This involves a sighted individual taking a picture of the device and correctly identifying all sections and buttons, or a manufacturer providing a customized database so that the user can download the layout of the device to the reading machine 10.
  • ATM automatic teller machine
  • the knowledge base can include a range of relevant information.
  • the mode knowledge base includes general information, such as the expected fonts, vocabulary or language most commonly encountered for that device .
  • the knowledge base can also include very specific information, such as templates that specify the layout or contents of specific screens.
  • the reading machine can detect many kinds of pointers, such as a stylus which may be used with a touchscreen, a pen, or any other similar pointing device.
  • the video input device starts 176 taking images at a high frame rate with low resolution.
  • Low resolution images may be used during this stage of pointer detection, since no text is being detected.
  • Using low resolution images will speed processing, because the low resolution images require fewer bits than high resolution images and thus there are fewer bits to process.
  • the reading machine processes those low resolution images to detect 178 the location of the user's pointer.
  • the reading machine determines 180 what is in the image underlying, adjacent, etc. the pointer.
  • the outdoor mode is intended the help the user with physical navigation.
  • the portable reading machine may look for street signs and building signs. It may look for traffic lights and their status. It may give indications of streets, buildings or other landmarks.
  • the portable reading machine may use GPS or compass and maps to help the user get around.
  • the portable reading machine may take images at a faster rate and lower resolution process those images faster (do to low resolution) , at relatively more current positions (do to high frame rate) to provide more "real-time" information such as looking for larger physical objects, such as buildings, trees, people, cars, etc.
  • the portable reading machine can include an "Indoor Navigation Mode.”
  • the indoor navigation mode helps a person navigate indoors, e.g., in an office environment.
  • the site may be specialized to handle time-consuming tasks such as OCR, using multiple servers and large databases in order to process efficiently.
  • the ability of the processing subsystem to hold the reference information reduces the amount of I/O traffic between the card and the portable reader.
  • the reader 10 may only need to send captured image data to the subsystem once and then make many requests to the subsystem to process and analyze the different sections of the image for text or shapes.
  • the portable reading machine 10 includes features to improve the quality of a captured image. For instance, the portable reading machine could use image stabilization technology found in digital camcorders to keep the text from becoming blurry. This is especially important for smaller print or features and for the mobile environment.
  • the portable reading machine 10 can ' include a digital camera system that uses a zoom capability to get more resolution for specific areas of the image.
  • the portable reading machine can use auto balancing or a range of other image enhancement techniques to improve the image quality.
  • the portable reading machine could have special enhancement modes to enhance images from electronic displays such as LCD displays.
  • the projection yields a graph that that is has periodic amplitude fluctuations, with the peaks corresponding to lines of text and the valleys corresponding to the gaps between.
  • the resulting graph is smoother. Finding the angles that yield a high amplitude periodicity, one can provide a good estimate for an angle that is likely to yield good OCR results.
  • the spatial frequency of the periodicity gives the line spacing, and is likely to be a good indicator of the font size, which is one of the factors that determine the performance of an OCR algorithm.
  • the reading machine can compute the distance D from the camera to a point on the page X using the input distance measurements. Using the distance D and the angle A between any other point Y on the page and X, the distance between X and Y can be computed using basic geometry, and also the distance between any two points on the page. The reading machine computes 285 the distance D from the camera to a point on the page X using the input distance measurements. Returning to FIG. 12, the reading machine computes 286 the distances of the detected edges. The reading machine uses the measured distances of the detected edges and the data on standard sizes of pages to determine 287 whether part of a page is missing.
  • Object re-sizing One of the difficulties in dealing with real-world information is that the object in question can appear as a small part of an image or as a dominant element of an image . To deal with this, the image is processed at different levels of pixel resolution. For example, consider text processing. Text can occur in an object in variety of font sizes. For example, commercially available OCR software packages will recognize text in a digitized image if it is approximately 20 to 170 pixels in height . Referring to FIG. 13, an object re-sizing process 300 that re-sizes text to allow successful OCR is shown. The process receives 302 an image and decides 304 if the text is too large or small for OCR. The Hough transform, described above, can provide an estimate of text size.
  • Image Stitching Referring to FIG. 14, an image stitching process 340 is shown.
  • the reading machine 10 stitches multiple images together to allow larger scenes to be read.
  • Image stitching is used in other contexts, such as producing a panorama from several separate images that have some overlap.
  • the stitching attempts to transform two or more images to a common image.
  • the reading machine may allow the user to take several pictures of a scene and may piece together the scene using mathematical stitching. Because the visually impaired person is not as able to control the amount of scene overlap that exists between the individual images, the portable reading machine may need to implement more sophisticated stitching algorithms.
  • gesturing processing In gesturing processing, the user makes a gesture (e.g. with the user's hand) and the reading machine 10 captures the gesture and interprets the gesture as a command.
  • the reading machine may capture the motion of a user's hand, or other pointing device, with a video camera, using high frame rates to capture the motion, and low resolution images to allow faster data transfer and processing.
  • a gesture could also be captured by using a stylus on a touch screen, e.g., circling the area of the image on the screen that the user wishes to be read.
  • Another option is to apply sensors to the user's hand or other body part, such as accelerometers or position sensors. Referring to FIG.
  • the portable reading machine 10 allows the user to select and specify a feature to find in the scene (e.g. stairs, exit, specific street sign or door number) .
  • One method to achieve this is through speech input. For example, if the user is in a building and looking for an exit, the user may simply speak "find exit” to direct the portable reading machine to look for an item that corresponds to an "exit sign" in the scene and announce the location to the user.
  • the usefulness of the portable reading machine 10 in helping the user navigate the physical environment can be augmented in several ways. For instance, the portable reading machine 10 will store in a knowledge base a layout of the relevant building or environment. Having this information, the portable reading machine 10 correlates features that it detects in the images to features in its knowledge base. By detecting the features, the portable reading machine 10 helps the user identify his/her location or provide information on the location of exits, elevators, rest rooms, etc.
  • the portable reading machine may incorporate the functionality of a compass to help orient the user and help in navigation.
  • the portable reading machine 10 may give the user feedback if the conditions for accurate reading are not present. For example, the portable reading machine 10 determines 442 lighting conditions in a captured image or set of images. The reading machine 10 determines lighting conditions by examining contrast characteristics of different parts of the image. Such regional contrast of an image is computed by examining a distribution of light intensities across a captured image. Regions of the captured image that have poor contrast will be characterized by a relatively narrow distribution of light intensity values compared to regions of good contrast . Poor contrast may be present due to lighting that is too dim or too bright.
  • a knowledge base in the reading machine 10 stores information about a particular function of the reading machine 10, such as a mode (e.g. document mode or clothing mode), or a type of hardware (e.g. a camera and its settings), or image processing algorithms.
  • the knowledge base is a collection of reference data, templates, formulas and rules that are used by the portable reader.
  • the data in a knowledge base (or set of knowledge bases) together with algorithms in the reading machine 10 are used to carry out a particular function in the reading machine 10.
  • a knowledge base for document mode could include all the document templates (as previously discussed) , the rules for using the different templates, and a model of document processing.
  • a knowledge base for using an ATM would include all the templates for each screen, plus the rules and other knowledge needed for handling ATMs.
  • the knowledge bases may be hierarchical. For example, one knowledge base helps the reader device determine the most appropriate knowledge base to use to process an image.
  • a model describes an organization of data and procedures that model (or produce a simplified imitation of) some process.
  • a model provides a framework for dealing with the process.
  • a model ties together the necessary knowledge bases, rules, procedures, templates and so on, into a framework for dealing with the mode or interaction or process.
  • the reading machine 10 In document mode, the reading machine 10 has a model of how to read a document to the user.
  • a document speed-reading model may collect together rules that read only the section title and first paragraph from each section, and skip the reading of page numbers, whereas other document reading models may collect different reading rules.
  • the model may be stored in a knowledge base, or the software for the model processing may be implicit in the software of the reading machine 10.
  • a model may be used to help stitch together the content from multiple images with a common theme or context.
  • the device may be set up to skip reading of the title on the top of each page, or to suppress reading the page numbers when reading to the user.
  • Voice recognition for finding stored materials Images or OCR-processed documents may be stored for later recall .
  • a voice note or file name may be specified for the document.
  • the system may allow an interactive search for the stored files based on the stored voice note or on the title or contents of the document.
  • the user can specify the file name, or may specify the keywords.
  • the system specifies how many candidate files were found and may read their names and/or attached voice notes to the user.
  • Process Flow Overview Referring to FIG. 22, an example 500 of the process flow of a document mode is shown. The templates, layout models, and rules that support the mode are retrieved from a Mode ?Knowledge base 501.
  • the section of the image may undergo additional processing 508 prior to producing a binary or grayscale image for OCR.
  • additional processing includes text angle measurement or refinement and contrast/brightness enhancement using filters chosen based on the size of the text lines.
  • the image region is "read" 510 using OCR.
  • the region may also look for patterns that correspond to logos, marks or special symbols.
  • the OCR is assessed 512 by quality measures from the OCR module and by the match of the words against a dictionary, and grammar rules.
  • the reading machine determines if the text detection was satisfactory. If the text detection quality is satisfactory, the device starts reading 514 to the user using text-to-speech (TTS) software.
  • TTS text-to-speech
  • the reading to the user can incorporate auditory cues that indicate transitions such as font changes and paragraph or column transitions.
  • the device Because of the device's ability to provide quick feedback to the user about the quality of the OCR attempt, including specific feedback such as lighting, text being cut off, and text being too large or too small, the device has an advantage for those situations where access time to the text is limited. In other embodiments, the device can automatically translate the text into another language, and either speak the translation or display the translated text. Thus, it will be understood that various modifications may be made without departing from the spirit and scope of the invention. A ccordingly, other embodiments are within the scope of the following claims.

Abstract

L'invention concerne un dispositif de lecture, qui comprend un ordinateur et un dispositif de saisie d'images couplé à l'ordinateur pour capturer des images à basse résolution et des images à haute résolution. Le dispositif de lecture contient également un progiciel résidant dans un support lisible par ordinateur. Le support communique avec l'ordinateur et contient des instructions qui lui permettent de fonctionner dans une pluralité de modes pour optimiser la performance pour des utilisations spécifiques du dispositif de lecture, et pour traiter des images à basse résolution et des images à haute résolution quand il fonctionne dans au moins un desdits modes.
PCT/US2005/011180 2004-04-02 2005-04-04 Dispositif de lecture portatif avec traitement modal WO2005096760A2 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
EP05733206A EP1756802A2 (fr) 2004-04-02 2005-04-04 Dispositif de lecture portatif avec traitement modal

Applications Claiming Priority (28)

Application Number Priority Date Filing Date Title
US55909104P 2004-04-02 2004-04-02
US60/559,091 2004-04-02
US11/097,956 US7659915B2 (en) 2004-04-02 2005-04-01 Portable reading device with mode processing
US11/097,447 2005-04-01
US11/096,711 US7325735B2 (en) 2004-04-02 2005-04-01 Directed reading mode for portable reading machine
US11/096,703 US8249309B2 (en) 2004-04-02 2005-04-01 Image evaluation for reading mode in a reading machine
US11/097,958 2005-04-01
US11/096,712 US7627142B2 (en) 2004-04-02 2005-04-01 Gesture processing with low resolution images with high resolution processing for optical character recognition for a reading machine
US11/096,714 US7641108B2 (en) 2004-04-02 2005-04-01 Device and method to assist user in conducting a transaction with a machine
US11/097,978 2005-04-01
US11/096,711 2005-04-01
US11/097,956 2005-04-01
US11/097,448 US7629989B2 (en) 2004-04-02 2005-04-01 Reducing processing latency in optical character recognition for portable reading machine
US11/097,944 US7505056B2 (en) 2004-04-02 2005-04-01 Mode processing in portable reading machine
US11/096,703 2005-04-01
US11/097,094 2005-04-01
US11/097,986 US8873890B2 (en) 2004-04-02 2005-04-01 Image resizing for optical character recognition in portable reading machine
US11/096,714 2005-04-01
US11/097,447 US9236043B2 (en) 2004-04-02 2005-04-01 Document mode processing for portable reading machine enabling document navigation
US11/097,944 2005-04-01
US11/097,978 US8320708B2 (en) 2004-04-02 2005-04-01 Tilt adjustment for optical character recognition in portable reading machine
US11/097,986 2005-04-01
US11/097,958 US8036895B2 (en) 2004-04-02 2005-04-01 Cooperative processing for portable reading machine
US11/096,712 2005-04-01
US11/097,959 US20060020486A1 (en) 2004-04-02 2005-04-01 Machine and method to assist user in selecting clothing
US11/097,959 2005-04-01
US11/097,448 2005-04-01
US11/097,094 US7840033B2 (en) 2004-04-02 2005-04-01 Text stitching from multiple images

Publications (2)

Publication Number Publication Date
WO2005096760A2 true WO2005096760A2 (fr) 2005-10-20
WO2005096760A3 WO2005096760A3 (fr) 2006-02-16

Family

ID=35125540

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2005/011180 WO2005096760A2 (fr) 2004-04-02 2005-04-04 Dispositif de lecture portatif avec traitement modal

Country Status (2)

Country Link
EP (1) EP1756802A2 (fr)
WO (1) WO2005096760A2 (fr)

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2007144705A1 (fr) * 2006-06-09 2007-12-21 Sony Ericsson Mobile Communications Ab Identification de contenu multimédia
WO2008096192A1 (fr) * 2007-02-06 2008-08-14 Alpnacom Procédé et dispositifs de traitement d'informations se présentant sous forme de texte écrit au moyen d'un appareil ocr mobile
EP2065871A1 (fr) 2007-11-30 2009-06-03 beyo GmbH Dispositif de lecture pour aveugles ou personnes dont la vue est altérée
WO2011080361A1 (fr) * 2009-12-30 2011-07-07 Fundacion Robotiker Procédé, dispositif et système de lecture d'écrans
CN101558416B (zh) * 2006-11-20 2012-07-11 微软公司 移动通信设备的文本检测
CN102778207A (zh) * 2012-07-10 2012-11-14 中联重科股份有限公司 一种结构件应力应变的测量方法、装置及系统
US8913138B2 (en) 2012-12-21 2014-12-16 Technologies Humanware Inc. Handheld magnification device with a two-camera module
US9298661B2 (en) 2012-12-21 2016-03-29 Technologies Humanware Inc. Docking assembly with a reciprocally movable handle for docking a handheld device
US9596388B2 (en) 2008-07-07 2017-03-14 Gopro, Inc. Camera housing with integrated expansion module
CN107979383A (zh) * 2016-10-21 2018-05-01 上海闻通信息科技有限公司 一种阅读方法、系统以及一种阅读机
USD894256S1 (en) 2018-08-31 2020-08-25 Gopro, Inc. Camera mount
CN111698384A (zh) * 2020-06-22 2020-09-22 上海肇观电子科技有限公司 图像处理设备
USD905786S1 (en) 2018-08-31 2020-12-22 Gopro, Inc. Camera mount
US10928711B2 (en) 2018-08-07 2021-02-23 Gopro, Inc. Camera and camera mount
CN113408519A (zh) * 2021-07-09 2021-09-17 深圳市铁越电气有限公司 一种基于模板旋转匹配的指针仪表读数的方法及系统
USD991318S1 (en) 2020-08-14 2023-07-04 Gopro, Inc. Camera
USD997232S1 (en) 2019-09-17 2023-08-29 Gopro, Inc. Camera

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020188772A1 (en) * 2001-04-02 2002-12-12 Mark Radcliffe Media production methods and systems
US6510177B1 (en) * 2000-03-24 2003-01-21 Microsoft Corporation System and method for layered video coding enhancement
US20030095688A1 (en) * 2001-10-30 2003-05-22 Kirmuss Charles Bruno Mobile motor vehicle identification
US20030234772A1 (en) * 2002-06-19 2003-12-25 Zhengyou Zhang System and method for whiteboard and audio capture

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6510177B1 (en) * 2000-03-24 2003-01-21 Microsoft Corporation System and method for layered video coding enhancement
US20020188772A1 (en) * 2001-04-02 2002-12-12 Mark Radcliffe Media production methods and systems
US20030095688A1 (en) * 2001-10-30 2003-05-22 Kirmuss Charles Bruno Mobile motor vehicle identification
US20030234772A1 (en) * 2002-06-19 2003-12-25 Zhengyou Zhang System and method for whiteboard and audio capture

Cited By (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7787697B2 (en) 2006-06-09 2010-08-31 Sony Ericsson Mobile Communications Ab Identification of an object in media and of related media objects
WO2007144705A1 (fr) * 2006-06-09 2007-12-21 Sony Ericsson Mobile Communications Ab Identification de contenu multimédia
US8165409B2 (en) 2006-06-09 2012-04-24 Sony Mobile Communications Ab Mobile device identification of media objects using audio and image recognition
CN101558416B (zh) * 2006-11-20 2012-07-11 微软公司 移动通信设备的文本检测
WO2008096192A1 (fr) * 2007-02-06 2008-08-14 Alpnacom Procédé et dispositifs de traitement d'informations se présentant sous forme de texte écrit au moyen d'un appareil ocr mobile
EP2065871A1 (fr) 2007-11-30 2009-06-03 beyo GmbH Dispositif de lecture pour aveugles ou personnes dont la vue est altérée
EP2096614A1 (fr) 2007-11-30 2009-09-02 beyo GmbH Dispositif de lecture pour personnes aveugles ou malvoyantes
US10356291B2 (en) 2008-07-07 2019-07-16 Gopro, Inc. Camera housing with integrated expansion module
US9596388B2 (en) 2008-07-07 2017-03-14 Gopro, Inc. Camera housing with integrated expansion module
US9699360B2 (en) 2008-07-07 2017-07-04 Gopro, Inc. Camera housing with integrated expansion module
US11025802B2 (en) 2008-07-07 2021-06-01 Gopro, Inc. Camera housing with expansion module
US10986253B2 (en) 2008-07-07 2021-04-20 Gopro, Inc. Camera housing with expansion module
WO2011080361A1 (fr) * 2009-12-30 2011-07-07 Fundacion Robotiker Procédé, dispositif et système de lecture d'écrans
CN102778207A (zh) * 2012-07-10 2012-11-14 中联重科股份有限公司 一种结构件应力应变的测量方法、装置及系统
US8913138B2 (en) 2012-12-21 2014-12-16 Technologies Humanware Inc. Handheld magnification device with a two-camera module
US9298661B2 (en) 2012-12-21 2016-03-29 Technologies Humanware Inc. Docking assembly with a reciprocally movable handle for docking a handheld device
CN107979383A (zh) * 2016-10-21 2018-05-01 上海闻通信息科技有限公司 一种阅读方法、系统以及一种阅读机
US10928711B2 (en) 2018-08-07 2021-02-23 Gopro, Inc. Camera and camera mount
US11662651B2 (en) 2018-08-07 2023-05-30 Gopro, Inc. Camera and camera mount
USD905786S1 (en) 2018-08-31 2020-12-22 Gopro, Inc. Camera mount
USD894256S1 (en) 2018-08-31 2020-08-25 Gopro, Inc. Camera mount
USD989165S1 (en) 2018-08-31 2023-06-13 Gopro, Inc. Camera mount
USD1023115S1 (en) 2018-08-31 2024-04-16 Gopro, Inc. Camera mount
USD997232S1 (en) 2019-09-17 2023-08-29 Gopro, Inc. Camera
USD1024165S1 (en) 2019-09-17 2024-04-23 Gopro, Inc. Camera
CN111698384A (zh) * 2020-06-22 2020-09-22 上海肇观电子科技有限公司 图像处理设备
USD991318S1 (en) 2020-08-14 2023-07-04 Gopro, Inc. Camera
USD1004676S1 (en) 2020-08-14 2023-11-14 Gopro, Inc. Camera
CN113408519A (zh) * 2021-07-09 2021-09-17 深圳市铁越电气有限公司 一种基于模板旋转匹配的指针仪表读数的方法及系统
CN113408519B (zh) * 2021-07-09 2023-07-18 深圳市铁越电气有限公司 一种基于模板旋转匹配的指针仪表读数的方法及系统

Also Published As

Publication number Publication date
EP1756802A2 (fr) 2007-02-28
WO2005096760A3 (fr) 2006-02-16

Similar Documents

Publication Publication Date Title
US10741167B2 (en) Document mode processing for portable reading machine enabling document navigation
US9626000B2 (en) Image resizing for optical character recognition in portable reading machine
US7659915B2 (en) Portable reading device with mode processing
US7627142B2 (en) Gesture processing with low resolution images with high resolution processing for optical character recognition for a reading machine
US7505056B2 (en) Mode processing in portable reading machine
US7629989B2 (en) Reducing processing latency in optical character recognition for portable reading machine
US8284999B2 (en) Text stitching from multiple images
US8320708B2 (en) Tilt adjustment for optical character recognition in portable reading machine
US8626512B2 (en) Cooperative processing for portable reading machine
US7325735B2 (en) Directed reading mode for portable reading machine
US8186581B2 (en) Device and method to assist user in conducting a transaction with a machine
US8249309B2 (en) Image evaluation for reading mode in a reading machine
US20150043822A1 (en) Machine And Method To Assist User In Selecting Clothing
US20160344860A1 (en) Document and image processing
EP1756802A2 (fr) Dispositif de lecture portatif avec traitement modal
EP2065871A1 (fr) Dispositif de lecture pour aveugles ou personnes dont la vue est altérée
US20090186321A1 (en) Reading Device for Blind or Visually Impaired Persons

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

WWE Wipo information: entry into national phase

Ref document number: 2005733206

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: DE

WWW Wipo information: withdrawn in national office

Country of ref document: DE

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWP Wipo information: published in national office

Ref document number: 2005733206

Country of ref document: EP