US20140168076A1 - Touch sensitive device with concentration mode - Google Patents

Touch sensitive device with concentration mode Download PDF

Info

Publication number
US20140168076A1
US20140168076A1 US13/715,221 US201213715221A US2014168076A1 US 20140168076 A1 US20140168076 A1 US 20140168076A1 US 201213715221 A US201213715221 A US 201213715221A US 2014168076 A1 US2014168076 A1 US 2014168076A1
Authority
US
United States
Prior art keywords
user
reading guide
guide feature
reading
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US13/715,221
Other versions
US8963865B2 (en
Inventor
Kourtny M. Hicks
Dale J. Brewer
Amir Mesguich Havilio
Gerald B. Cueto
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Barnes and Noble College Booksellers LLC
Original Assignee
Nook Digital LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nook Digital LLC filed Critical Nook Digital LLC
Priority to US13/715,221 priority Critical patent/US8963865B2/en
Assigned to BARNESANDNOBLE.COM LLC reassignment BARNESANDNOBLE.COM LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BREWER, DALE J., CUETO, GERALD B., HARGREAVES, ANDREW, HAVILIO, AMIR MESGUICH, HICKS, KOURTNY M.
Publication of US20140168076A1 publication Critical patent/US20140168076A1/en
Application granted granted Critical
Publication of US8963865B2 publication Critical patent/US8963865B2/en
Assigned to NOOK DIGITAL LLC reassignment NOOK DIGITAL LLC CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: BARNESANDNOBLE.COM LLC
Assigned to NOOK DIGITAL, LLC reassignment NOOK DIGITAL, LLC CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: NOOK DIGITAL LLC
Assigned to BARNES & NOBLE COLLEGE BOOKSELLERS, LLC reassignment BARNES & NOBLE COLLEGE BOOKSELLERS, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: NOOK DIGITAL, LLC
Assigned to NOOK DIGITAL, LLC reassignment NOOK DIGITAL, LLC CORRECTIVE ASSIGNMENT TO REMOVE APPLICATION NUMBERS 13924129 AND 13924362 PREVIOUSLY RECORDED ON REEL 035187 FRAME 0476. ASSIGNOR(S) HEREBY CONFIRMS THE CHANGE OF NAME. Assignors: NOOK DIGITAL LLC
Assigned to NOOK DIGITAL LLC reassignment NOOK DIGITAL LLC CORRECTIVE ASSIGNMENT TO REMOVE APPLICATION NUMBERS 13924129 AND 13924362 PREVIOUSLY RECORDED ON REEL 035187 FRAME 0469. ASSIGNOR(S) HEREBY CONFIRMS THE CHANGE OF NAME. Assignors: BARNESANDNOBLE.COM LLC
Assigned to BANK OF AMERICA, N.A., AS COLLATERAL AGENT reassignment BANK OF AMERICA, N.A., AS COLLATERAL AGENT SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BARNES & NOBLE COLLEGE BOOKSELELRS, LLC
Assigned to TOPLIDS LENDCO, LLC AS COLLATERAL AGENT reassignment TOPLIDS LENDCO, LLC AS COLLATERAL AGENT SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BARNES & NOBLE COLLEGE BOOKSELLERS, LLC, BNED LOUDCLOUD, LLC
Assigned to BANK OF AMERICA, N.A., AS COLLATERAL AGENT reassignment BANK OF AMERICA, N.A., AS COLLATERAL AGENT CORRECTIVE ASSIGNMENT TO CORRECT THE REPLACE THE GRANT OF SECURITY INTEREST IN PATENTS AND SCHEDULE I ATTACHED TO FILING PREVIOUSLY RECORDED AT REEL: 056784 FRAME: 0426. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY INTEREST. Assignors: BARNES & NOBLE COLLEGE BOOKSELLERS, LLC
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S345/00Computer graphics processing and selective visual display systems
    • Y10S345/901Electronic book with display

Definitions

  • This disclosure relates to electronic display devices, and more particularly, to user interface techniques for interacting with touch sensitive devices.
  • Electronic display devices such as tablets, eReaders, mobile phones, smart phones, personal digital assistants (PDAs), and other such touch screen electronic display devices are commonly used for displaying consumable content.
  • the content may be, for example, an eBook, an online article or blog, images, a movie or video, a map, just to name a few types.
  • Such display devices are also useful for displaying a user interface that allows a user to interact with an application running on the device.
  • the user interface may include, for example, one or more touch screen controls and/or one or more displayed labels that correspond to nearby hardware buttons.
  • the touch screen display may be backlit or not, and may be implemented for instance with an LED screen or an electrophoretic display.
  • Such devices may also include other touch sensitive surfaces, such as a track pad (e.g., capacitive or resistive touch sensor) or touch sensitive housing (e.g., acoustic sensor).
  • FIGS. 1 a - b illustrate an example electronic touch sensitive device having a concentration mode configured in accordance with an embodiment of the present invention.
  • FIGS. 1 c - d illustrate example configuration screen shots of the user interface of the electronic touch sensitive device shown in FIGS. 1 a - b , configured in accordance with an embodiment of the present invention.
  • FIG. 2 a illustrates a block diagram of an electronic touch sensitive device configured in accordance with an embodiment of the present invention.
  • FIG. 2 b illustrates a block diagram of a communication system including the electronic touch sensitive device of FIG. 2 b , configured in accordance with an embodiment of the present invention.
  • FIGS. 3 a - b illustrate an example user interface for engaging a concentration mode of an electronic touch sensitive device configured in accordance with an embodiment of the present invention.
  • FIGS. 4 a - f each illustrates an example user interface screen shot showing an electronic touch sensitive device in a concentration mode, in accordance with an embodiment of the present invention.
  • FIG. 5 illustrates a method for providing a concentration mode in an electronic touch sensitive device, in accordance with an embodiment of the present invention.
  • FIG. 6 illustrates a method for providing a concentration mode in an electronic touch sensitive device, in accordance with another embodiment of the present invention.
  • the concentration mode can be used to highlight given textual content on a line by line basis, or other consumable content (e.g., line of photos or graphics, etc).
  • the user can engage the mode, for instance, with a particular touch screen/pad gesture.
  • a user interface (UI) feature is displayed that assists the user in viewing the given content.
  • the UI feature includes a reading pane that encompasses one to three complete lines of displayed text or other content. Background content outside the reading pane can be faded or otherwise softened, if so desired.
  • the UI feature also includes, or alternatively includes, a straight-edged reading guide.
  • the initial placement of the UI feature can be set, for instance, based on the initial contact point touched by the user, or at some predetermined location. In some cases, as the user moves the contact point, the UI feature moves accordingly; in other cases, once engaged, the UI feature advances automatically at a pace suitable to the user's consumption speed. Concentration mode may be paused or disengaged, for example, based on a further specific user contact or release of contact point.
  • electronic display devices such as tablets, eReaders, and smart phones are commonly used for displaying user interfaces and consumable content.
  • the user of such devices can typically read the displayed content with relative ease.
  • the user might prefer a viewing guide to enhance the consumption experience. For instance, some people like to read physical books with a ruler or straight edge to accentuate the current line being read.
  • some people prefer to read with a magnifying glass or large print books. While some electronic devices are provided with large print or have a magnifying glass feature, there does not appear to be a user interface reading guide that can be intuitively engaged and disengaged by the reader.
  • the concentration mode can be used, for example, to highlight given textual content on a line by line basis, or other consumable content.
  • the mode can be engaged globally, so as to allow for use in any number of different applications available on a given device, such as a tablet or mobile phone.
  • the concentration mode may be used in an eBook application, a browser application, an email application, a word processing application, a PDF viewer application, or any other application where it might be useful to the user to have a reading guide as described herein.
  • the user can engage the mode with a particular gesture that uniquely and globally indicates a desired intent to engage the mode, such as a touch or press-and-hold at the edge of the device touch screen or track pad, or on the device home button (or other such physical feature), or a tap or sequence of taps on a housing configured with acoustics sensors.
  • a desired intent to engage the mode such as a touch or press-and-hold at the edge of the device touch screen or track pad, or on the device home button (or other such physical feature), or a tap or sequence of taps on a housing configured with acoustics sensors.
  • a particular gesture that uniquely and globally indicates a desired intent to engage the mode, such as a touch or press-and-hold at the edge of the device touch screen or track pad, or on the device home button (or other such physical feature), or a tap or sequence of taps on a housing configured with acoustics sensors.
  • any uniquely identifiable touch of a touch sensitive surface by the user can be used to engage the concentration mode.
  • the uniqueness of the touch may be based, for example, on the particular location of the touch and/or the nature of the touch (e.g., a press-and-hold for a predetermined duration, the number of simultaneously applied contact points, a sequence of taps, drags, swipes, circular or elliptical gestures, etc).
  • a UI feature is displayed to the user that assists the user in focusing on the given content.
  • the UI feature includes a reading pane that encompasses one to three complete lines of displayed text or other content to be focused on.
  • a ‘complete line’ refers to, for example, a line of content that spans from the farthest left piece of content displayed on a given line to the farthest right piece of content displayed on that same line.
  • the line of content may be a string of words, or one or more graphics provisioned along a line or otherwise spanning from the farthest left edge to the farthest right edge, or a string of Magnification can be provided if so desired, and background content outside the reading pane can be faded or otherwise softened (e.g., blacked-out, blurred, etc) from the user's view so as to further accentuate the content within the reading pane.
  • the UI feature may also include, or alternatively include, a virtual ruler or other such straight-edged reading guide.
  • the initial point of focus provided by the UI feature can be set, for instance, based on the initial contact point (e.g., finger or stylus) designated by the user, or at a predetermined location or at the beginning of the text/content for a given page.
  • the contact point or points
  • the provisioned UI feature moves accordingly, thus allowing the user direct control over advancement of the reading guide UI feature.
  • the user may continue to maintain the engagement contact points while the UI feature moves in unison with the user's hand movement.
  • the user can keep just one of the engagement contact points in place while the UI feature moves in unison with the user dragging of that point.
  • the user may move the contact point back up a page (rather than down the page), for example, to review text previously read, if so desired. Once the review is done, the user can then speed-drag the contact point to the desired place and continue with consuming the content in the concentration mode.
  • numerous dragging schemes can be used to effectively advance the highlighted content region accordingly, in accordance with an embodiment. If no contact points remain, the concentration mode will disengage, in some embodiments.
  • the displayed UI feature will remain engaged and advance automatically at a predetermined speed (e.g., user-configurable speed) starting at the top of each page (or beginning of content for that page) and steadily progressing to the bottom of each page (or end of content for that page).
  • a predetermined speed e.g., user-configurable speed
  • the user can pause the auto-advance function, for instance, by touching a pause button UI feature provisioned somewhere on the touch screen display.
  • the auto-advance function will switch to a paused mode and the user may freely move the UI feature based on dragging contact with the touch screen or track pad.
  • the auto-advance mode can be re-started by, for example, touching a play button UI feature provisioned somewhere on the touch screen display (e.g., interchangeably located with pause button).
  • the user can exit or otherwise terminate the auto-advance concentration mode, for example, by selecting an end mode UI feature provisioned on the touch screen, or by depressing the home button or some other physical button of the device that can be purposed to terminate a given operation mode. Numerous termination schemes will be apparent in light of this disclosure.
  • Additional functions and features can be added to further enhance the user's experience. For example, once the reading guide is deployed, the user may touch the content within (and/or at the edge of) the reading guide area so as to further cause that text to be accentuated. For instance, some people like to read with their finger tip.
  • the user may touch the content within (and/or at the edge of) the reading guide area so as to further cause that text to be accentuated. For instance, some people like to read with their finger tip.
  • another UI feature is activated to highlight the word at the user's fingertip, in accordance with an embodiment.
  • the highlighting may be, for instance, bolding, underlining, magnifying, or a change in color. As the user moves her fingertip down the sentence, the word highlighting advances accordingly to follow the finger.
  • the user may resize the reading pane once it is deployed on screen so that, for example, additional lines of text can be viewed (e.g., change from one to three lines of text).
  • a highlighting feature as just described can also be provisioned.
  • the highlighting feature may include several words at a time (e.g., 2 to 5 words), if so desired, and in accordance with an embodiment of the present invention.
  • Such optional features can be user-configured or hard-coded, as will be appreciated.
  • a two-finger press-and-hold or other touch gesture along the border region of the display screen can be used to automatically create the UI feature (e.g., pane or straight edge) to accentuate one complete line of text at a time.
  • the two fingers may be, for example, the user's pointer and middle fingers (or the thumb and pointer finger) of one hand on one side of the display, or the user's thumbs on each side of the display.
  • a stylus could be used in combination with one or more fingers.
  • the left thumb could be placed at the left-side border of the display and the stylus tip at the opposing right-side border.
  • a stylus could be used on its own to engage the mode, for example, by laying the stylus on (or otherwise sufficiently near) the touch sensitive surface so as to effectively create multiple contact points that can be identified as a request for the concentration mode.
  • the use of two (or more) contact points can be used to uniquely identify the user's desire to engage the concentration mode.
  • Numerous uniquely identifiable engagement schemes that exploit a touch sensitive surface can be used as will be appreciated in light of this disclosure.
  • any touch sensitive device e.g., track pad, touch screen, or other touch sensitive surface, whether capacitive, resistive, acoustic or other touch detecting technology
  • touch sensitive device e.g., track pad, touch screen, or other touch sensitive surface, whether capacitive, resistive, acoustic or other touch detecting technology
  • the claimed invention is not intended to be limited to any particular type of touch sensitive technology, unless expressly stated.
  • FIGS. 1 a - b illustrate an example electronic touch sensitive device having a concentration mode configured in accordance with an embodiment of the present invention.
  • the device could be, for example, a tablet such as the NOOK® Tablet by Barnes & Noble.
  • the device may be any electronic device having a touch sensitive user interface and capability for displaying content to a user, such as a mobile phone or mobile computing device such as a eReader or laptop, a desktop computing system, a television, a smart display screen, or any other device having a touch screen display or a non-touch display screen that can be used in conjunction with a touch sensitive surface.
  • the claimed invention is not intended to be limited to any particular kind or type of electronic device.
  • the device comprises a housing that includes a number of hardware features such as a power button, control features, and a press-button (sometimes called a home button herein).
  • a touch screen based user interface is also provided, which in this example embodiment includes a quick navigation menu having six main categories to choose from (Home, Library, Shop, Search, Light, and Settings) and a status bar that includes a number of icons (a night-light icon, a wireless network icon, and a book icon), a battery indicator, and a clock.
  • Other embodiments may have fewer or additional such UI touch screen controls and features, or different UI touch screen controls and features altogether, depending on the target application of the device. Any such general UI controls and features can be implemented using any suitable conventional or custom technology, as will be appreciated.
  • the hardware control features provided on the device housing in this example embodiment are configured as elongated press-bars and can be used, for example, to page forward (using the top press-bar) or to page backward (using the bottom press-bar), such as might be useful in an eReader application.
  • the power button can be used to turn the device on and off, and may be used in conjunction with a touch-based UI control feature that allows the user to confirm a given power transition action request (e.g., such as a slide bar or tap point graphic to turn power off).
  • a touch-based UI control feature that allows the user to confirm a given power transition action request (e.g., such as a slide bar or tap point graphic to turn power off).
  • the home button is a physical press-button that can be used as follows: when the device is awake and in use, tapping the button will display the quick navigation menu, which is a toolbar that provides quick access to various features of the device.
  • the home button may also be configured to cease an active function that is currently executing on the device, such as an auto-advance reading pane UI feature as described herein.
  • the button may further control other functionality if, for example, the user presses and holds the home button. For instance, an example such push-and-hold function could engage a power conservation routine where the device is put to sleep or an otherwise lower power consumption mode. So, a user could grab the device by the button, press and keep holding as the device was stowed into a bag or purse.
  • the home button may be associated with and control different and unrelated actions: 1) show the quick navigation menu; 2) exit the auto-advance reading pane UI feature mode, but keep the page being read or otherwise consumed displayed; and 3) put the device to sleep.
  • the status bar may also include a book icon (upper left corner).
  • the user can access a sub-menu that provides access to a concentration mode configuration sub-menu by tapping the book icon of the status bar. For example, upon receiving an indication that the user has touched the book icon, the device can then display the concentration mode configuration sub-menu shown in FIG. 1 d . In other cases, tapping the book icon just provides bibliographic information on the content being consumed.
  • Another example way for the user to access a concentration mode configuration sub-menu such as the one shown in FIG. 1 d is to tap or otherwise touch the Settings option in the quick navigation menu, which causes the device to display the general sub-menu shown in FIG. 1 c .
  • the user can select any one of a number of options, including one designated Screen/UI in this specific example case. Selecting this sub-menu item (with an appropriately placed screen tap) may cause the concentration mode configuration sub-menu of FIG. 1 d to be displayed, in accordance with an embodiment.
  • selecting the Screen/UI option may present the user with a number of additional sub-options, one of which may include a so-called concentration mode option, which may then be selected by the user so as to cause the concentration mode configuration sub-menu of FIG. 1 d to be displayed. Any number of such menu schemes and nested hierarchies can be used, as will be appreciated in light of this disclosure.
  • UI touch screen controls are implemented as UI touch screen controls in this example embodiment.
  • Such UI touch screen controls can be programmed or otherwise configured using any number of conventional or custom technologies.
  • the touch screen translates the user touch in a given location into an electrical signal which is then received and processed by the underlying operating system (OS) and circuitry (processor, etc). Additional example details of the underlying OS and circuitry in accordance with some embodiments will be discussed in turn with reference to FIG. 2 a.
  • the concentration mode configuration sub-menu shown in FIG. 1 d can be provided to the user.
  • the concentration mode configuration sub-menu includes a UI check box that when checked or otherwise selected by the user, effectively enables the concentration mode (shown in the enabled state); unchecking the box disables the mode.
  • Other embodiments may have the concentration mode always enabled, or enabled by a physical switch or button, for example. The user can configure a number of features with respect to the concentration mode, in this example embodiment.
  • the user can select a reading pane based concentration mode which allows a reading pane to be displayed during concentration mode.
  • the reading pane effectively encompasses or otherwise highlights one line of text at a time, in some example cases.
  • the user may also check the Borderless check box to provide a borderless reading pane and/or the Magnify check box to provide a reading pane having a desired degree of magnification (e.g., 1.5 ⁇ , 2 ⁇ , 3 ⁇ , which may be fixed or user configurable), if so desired.
  • the Ruler check box the user can select a ruler-based concentration mode which allows a virtual ruler or straight-edge to be displayed during concentration mode.
  • the ruler is deployed below the line of text to be read by default, but the user can further check the Above Text check box to provide the ruler above the line of text to be read, if so desired.
  • both the Ruler and Reading Pane check boxes may be checked to provide a hybrid UI feature to assist the reading experience.
  • the reading pane/ruler UI feature is advanced manually by the user as the default setting, but the user can further check the Auto-advance check box which allows the UI feature to advance automatically at a pace suitable to the user's consumption speed, if so desired.
  • user can further specify the delay time the UI feature dwells on each line of text (or other content being consumed).
  • the user may also set certain background features in this example embodiment.
  • the text/content outside the area of focus provided by the reading pane/rule UI feature is effectively left as-is by default setting, but the user can further check the Content Fade check box which grays out or otherwise softens background content not within the highlighted reading area, if so desired.
  • the user may also select the color scheme for the highlighted reading area and/or background, by selecting the Color UI control and selecting the desired scheme, for example.
  • Also configurable in this example embodiment is the mechanism by which the user engages the concentration mode as well as the activation time.
  • the user can assign the mode engagement function to a hardware switch/feature such as the home button by checking the Button Press check box, or to an area of the touch screen by checking the Edge Touch check box.
  • a hardware switch/feature such as the home button by checking the Button Press check box, or to an area of the touch screen by checking the Edge Touch check box.
  • the duration which the user's press-and-hold gesture (on either the hardware button or the touch screen) to engage the concentration mode is set to 3 seconds in this example case. Other embodiments may not use such a delay and activate upon contact in the activation region.
  • a back button arrow UI control feature is provisioned on the touch screen, so that the user can go back to the Settings menu after the concentration mode has been configured, if so desired. While a number of user-configurable features are provided in some embodiments, other embodiments may have any combination of similar features wherein the features are set as a given default or hard-coded, such that no user configuration is necessary.
  • FIG. 2 a illustrates a block diagram of an electronic touch screen device configured in accordance with an embodiment of the present invention.
  • this example device includes a processor, random access memory (RAM), additional storage/memory, a communications module, a touch screen, and an audio module.
  • RAM random access memory
  • a communications bus and interconnect is also provided to allow inter-device communication.
  • Other typical componentry and functionality not reflected in the block diagram will be apparent (e.g., battery, co-processor, etc).
  • a touch screen display is provided, other embodiments may include a non-touch screen and a touch sensitive surface such as a track pad, or a touch sensitive housing configured with one or more acoustic sensors, etc.
  • the touch sensitive surface is generally capable of translating a user's physical contact with the surface into an electronic signal that can be manipulated or otherwise used to trigger a specific user interface action, such as those provided herein.
  • the principles provided herein equally apply to any such touch sensitive devices. For ease of description, examples are provided with touch screen technology.
  • the RAM includes a number of modules stored therein that can be accessed and executed by the processor (and/or a co-processor).
  • the modules include an operating system (OS), a user interface (UI), and a power conservation routine (Power).
  • OS operating system
  • UI user interface
  • Power power conservation routine
  • the modules can be implemented, for example, in any suitable programming language (e.g., C, C++, objective C, JavaScript, custom or proprietary instruction sets, etc), and encoded on a machine readable medium, that when executed by the processor (and/or co-processors), carries out the functionality of the device including a UI having a concentration mode as described herein.
  • inventions can be implemented, for instance, with gate-level logic or an application-specific integrated circuit (ASIC) or chip set or other such purpose built logic, or a microcontroller having input/output capability (e.g., inputs for receiving user inputs and outputs for directing other components) and a number of embedded routines for carrying out the device functionality.
  • ASIC application-specific integrated circuit
  • the functional modules can be implemented in hardware, software, firmware, or a combination thereof.
  • the processor can be any suitable processor (e.g., 800 MHz Texas Instruments OMAP3621 applications processor), and may include one or more co-processors or controllers to assist in device control.
  • the processor receives input from the user, including input from or otherwise derived from the power button and the home button.
  • the processor can also have a direct connection to a battery so that it can perform base level tasks even during sleep or low power modes.
  • the RAM can be any suitable type of memory and size (e.g., 256 or 512 Mbytes SDRAM), and in other embodiments may be implemented with non-volatile memory or a combination of non-volatile and volatile memory technologies.
  • the storage can also be implemented with any suitable memory and size (e.g., 2 GBytes of flash memory).
  • the display can be implemented, for example, with a 6-inch E-ink Pearl 800 ⁇ 600 pixel screen with Neonode® zForce® touch screen, or any other suitable display and touch screen interface technology.
  • the communications module can be, for instance, any suitable 802.11 b/g/n WLAN chip or chip set, which allows for connection to a local network so that content can be downloaded to the device from a remote location (e.g., content provider, etc, depending on the application of the display device).
  • the device housing that contains all the various componentry measures about 6.5′′ high by about 5′′ wide by about 0.5′′ thick, and weighs about 6.9 ounces. Any number of suitable form factors can be used, depending on the target application (e.g., laptop, desktop, mobile phone, etc).
  • the device may be smaller, for example, for smartphone and tablet applications and larger for smart computer monitor applications.
  • the operating system (OS) module can be implemented with any suitable OS, but in some example embodiments is implemented with Google Android OS or Linux OS or Microsoft OS or Apple OS. As will be appreciated in light of this disclosure, the techniques provided herein can be implemented on any such platforms.
  • the power management (Power) module can be configured as typically done, such as to automatically transition the device to a low power consumption or sleep mode after a period of non-use. A wake-up from that sleep mode can be achieved, for example, by a physical button press and/or a touch screen swipe or other action.
  • the user interface (UI) module can be, for example, based on touch screen technology and the various example screen shots shown in FIGS.
  • the audio module can be configured, for example, to speak or otherwise aurally present a selected eBook or other textual content, if preferred by the user.
  • Numerous commercially available text-to-speech modules can be used, such as Verbose text-to-speech software by NCH Software.
  • storage can be expanded via a microSD card or other suitable memory expansion technology (e.g., 32 GBytes, or higher).
  • FIG. 2 b illustrates a block diagram of a communication system including the electronic touch sensitive device of FIG. 2 b , configured in accordance with an embodiment of the present invention.
  • the system generally includes an electronic touch sensitive device that is capable of communicating with a server via a network/cloud.
  • the electronic touch sensitive device may be, for example, an eBook reader, a mobile cell phone, a laptop, a tablet, desktop, or any other touch sensitive computing device.
  • the network/cloud may be a public and/or private network, such as a private local area network operatively coupled to a wide area network such as the Internet.
  • the server may be programmed or otherwise configured to receive content requests from a user via the touch sensitive device and to respond to those requests by providing the user with requested or otherwise recommended content.
  • the server is configured to remotely provision a reading pane/rule UI feature as provided herein to the touch sensitive device (e.g., via JavaScript or other browser based technology).
  • portions of the methodology are executed on the server and other portions of the methodology are executed on the device. Numerous server-side/client-side execution schemes can be implemented, as will be apparent in light of this disclosure.
  • FIGS. 3 a - b illustrate an example user interface for engaging a concentration mode of an electronic touch sensitive device configured in accordance with an embodiment of the present invention.
  • the touch screen generally includes a content portion (within the dashed line area) and an edge portion.
  • a physical frame or support structure is provided about the screen.
  • the user is using two contact points on opposite edges of the touch screen. Note that the two points do not need to be aligned in the horizontal direction, but rather can each be on any point along the respective edge, in some such embodiments.
  • the user may use any two fingers/thumbs or stylus to create the two (or more) contact points.
  • FIG. 3 b is similar to FIG.
  • any portion of the touch screen display can be used as a concentration mode activation zone.
  • a hardware button may also be used to activate the concentration mode, if so desired. Numerous uniquely identifiable engagement schemes can be used as will be appreciated in light of this disclosure.
  • FIGS. 4 a - f each illustrates a screen shot showing example such UI features, in accordance with an embodiment of the present invention.
  • a borderless reading pane is provided that highlights a single line of text/content to the user, and all other background content is faded out. As the user's hand or stylus moves down the virtual page, so does the reading pane.
  • the example reading pane illustrated in FIG. 4 b is similar to that shown in FIG. 4 a , but is provided with a border and magnification.
  • the user can specify the degree of magnification, in accordance with some embodiments. Further recall that the user may also resize the reading guide, for example, so that additional lines of text can be viewed (e.g., by simultaneously pulling top and bottom edges of the reading pane in opposite directions, or any other such suitable resizing gesture). Further recall that content highlighting within or otherwise at the reading guide can be used if so desired (e.g., reading with fingertip scenario). With respect to the specific example of FIG. 4 c , a virtual ruler or straight-edge reading guide is provided that highlights a single line of text/content to the user. In this example case, the background text is not faded out, but could be if so desired. The user manually controls the placement of the ruler based on dragging of the contact point.
  • a borderless reading pane is provided that highlights a single line of text/content to the user, and the background text is not faded out.
  • the auto-advance function is selected, so the reading pane automatically advances at a rate suitable to the user's consumption speed.
  • the user may pause the auto-advancement of the UI feature by tapping the pause button provided in the lower right corner of the touch screen, in this example case.
  • the pause UI feature changes to a play UI feature that the user can tap to re-start the auto-advancement.
  • the user can at any time manually control the placement of the reading pane by contacting the touch screen over the pane (or proximate to the pane) and dragging to a point of interest in the displayed content.
  • the initial touch by the user causes the auto-advance to pause (and the UI play feature then appears), and when the user is done dragging the reading pane, the UI play feature can then be tapped to resume auto-advancement (e.g., from the point where the reading pane was left by the user, or from the point where auto-advancement was paused).
  • the user can exit auto-advancement, for example, by pressing the home button or some other hardware or software feature provisioned for allowing exit from the auto-advancement routine.
  • the example auto-advancing reading pane illustrated in FIG. 4 e is similar to that shown in FIG. 4 d , but is provided with a border and magnification, and faded background content.
  • FIG. 4 f shows an example of an auto-advancing reading pane and ruler combination, with faded background content. Numerous configurations will be apparent.
  • FIG. 5 illustrates a method for providing a concentration mode in an electronic touch screen device, in accordance with an embodiment of the present invention.
  • This example methodology may be implemented, for instance, by the UI module of the touch sensitive device shown in FIG. 2 a , or the touch sensitive device shown in FIG. 2 b (e.g., with the UI provisioned to the client by the server).
  • the UI can be implemented in software, hardware, firmware, or any combination thereof, as will be appreciated in light of this disclosure.
  • the method includes an activation phase and a maintain/exit phase.
  • Each phase includes sensing a user's input by a touch sensitive surface.
  • any touch sensitive device may be used to detect contact with it by one or more fingers and/or styluses.
  • the UI code and/or hardware
  • the UI code can assume a drag gesture has been engaged and track the path of each contact point with respect to any fixed point within the touch surface until the user stops engaging the touch sensitive surface.
  • the release point can also be captured by the UI as it may be used to commit the action started when the user pressed on the touch sensitive surface.
  • a press or press-and-hold command may be assumed depending on the amount of time the user was continually pressing on the touch sensitive surface.
  • the method includes detecting 501 user contact at the touch sensitive interface.
  • the touch monitoring is effectively continuous.
  • the method continues with determining 503 if the contact is in the concentration mode activation zone.
  • one example activation zone is the edge of the touch sensitive surface.
  • Other embodiments may have the activation zone anywhere on the touch sensitive surface that is specified by the user via a configuration menu.
  • the concentration mode activation zone can be pre-established in any specific location on the touch sensitive surface of the device.
  • the method may continue with reviewing 504 the contact for some other UI request (e.g., select a file, send an email, etc).
  • the method continues with determining 505 if the contact is persistent. For instance, if the contact is a press-and-hold that lasts greater than 3 seconds (or some other designated time period), then the UI code can assume the contact is persistent.
  • the method may continue with reviewing 504 the contact for some other UI request (e.g., select a file, send an email, etc).
  • the method continues with activating 507 the concentration mode, or maintaining that mode along with the current reading guide feature position on the display if already activated.
  • the method continues with determining 509 if the contact on the touch sensitive surface is persistent and moving. If the contact is persistent but not moving (e.g., movement ⁇ 2 mm), then the UI code assumes the user is actively consuming content of the current line and the method continues with maintaining 507 the concentration mode along with the current reading guide feature position on the display. If the contact is not persistent, then the method continues with exiting 512 the concentration mode. However, if the contact is both persistent and moving (e.g., movement >2 mm), then the method continues with moving 511 the reading pane/ruler UI feature to next line or to the otherwise desired area of the content indicated by the contact movement on the touch sensitive surface. As can be further seen, the method includes continuously monitoring the contact point for persistence and movement.
  • FIG. 6 illustrates a method for providing a concentration mode in an electronic touch screen device, in accordance with another embodiment of the present invention.
  • this methodology includes an activation phase and a maintain/exit phase, and the previous relevant discussion is equally applicable here.
  • the methodology employs an auto-advance feature.
  • the activation phase includes detecting 601 user contact at the touch sensitive interface, and determining 603 if that contact indicates that the concentration mode is desired (e.g., based on the unique set of multiple contact points provided by the user). If not, then the method may continue with reviewing 604 the contact for some other UI request (e.g., select a file, send an email, etc). On the other hand, if the contact indicates that the concentration mode is desired, the method continues with activating 605 the concentration mode, or maintaining that mode if already activated and auto-advance the reading pane/ruler UI feature position.
  • the method continues in the maintain/exit phase with determining 607 if the user has selected pause or wishes to exit the concentration mode (this monitoring is continuous in this example embodiment). If the user has selected pause, then the method continues with holding 609 the current reading pane/ruler UI feature position, and the continuous monitoring continues. If the user releases the pause by selecting a play UI feature, then the method continues with maintaining 605 the concentration mode and auto-advancement of the reading pane/ruler UI feature position, and the continuous monitoring continues. If the user signal an exit is desired (e.g., by selecting an exit UI feature or pressing the home button), then the method continues with exiting 611 the concentration mode, and the continuous touch monitoring continues.
  • this monitoring is continuous in this example embodiment. If the user has selected pause, then the method continues with holding 609 the current reading pane/ruler UI feature position, and the continuous monitoring continues. If the user releases the pause by selecting a play UI feature, then the method continues with maintaining 605 the concentration mode and auto-ad
  • one example embodiment provides a device that includes a display for displaying content to a user, and a touch sensitive surface for allowing user input.
  • the device further includes a user interface including a reading guide feature for highlighting a line of the content displayed to a user, the reading guide feature configured to be displayed on the display in response to user contact via the touch sensitive interface, the user contact uniquely identifying the user's desire to use the reading guide feature and including multiple contact points.
  • the display is a touch screen display that includes the touch sensitive surface.
  • the touch sensitive surface is a track pad.
  • the reading guide feature includes at least one of a reading pane and/or a straight-edge.
  • content displayed to the user that is not highlighted by the reading guide feature is faded (e.g., grayed out, blacked-out, blurred, etc).
  • the reading guide feature provides magnification and/or is user-resizable (so as to view more lines of text). In some such cases, the degree of magnification is user-configurable.
  • the reading guide feature advances in response to user dragging contact on the touch sensitive surface. In some cases, the reading guide feature advances automatically. In some cases, the reading guide feature advances automatically at a rate that is user-configurable.
  • Another example embodiment provides a device that includes a display having a touch screen interface and for displaying content to a user, and a user interface including a reading guide feature for highlighting a complete line of the content displayed to a user, the reading guide feature configured to be displayed on the display in response to user contact via the touch screen interface, the user contact uniquely identifying the user's desire to use the reading guide feature and including multiple contact points, wherein the reading guide feature includes at least one of a reading pane and/or a straight-edge.
  • content displayed to the user that is not highlighted by the reading guide feature is faded, and/or content that is highlighted by the reading guide can be selectively amplified (e.g., amplified by virtue of formatting such as bolding, underlining, color change, or by virtue of magnification) by the user.
  • the reading guide feature provides magnification and/or is user-resizable.
  • the reading guide feature advances automatically at a rate that is user-configurable.
  • the device is an eReader device or a tablet computer or a smartphone.
  • Another example embodiment provides a computer readable medium encoded with instructions that when executed by one or more processors, cause a process to be carried out.
  • the process includes, in response to user contact via a touch sensitive interface, providing for display to the user a user interface including a reading guide feature for highlighting a line of content displayed to a user, the user contact uniquely identifying the user's desire to use the reading guide feature and including multiple contact points, and advancing the reading guide feature as the user consumes the content.
  • the reading guide feature includes at least one of a reading pane and/or a straight-edge.
  • the process further includes causing content displayed to the user that is not highlighted by the reading guide feature to be faded.
  • the reading guide feature advances in response to user dragging contact on the touch sensitive surface. In other example cases, the reading guide feature advances automatically.

Abstract

Techniques are disclosed for providing a concentration mode in electronic touch sensitive devices. The concentration mode can be used to highlight given textual content on a line by line basis, or other consumable content. The user can engage the mode with a particular gesture. Once engaged, a UI feature is displayed that assists the user in viewing the given content. In some cases, the UI feature includes a reading pane that encompasses one to three complete lines of displayed text or other content. Background content outside the reading pane can be faded or otherwise softened. In some cases, the UI feature also includes, or alternatively includes, a straight-edged reading guide. The initial placement of the UI feature can be set, for instance, based on the initial contact point touched by the user. As the contact point moves, the UI feature moves accordingly. Concentration mode disengages when user releases contact point.

Description

    FIELD OF THE DISCLOSURE
  • This disclosure relates to electronic display devices, and more particularly, to user interface techniques for interacting with touch sensitive devices.
  • BACKGROUND
  • Electronic display devices such as tablets, eReaders, mobile phones, smart phones, personal digital assistants (PDAs), and other such touch screen electronic display devices are commonly used for displaying consumable content. The content may be, for example, an eBook, an online article or blog, images, a movie or video, a map, just to name a few types. Such display devices are also useful for displaying a user interface that allows a user to interact with an application running on the device. The user interface may include, for example, one or more touch screen controls and/or one or more displayed labels that correspond to nearby hardware buttons. The touch screen display may be backlit or not, and may be implemented for instance with an LED screen or an electrophoretic display. Such devices may also include other touch sensitive surfaces, such as a track pad (e.g., capacitive or resistive touch sensor) or touch sensitive housing (e.g., acoustic sensor).
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIGS. 1 a-b illustrate an example electronic touch sensitive device having a concentration mode configured in accordance with an embodiment of the present invention.
  • FIGS. 1 c-d illustrate example configuration screen shots of the user interface of the electronic touch sensitive device shown in FIGS. 1 a-b, configured in accordance with an embodiment of the present invention.
  • FIG. 2 a illustrates a block diagram of an electronic touch sensitive device configured in accordance with an embodiment of the present invention.
  • FIG. 2 b illustrates a block diagram of a communication system including the electronic touch sensitive device of FIG. 2 b, configured in accordance with an embodiment of the present invention.
  • FIGS. 3 a-b illustrate an example user interface for engaging a concentration mode of an electronic touch sensitive device configured in accordance with an embodiment of the present invention.
  • FIGS. 4 a-f each illustrates an example user interface screen shot showing an electronic touch sensitive device in a concentration mode, in accordance with an embodiment of the present invention.
  • FIG. 5 illustrates a method for providing a concentration mode in an electronic touch sensitive device, in accordance with an embodiment of the present invention.
  • FIG. 6 illustrates a method for providing a concentration mode in an electronic touch sensitive device, in accordance with another embodiment of the present invention.
  • DETAILED DESCRIPTION
  • Techniques are disclosed for providing a concentration mode in electronic touch sensitive devices, such as touch screen and track pad devices. The concentration mode can be used to highlight given textual content on a line by line basis, or other consumable content (e.g., line of photos or graphics, etc). The user can engage the mode, for instance, with a particular touch screen/pad gesture. Once engaged, a user interface (UI) feature is displayed that assists the user in viewing the given content. In some cases, the UI feature includes a reading pane that encompasses one to three complete lines of displayed text or other content. Background content outside the reading pane can be faded or otherwise softened, if so desired. In some cases, the UI feature also includes, or alternatively includes, a straight-edged reading guide. The initial placement of the UI feature can be set, for instance, based on the initial contact point touched by the user, or at some predetermined location. In some cases, as the user moves the contact point, the UI feature moves accordingly; in other cases, once engaged, the UI feature advances automatically at a pace suitable to the user's consumption speed. Concentration mode may be paused or disengaged, for example, based on a further specific user contact or release of contact point.
  • General Overview
  • As previously explained, electronic display devices such as tablets, eReaders, and smart phones are commonly used for displaying user interfaces and consumable content. The user of such devices can typically read the displayed content with relative ease. In some instances, however, the user might prefer a viewing guide to enhance the consumption experience. For instance, some people like to read physical books with a ruler or straight edge to accentuate the current line being read. Likewise, some people prefer to read with a magnifying glass or large print books. While some electronic devices are provided with large print or have a magnifying glass feature, there does not appear to be a user interface reading guide that can be intuitively engaged and disengaged by the reader.
  • Thus, and in accordance with an embodiment of the present invention, techniques are provided for engaging, maintaining, and exiting a concentration mode in electronic touch sensitive devices. The concentration mode can be used, for example, to highlight given textual content on a line by line basis, or other consumable content. In some embodiments, the mode can be engaged globally, so as to allow for use in any number of different applications available on a given device, such as a tablet or mobile phone. For instance, the concentration mode may be used in an eBook application, a browser application, an email application, a word processing application, a PDF viewer application, or any other application where it might be useful to the user to have a reading guide as described herein. The user can engage the mode with a particular gesture that uniquely and globally indicates a desired intent to engage the mode, such as a touch or press-and-hold at the edge of the device touch screen or track pad, or on the device home button (or other such physical feature), or a tap or sequence of taps on a housing configured with acoustics sensors. In a more general sense, any uniquely identifiable touch of a touch sensitive surface by the user can be used to engage the concentration mode. The uniqueness of the touch may be based, for example, on the particular location of the touch and/or the nature of the touch (e.g., a press-and-hold for a predetermined duration, the number of simultaneously applied contact points, a sequence of taps, drags, swipes, circular or elliptical gestures, etc). Once engaged, a UI feature is displayed to the user that assists the user in focusing on the given content. In some cases, the UI feature includes a reading pane that encompasses one to three complete lines of displayed text or other content to be focused on. A ‘complete line’ refers to, for example, a line of content that spans from the farthest left piece of content displayed on a given line to the farthest right piece of content displayed on that same line. The line of content may be a string of words, or one or more graphics provisioned along a line or otherwise spanning from the farthest left edge to the farthest right edge, or a string of Magnification can be provided if so desired, and background content outside the reading pane can be faded or otherwise softened (e.g., blacked-out, blurred, etc) from the user's view so as to further accentuate the content within the reading pane. In some cases, the UI feature may also include, or alternatively include, a virtual ruler or other such straight-edged reading guide. The initial point of focus provided by the UI feature can be set, for instance, based on the initial contact point (e.g., finger or stylus) designated by the user, or at a predetermined location or at the beginning of the text/content for a given page. In some specific embodiments, as the contact point (or points) moves on the touch screen or track pad, the provisioned UI feature moves accordingly, thus allowing the user direct control over advancement of the reading guide UI feature.
  • Thus, once the concentration mode is engaged, the user may continue to maintain the engagement contact points while the UI feature moves in unison with the user's hand movement. In some cases, the user can keep just one of the engagement contact points in place while the UI feature moves in unison with the user dragging of that point. In some such embodiments, the user may move the contact point back up a page (rather than down the page), for example, to review text previously read, if so desired. Once the review is done, the user can then speed-drag the contact point to the desired place and continue with consuming the content in the concentration mode. As will be appreciated in light of this disclosure, numerous dragging schemes can be used to effectively advance the highlighted content region accordingly, in accordance with an embodiment. If no contact points remain, the concentration mode will disengage, in some embodiments.
  • In other embodiments, once the concentration mode is engaged, the displayed UI feature will remain engaged and advance automatically at a predetermined speed (e.g., user-configurable speed) starting at the top of each page (or beginning of content for that page) and steadily progressing to the bottom of each page (or end of content for that page). In such an embodiment, the user can pause the auto-advance function, for instance, by touching a pause button UI feature provisioned somewhere on the touch screen display. In some such cases, if the user manually moves the reading pane/ruler UI feature (via contact with the touch screen or track pad), the auto-advance function will switch to a paused mode and the user may freely move the UI feature based on dragging contact with the touch screen or track pad. Once the user is done with manual manipulation of the reading pane UI feature, the auto-advance mode can be re-started by, for example, touching a play button UI feature provisioned somewhere on the touch screen display (e.g., interchangeably located with pause button). In addition, the user can exit or otherwise terminate the auto-advance concentration mode, for example, by selecting an end mode UI feature provisioned on the touch screen, or by depressing the home button or some other physical button of the device that can be purposed to terminate a given operation mode. Numerous termination schemes will be apparent in light of this disclosure.
  • Additional functions and features can be added to further enhance the user's experience. For example, once the reading guide is deployed, the user may touch the content within (and/or at the edge of) the reading guide area so as to further cause that text to be accentuated. For instance, some people like to read with their finger tip. Thus, in one example case for an eBook application, as the user places a finger at the beginning of a sentence within/at the reading guide, another UI feature is activated to highlight the word at the user's fingertip, in accordance with an embodiment. The highlighting may be, for instance, bolding, underlining, magnifying, or a change in color. As the user moves her fingertip down the sentence, the word highlighting advances accordingly to follow the finger. In another embodiment, where the reading guide includes a reading pane, the user may resize the reading pane once it is deployed on screen so that, for example, additional lines of text can be viewed (e.g., change from one to three lines of text). In some such cases, a highlighting feature as just described can also be provisioned. With the expanded reading pane, the highlighting feature may include several words at a time (e.g., 2 to 5 words), if so desired, and in accordance with an embodiment of the present invention. Such optional features can be user-configured or hard-coded, as will be appreciated.
  • The techniques can be implemented in an intuitive manner, so that the user can seamlessly and selectively engage, pause, and/or disengage the concentration mode without disrupting the reading experience, in accordance with some embodiments. For instance, in one specific such embodiment, a two-finger press-and-hold or other touch gesture along the border region of the display screen can be used to automatically create the UI feature (e.g., pane or straight edge) to accentuate one complete line of text at a time. The two fingers may be, for example, the user's pointer and middle fingers (or the thumb and pointer finger) of one hand on one side of the display, or the user's thumbs on each side of the display. Alternatively, a stylus could be used in combination with one or more fingers. For instance, for a right-handed user, the left thumb could be placed at the left-side border of the display and the stylus tip at the opposing right-side border. Alternatively, a stylus could be used on its own to engage the mode, for example, by laying the stylus on (or otherwise sufficiently near) the touch sensitive surface so as to effectively create multiple contact points that can be identified as a request for the concentration mode. In any such cases, the use of two (or more) contact points can be used to uniquely identify the user's desire to engage the concentration mode. Numerous uniquely identifiable engagement schemes that exploit a touch sensitive surface can be used as will be appreciated in light of this disclosure. Further note that any touch sensitive device (e.g., track pad, touch screen, or other touch sensitive surface, whether capacitive, resistive, acoustic or other touch detecting technology) may be used to detect the user contact, and the claimed invention is not intended to be limited to any particular type of touch sensitive technology, unless expressly stated.
  • Architecture
  • FIGS. 1 a-b illustrate an example electronic touch sensitive device having a concentration mode configured in accordance with an embodiment of the present invention. The device could be, for example, a tablet such as the NOOK® Tablet by Barnes & Noble. In a more general sense, the device may be any electronic device having a touch sensitive user interface and capability for displaying content to a user, such as a mobile phone or mobile computing device such as a eReader or laptop, a desktop computing system, a television, a smart display screen, or any other device having a touch screen display or a non-touch display screen that can be used in conjunction with a touch sensitive surface. As will be appreciated, the claimed invention is not intended to be limited to any particular kind or type of electronic device.
  • As can be seen with this example embodiment, the device comprises a housing that includes a number of hardware features such as a power button, control features, and a press-button (sometimes called a home button herein). A touch screen based user interface is also provided, which in this example embodiment includes a quick navigation menu having six main categories to choose from (Home, Library, Shop, Search, Light, and Settings) and a status bar that includes a number of icons (a night-light icon, a wireless network icon, and a book icon), a battery indicator, and a clock. Other embodiments may have fewer or additional such UI touch screen controls and features, or different UI touch screen controls and features altogether, depending on the target application of the device. Any such general UI controls and features can be implemented using any suitable conventional or custom technology, as will be appreciated.
  • The hardware control features provided on the device housing in this example embodiment are configured as elongated press-bars and can be used, for example, to page forward (using the top press-bar) or to page backward (using the bottom press-bar), such as might be useful in an eReader application. The power button can be used to turn the device on and off, and may be used in conjunction with a touch-based UI control feature that allows the user to confirm a given power transition action request (e.g., such as a slide bar or tap point graphic to turn power off). Numerous variations will be apparent, and the claimed invention is not intended to be limited to any particular set of hardware buttons or features, or device form factor.
  • In this example configuration, the home button is a physical press-button that can be used as follows: when the device is awake and in use, tapping the button will display the quick navigation menu, which is a toolbar that provides quick access to various features of the device. The home button may also be configured to cease an active function that is currently executing on the device, such as an auto-advance reading pane UI feature as described herein. The button may further control other functionality if, for example, the user presses and holds the home button. For instance, an example such push-and-hold function could engage a power conservation routine where the device is put to sleep or an otherwise lower power consumption mode. So, a user could grab the device by the button, press and keep holding as the device was stowed into a bag or purse. One physical gesture that would safely put the device to sleep. Thus, in such an example embodiment, the home button may be associated with and control different and unrelated actions: 1) show the quick navigation menu; 2) exit the auto-advance reading pane UI feature mode, but keep the page being read or otherwise consumed displayed; and 3) put the device to sleep.
  • As can be further seen, the status bar may also include a book icon (upper left corner). In some such cases, the user can access a sub-menu that provides access to a concentration mode configuration sub-menu by tapping the book icon of the status bar. For example, upon receiving an indication that the user has touched the book icon, the device can then display the concentration mode configuration sub-menu shown in FIG. 1 d. In other cases, tapping the book icon just provides bibliographic information on the content being consumed. Another example way for the user to access a concentration mode configuration sub-menu such as the one shown in FIG. 1 d is to tap or otherwise touch the Settings option in the quick navigation menu, which causes the device to display the general sub-menu shown in FIG. 1 c. From this general sub-menu the user can select any one of a number of options, including one designated Screen/UI in this specific example case. Selecting this sub-menu item (with an appropriately placed screen tap) may cause the concentration mode configuration sub-menu of FIG. 1 d to be displayed, in accordance with an embodiment. In other example embodiments, selecting the Screen/UI option may present the user with a number of additional sub-options, one of which may include a so-called concentration mode option, which may then be selected by the user so as to cause the concentration mode configuration sub-menu of FIG. 1 d to be displayed. Any number of such menu schemes and nested hierarchies can be used, as will be appreciated in light of this disclosure.
  • As will be appreciated, the various UI control features and sub-menus displayed to the user are implemented as UI touch screen controls in this example embodiment. Such UI touch screen controls can be programmed or otherwise configured using any number of conventional or custom technologies. In general, the touch screen translates the user touch in a given location into an electrical signal which is then received and processed by the underlying operating system (OS) and circuitry (processor, etc). Additional example details of the underlying OS and circuitry in accordance with some embodiments will be discussed in turn with reference to FIG. 2 a.
  • As previously explained, and with further reference to FIGS. 1 c and 1 d, once the Settings sub-menu is displayed (FIG. 1 c), the user can then select the Screen/UI option. In response to such a selection, the concentration mode configuration sub-menu shown in FIG. 1 d can be provided to the user. In this example case, the concentration mode configuration sub-menu includes a UI check box that when checked or otherwise selected by the user, effectively enables the concentration mode (shown in the enabled state); unchecking the box disables the mode. Other embodiments may have the concentration mode always enabled, or enabled by a physical switch or button, for example. The user can configure a number of features with respect to the concentration mode, in this example embodiment. In more detail, by checking the Reading Pane check box, the user can select a reading pane based concentration mode which allows a reading pane to be displayed during concentration mode. The reading pane effectively encompasses or otherwise highlights one line of text at a time, in some example cases. As can be further seen, the user may also check the Borderless check box to provide a borderless reading pane and/or the Magnify check box to provide a reading pane having a desired degree of magnification (e.g., 1.5×, 2×, 3×, which may be fixed or user configurable), if so desired. In addition, or alternatively, by checking the Ruler check box, the user can select a ruler-based concentration mode which allows a virtual ruler or straight-edge to be displayed during concentration mode. In this example case, the ruler is deployed below the line of text to be read by default, but the user can further check the Above Text check box to provide the ruler above the line of text to be read, if so desired. Note that both the Ruler and Reading Pane check boxes may be checked to provide a hybrid UI feature to assist the reading experience. With further reference to the example embodiment of FIG. 1 d, note that the reading pane/ruler UI feature is advanced manually by the user as the default setting, but the user can further check the Auto-advance check box which allows the UI feature to advance automatically at a pace suitable to the user's consumption speed, if so desired. Note that user can further specify the delay time the UI feature dwells on each line of text (or other content being consumed).
  • With further reference to FIG. 1 d, the user may also set certain background features in this example embodiment. In particular, the text/content outside the area of focus provided by the reading pane/rule UI feature is effectively left as-is by default setting, but the user can further check the Content Fade check box which grays out or otherwise softens background content not within the highlighted reading area, if so desired. The user may also select the color scheme for the highlighted reading area and/or background, by selecting the Color UI control and selecting the desired scheme, for example. Also configurable in this example embodiment is the mechanism by which the user engages the concentration mode as well as the activation time. In particular, the user can assign the mode engagement function to a hardware switch/feature such as the home button by checking the Button Press check box, or to an area of the touch screen by checking the Edge Touch check box. Also, note that the duration which the user's press-and-hold gesture (on either the hardware button or the touch screen) to engage the concentration mode is set to 3 seconds in this example case. Other embodiments may not use such a delay and activate upon contact in the activation region. In addition, a back button arrow UI control feature is provisioned on the touch screen, so that the user can go back to the Settings menu after the concentration mode has been configured, if so desired. While a number of user-configurable features are provided in some embodiments, other embodiments may have any combination of similar features wherein the features are set as a given default or hard-coded, such that no user configuration is necessary.
  • FIG. 2 a illustrates a block diagram of an electronic touch screen device configured in accordance with an embodiment of the present invention. As can be seen, this example device includes a processor, random access memory (RAM), additional storage/memory, a communications module, a touch screen, and an audio module. A communications bus and interconnect is also provided to allow inter-device communication. Other typical componentry and functionality not reflected in the block diagram will be apparent (e.g., battery, co-processor, etc). Further note that although a touch screen display is provided, other embodiments may include a non-touch screen and a touch sensitive surface such as a track pad, or a touch sensitive housing configured with one or more acoustic sensors, etc. In any such cases, the touch sensitive surface is generally capable of translating a user's physical contact with the surface into an electronic signal that can be manipulated or otherwise used to trigger a specific user interface action, such as those provided herein. The principles provided herein equally apply to any such touch sensitive devices. For ease of description, examples are provided with touch screen technology.
  • In this example embodiment, the RAM includes a number of modules stored therein that can be accessed and executed by the processor (and/or a co-processor). The modules include an operating system (OS), a user interface (UI), and a power conservation routine (Power). The modules can be implemented, for example, in any suitable programming language (e.g., C, C++, objective C, JavaScript, custom or proprietary instruction sets, etc), and encoded on a machine readable medium, that when executed by the processor (and/or co-processors), carries out the functionality of the device including a UI having a concentration mode as described herein. Other embodiments can be implemented, for instance, with gate-level logic or an application-specific integrated circuit (ASIC) or chip set or other such purpose built logic, or a microcontroller having input/output capability (e.g., inputs for receiving user inputs and outputs for directing other components) and a number of embedded routines for carrying out the device functionality. In short, the functional modules can be implemented in hardware, software, firmware, or a combination thereof.
  • The processor can be any suitable processor (e.g., 800 MHz Texas Instruments OMAP3621 applications processor), and may include one or more co-processors or controllers to assist in device control. In this example case, the processor receives input from the user, including input from or otherwise derived from the power button and the home button. The processor can also have a direct connection to a battery so that it can perform base level tasks even during sleep or low power modes. The RAM can be any suitable type of memory and size (e.g., 256 or 512 Mbytes SDRAM), and in other embodiments may be implemented with non-volatile memory or a combination of non-volatile and volatile memory technologies. The storage can also be implemented with any suitable memory and size (e.g., 2 GBytes of flash memory). The display can be implemented, for example, with a 6-inch E-ink Pearl 800×600 pixel screen with Neonode® zForce® touch screen, or any other suitable display and touch screen interface technology. The communications module can be, for instance, any suitable 802.11 b/g/n WLAN chip or chip set, which allows for connection to a local network so that content can be downloaded to the device from a remote location (e.g., content provider, etc, depending on the application of the display device). In some specific example embodiments, the device housing that contains all the various componentry measures about 6.5″ high by about 5″ wide by about 0.5″ thick, and weighs about 6.9 ounces. Any number of suitable form factors can be used, depending on the target application (e.g., laptop, desktop, mobile phone, etc). The device may be smaller, for example, for smartphone and tablet applications and larger for smart computer monitor applications.
  • The operating system (OS) module can be implemented with any suitable OS, but in some example embodiments is implemented with Google Android OS or Linux OS or Microsoft OS or Apple OS. As will be appreciated in light of this disclosure, the techniques provided herein can be implemented on any such platforms. The power management (Power) module can be configured as typically done, such as to automatically transition the device to a low power consumption or sleep mode after a period of non-use. A wake-up from that sleep mode can be achieved, for example, by a physical button press and/or a touch screen swipe or other action. The user interface (UI) module can be, for example, based on touch screen technology and the various example screen shots shown in FIGS. 1 a-d, 3 a-b, and 4 a-f in conjunction with the concentration mode methodologies demonstrated in FIGS. 5 and 6, which will be discussed in turn. The audio module can be configured, for example, to speak or otherwise aurally present a selected eBook or other textual content, if preferred by the user. Numerous commercially available text-to-speech modules can be used, such as Verbose text-to-speech software by NCH Software. In some example cases, if additional space is desired, for example, to store digital books or other content and media, storage can be expanded via a microSD card or other suitable memory expansion technology (e.g., 32 GBytes, or higher).
  • Client-Server System
  • FIG. 2 b illustrates a block diagram of a communication system including the electronic touch sensitive device of FIG. 2 b, configured in accordance with an embodiment of the present invention. As can be seen, the system generally includes an electronic touch sensitive device that is capable of communicating with a server via a network/cloud. In this example embodiment, the electronic touch sensitive device may be, for example, an eBook reader, a mobile cell phone, a laptop, a tablet, desktop, or any other touch sensitive computing device. The network/cloud may be a public and/or private network, such as a private local area network operatively coupled to a wide area network such as the Internet. In this example embodiment, the server may be programmed or otherwise configured to receive content requests from a user via the touch sensitive device and to respond to those requests by providing the user with requested or otherwise recommended content. Is some such embodiments, the server is configured to remotely provision a reading pane/rule UI feature as provided herein to the touch sensitive device (e.g., via JavaScript or other browser based technology). In other embodiments, portions of the methodology are executed on the server and other portions of the methodology are executed on the device. Numerous server-side/client-side execution schemes can be implemented, as will be apparent in light of this disclosure.
  • Methodologies
  • FIGS. 3 a-b illustrate an example user interface for engaging a concentration mode of an electronic touch sensitive device configured in accordance with an embodiment of the present invention. As can be seen, the touch screen generally includes a content portion (within the dashed line area) and an edge portion. A physical frame or support structure is provided about the screen. In the example engagement mode shown in FIG. 3 a, the user is using two contact points on opposite edges of the touch screen. Note that the two points do not need to be aligned in the horizontal direction, but rather can each be on any point along the respective edge, in some such embodiments. As can be further seen, the user may use any two fingers/thumbs or stylus to create the two (or more) contact points. FIG. 3 b is similar to FIG. 3 a, except that the two contact points are on the same side of the screen. Further note that the contact points may be in both the edge portion and the content portion. In a more general sense, any portion of the touch screen display can be used as a concentration mode activation zone. Likewise, as previously explained, a hardware button may also be used to activate the concentration mode, if so desired. Numerous uniquely identifiable engagement schemes can be used as will be appreciated in light of this disclosure.
  • Once engaged, the concentration mode provides the user a UI reading assist feature as configured by the user, as previously discussed with respect to FIG. 1 d, for instance. FIGS. 4 a-f each illustrates a screen shot showing example such UI features, in accordance with an embodiment of the present invention. With respect to the example of FIG. 4 a, a borderless reading pane is provided that highlights a single line of text/content to the user, and all other background content is faded out. As the user's hand or stylus moves down the virtual page, so does the reading pane. The example reading pane illustrated in FIG. 4 b is similar to that shown in FIG. 4 a, but is provided with a border and magnification. Recall that the user can specify the degree of magnification, in accordance with some embodiments. Further recall that the user may also resize the reading guide, for example, so that additional lines of text can be viewed (e.g., by simultaneously pulling top and bottom edges of the reading pane in opposite directions, or any other such suitable resizing gesture). Further recall that content highlighting within or otherwise at the reading guide can be used if so desired (e.g., reading with fingertip scenario). With respect to the specific example of FIG. 4 c, a virtual ruler or straight-edge reading guide is provided that highlights a single line of text/content to the user. In this example case, the background text is not faded out, but could be if so desired. The user manually controls the placement of the ruler based on dragging of the contact point.
  • With respect to the specific example of FIG. 4 d, a borderless reading pane is provided that highlights a single line of text/content to the user, and the background text is not faded out. In this example case, the auto-advance function is selected, so the reading pane automatically advances at a rate suitable to the user's consumption speed. As previously explained, the user may pause the auto-advancement of the UI feature by tapping the pause button provided in the lower right corner of the touch screen, in this example case. Upon pausing, the pause UI feature changes to a play UI feature that the user can tap to re-start the auto-advancement. Also recall that the user can at any time manually control the placement of the reading pane by contacting the touch screen over the pane (or proximate to the pane) and dragging to a point of interest in the displayed content. In some such example cases, the initial touch by the user causes the auto-advance to pause (and the UI play feature then appears), and when the user is done dragging the reading pane, the UI play feature can then be tapped to resume auto-advancement (e.g., from the point where the reading pane was left by the user, or from the point where auto-advancement was paused). The user can exit auto-advancement, for example, by pressing the home button or some other hardware or software feature provisioned for allowing exit from the auto-advancement routine. The example auto-advancing reading pane illustrated in FIG. 4 e is similar to that shown in FIG. 4 d, but is provided with a border and magnification, and faded background content. FIG. 4 f shows an example of an auto-advancing reading pane and ruler combination, with faded background content. Numerous configurations will be apparent.
  • FIG. 5 illustrates a method for providing a concentration mode in an electronic touch screen device, in accordance with an embodiment of the present invention. This example methodology may be implemented, for instance, by the UI module of the touch sensitive device shown in FIG. 2 a, or the touch sensitive device shown in FIG. 2 b (e.g., with the UI provisioned to the client by the server). To this end, the UI can be implemented in software, hardware, firmware, or any combination thereof, as will be appreciated in light of this disclosure.
  • As can be seen, the method includes an activation phase and a maintain/exit phase. Each phase includes sensing a user's input by a touch sensitive surface. In general, any touch sensitive device may be used to detect contact with it by one or more fingers and/or styluses. As soon as the user begins to drag or otherwise move the contact points, the UI code (and/or hardware) can assume a drag gesture has been engaged and track the path of each contact point with respect to any fixed point within the touch surface until the user stops engaging the touch sensitive surface. The release point can also be captured by the UI as it may be used to commit the action started when the user pressed on the touch sensitive surface. In a similar fashion, if the user releases hold without moving the contact point, a press or press-and-hold command may be assumed depending on the amount of time the user was continually pressing on the touch sensitive surface. These main detections can be used in various ways to implement UI functionality.
  • In this example case, the method includes detecting 501 user contact at the touch sensitive interface. In general, the touch monitoring is effectively continuous. The method continues with determining 503 if the contact is in the concentration mode activation zone. As previously indicated, one example activation zone is the edge of the touch sensitive surface. Other embodiments may have the activation zone anywhere on the touch sensitive surface that is specified by the user via a configuration menu. In a more general sense, the concentration mode activation zone can be pre-established in any specific location on the touch sensitive surface of the device.
  • In any case, if the contact is not in the concentration mode activation zone, then the method may continue with reviewing 504 the contact for some other UI request (e.g., select a file, send an email, etc). On the other hand, if the contact is in the concentration mode activation zone, the method continues with determining 505 if the contact is persistent. For instance, if the contact is a press-and-hold that lasts greater than 3 seconds (or some other designated time period), then the UI code can assume the contact is persistent. In any case, if the contact is not sufficiently persistent, then the method may continue with reviewing 504 the contact for some other UI request (e.g., select a file, send an email, etc). On the other hand, if the contact is deemed to be sufficiently persistent, then the method continues with activating 507 the concentration mode, or maintaining that mode along with the current reading guide feature position on the display if already activated.
  • In the maintain/exit phase, the method continues with determining 509 if the contact on the touch sensitive surface is persistent and moving. If the contact is persistent but not moving (e.g., movement ≦2 mm), then the UI code assumes the user is actively consuming content of the current line and the method continues with maintaining 507 the concentration mode along with the current reading guide feature position on the display. If the contact is not persistent, then the method continues with exiting 512 the concentration mode. However, if the contact is both persistent and moving (e.g., movement >2 mm), then the method continues with moving 511 the reading pane/ruler UI feature to next line or to the otherwise desired area of the content indicated by the contact movement on the touch sensitive surface. As can be further seen, the method includes continuously monitoring the contact point for persistence and movement.
  • FIG. 6 illustrates a method for providing a concentration mode in an electronic touch screen device, in accordance with another embodiment of the present invention. Just as with the method of FIG. 5, this methodology includes an activation phase and a maintain/exit phase, and the previous relevant discussion is equally applicable here. However, the methodology employs an auto-advance feature. The activation phase includes detecting 601 user contact at the touch sensitive interface, and determining 603 if that contact indicates that the concentration mode is desired (e.g., based on the unique set of multiple contact points provided by the user). If not, then the method may continue with reviewing 604 the contact for some other UI request (e.g., select a file, send an email, etc). On the other hand, if the contact indicates that the concentration mode is desired, the method continues with activating 605 the concentration mode, or maintaining that mode if already activated and auto-advance the reading pane/ruler UI feature position.
  • The method continues in the maintain/exit phase with determining 607 if the user has selected pause or wishes to exit the concentration mode (this monitoring is continuous in this example embodiment). If the user has selected pause, then the method continues with holding 609 the current reading pane/ruler UI feature position, and the continuous monitoring continues. If the user releases the pause by selecting a play UI feature, then the method continues with maintaining 605 the concentration mode and auto-advancement of the reading pane/ruler UI feature position, and the continuous monitoring continues. If the user signal an exit is desired (e.g., by selecting an exit UI feature or pressing the home button), then the method continues with exiting 611 the concentration mode, and the continuous touch monitoring continues.
  • Numerous variations will be apparent in light of this disclosure. For instance, one example embodiment provides a device that includes a display for displaying content to a user, and a touch sensitive surface for allowing user input. The device further includes a user interface including a reading guide feature for highlighting a line of the content displayed to a user, the reading guide feature configured to be displayed on the display in response to user contact via the touch sensitive interface, the user contact uniquely identifying the user's desire to use the reading guide feature and including multiple contact points. In some cases, the display is a touch screen display that includes the touch sensitive surface. In some cases, the touch sensitive surface is a track pad. In some cases, the reading guide feature includes at least one of a reading pane and/or a straight-edge. In some cases, content displayed to the user that is not highlighted by the reading guide feature is faded (e.g., grayed out, blacked-out, blurred, etc). In some cases, the reading guide feature provides magnification and/or is user-resizable (so as to view more lines of text). In some such cases, the degree of magnification is user-configurable. In some cases, the reading guide feature advances in response to user dragging contact on the touch sensitive surface. In some cases, the reading guide feature advances automatically. In some cases, the reading guide feature advances automatically at a rate that is user-configurable.
  • Another example embodiment provides a device that includes a display having a touch screen interface and for displaying content to a user, and a user interface including a reading guide feature for highlighting a complete line of the content displayed to a user, the reading guide feature configured to be displayed on the display in response to user contact via the touch screen interface, the user contact uniquely identifying the user's desire to use the reading guide feature and including multiple contact points, wherein the reading guide feature includes at least one of a reading pane and/or a straight-edge. In some cases, content displayed to the user that is not highlighted by the reading guide feature is faded, and/or content that is highlighted by the reading guide can be selectively amplified (e.g., amplified by virtue of formatting such as bolding, underlining, color change, or by virtue of magnification) by the user. In some cases, the reading guide feature provides magnification and/or is user-resizable. In some cases, the reading guide feature advances automatically at a rate that is user-configurable. In some cases, the device is an eReader device or a tablet computer or a smartphone.
  • Another example embodiment provides a computer readable medium encoded with instructions that when executed by one or more processors, cause a process to be carried out. The process includes, in response to user contact via a touch sensitive interface, providing for display to the user a user interface including a reading guide feature for highlighting a line of content displayed to a user, the user contact uniquely identifying the user's desire to use the reading guide feature and including multiple contact points, and advancing the reading guide feature as the user consumes the content. In some cases, the reading guide feature includes at least one of a reading pane and/or a straight-edge. In some cases, the process further includes causing content displayed to the user that is not highlighted by the reading guide feature to be faded. In some cases, the reading guide feature advances in response to user dragging contact on the touch sensitive surface. In other example cases, the reading guide feature advances automatically.
  • The foregoing description of the embodiments of the invention has been presented for the purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise form disclosed. Many modifications and variations are possible in light of this disclosure. It is intended that the scope of the invention be limited not by this detailed description, but rather by the claims appended hereto.

Claims (20)

What is claimed is:
1. A device, comprising:
a display for displaying content to a user;
a touch sensitive surface for allowing user input; and
a user interface including a reading guide feature for highlighting a line of the content displayed to a user, the reading guide feature configured to be displayed on the display in response to user contact via the touch sensitive interface, the user contact uniquely identifying the user's desire to use the reading guide feature and including multiple contact points.
2. The device of claim 1 wherein the display is a touch screen display that includes the touch sensitive surface.
3. The device of claim 1 wherein the touch sensitive surface is a track pad.
4. The device of claim 1 wherein the reading guide feature includes at least one of a reading pane and/or a straight-edge.
5. The device of claim 1 wherein content displayed to the user that is not highlighted by the reading guide feature is faded.
6. The device of claim 1 wherein the reading guide feature provides magnification and/or is user-resizable.
7. The device of claim 6 wherein the degree of magnification is user-configurable.
8. The device of claim 1 wherein the reading guide feature advances in response to user dragging contact on the touch sensitive surface.
9. The device of claim 1 wherein the reading guide feature advances automatically.
10. The device of claim 1 wherein the reading guide feature advances automatically at a rate that is user-configurable.
11. A device, comprising:
a display having a touch screen interface and for displaying content to a user; and
a user interface including a reading guide feature for highlighting a complete line of the content displayed to a user, the reading guide feature configured to be displayed on the display in response to user contact via the touch screen interface, the user contact uniquely identifying the user's desire to use the reading guide feature and including multiple contact points, wherein the reading guide feature includes at least one of a reading pane and/or a straight-edge.
12. The device of claim 11 wherein content displayed to the user that is not highlighted by the reading guide feature is faded, and/or content that is highlighted by the reading guide can be selectively amplified by the user.
13. The device of claim 11 wherein the reading guide feature provides magnification and/or is user-resizable.
14. The device of claim 11 wherein the reading guide feature advances automatically at a rate that is user-configurable.
15. The device of claim 11 wherein the device is an eReader device or a tablet computer or a smartphone.
16. A computer readable medium encoded with instructions that when executed by one or more processors, cause a process to be carried out, the process comprising:
in response to user contact via a touch sensitive interface, providing for display to the user a user interface including a reading guide feature for highlighting a line of content displayed to a user, the user contact uniquely identifying the user's desire to use the reading guide feature and including multiple contact points; and
advancing the reading guide feature as the user consumes the content.
17. The computer readable medium of claim 16 wherein the reading guide feature includes at least one of a reading pane and/or a straight-edge.
18. The computer readable medium of claim 16 wherein the process further includes causing content displayed to the user that is not highlighted by the reading guide feature to be faded.
19. The computer readable medium of claim 16 wherein the reading guide feature advances in response to user dragging contact on the touch sensitive surface.
20. The computer readable medium of claim 16 wherein the reading guide feature advances automatically.
US13/715,221 2012-12-14 2012-12-14 Touch sensitive device with concentration mode Active 2033-04-08 US8963865B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/715,221 US8963865B2 (en) 2012-12-14 2012-12-14 Touch sensitive device with concentration mode

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US13/715,221 US8963865B2 (en) 2012-12-14 2012-12-14 Touch sensitive device with concentration mode

Publications (2)

Publication Number Publication Date
US20140168076A1 true US20140168076A1 (en) 2014-06-19
US8963865B2 US8963865B2 (en) 2015-02-24

Family

ID=50930278

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/715,221 Active 2033-04-08 US8963865B2 (en) 2012-12-14 2012-12-14 Touch sensitive device with concentration mode

Country Status (1)

Country Link
US (1) US8963865B2 (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140184531A1 (en) * 2012-12-27 2014-07-03 Kabushiki Kaisha Toshiba Electronic device, display method, and display program
US20140237427A1 (en) * 2013-02-20 2014-08-21 Fuji Xerox Co., Ltd. Browsing device, browsing system, and non-transitory computer readable medium
US20150095835A1 (en) * 2013-09-30 2015-04-02 Kobo Incorporated Providing a user specific reader mode on an electronic personal display
US20160012029A1 (en) * 2014-07-09 2016-01-14 Lg Electronics Inc. Mobile terminal and method of controlling the same
US20160034090A1 (en) * 2013-11-15 2016-02-04 Boe Technology Group Co., Ltd. Touch system and display device
US20170123624A1 (en) * 2015-11-02 2017-05-04 Guangzhou Ucweb Computer Technology Co., Ltd. Methods, apparatuses, and devices for processing interface displays
US20170124622A1 (en) * 2014-11-14 2017-05-04 The Joan and Irwin Jacobs Technion-Cornell Institute System and method for intuitive content browsing
US20170245017A1 (en) * 2014-06-12 2017-08-24 Apple Inc. Systems and Methods for Presenting and Interacting with a Picture-in-Picture Representation of Video Content on an Electronic Device with a Touch-Sensitive Display
US20180300019A1 (en) * 2013-03-14 2018-10-18 Samsung Electronics Co., Ltd. Electronic device and method for controlling screen display using temperature and humidity
US10732820B2 (en) 2014-06-12 2020-08-04 Apple Inc. Systems and methods for efficiently navigating between applications with linked content on an electronic device with a touch-sensitive display
US10824987B2 (en) 2014-11-14 2020-11-03 The Joan and Irwin Jacobs Technion-Cornell Institute Techniques for embedding virtual points of sale in electronic media content
US11966578B2 (en) 2019-05-16 2024-04-23 Apple Inc. Devices and methods for integrating video with user interface navigation

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9423825B2 (en) * 2013-02-08 2016-08-23 Nvidia Corporation Mobile computing device with expanded display size

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030068088A1 (en) * 2001-10-04 2003-04-10 International Business Machines Corporation Magnification of information with user controlled look ahead and look behind contextual information
US20030193481A1 (en) * 2002-04-12 2003-10-16 Alexander Sokolsky Touch-sensitive input overlay for graphical user interface
US20060026521A1 (en) * 2004-07-30 2006-02-02 Apple Computer, Inc. Gestures for touch sensitive input devices
US20100088653A1 (en) * 2008-10-07 2010-04-08 Research In Motion Limited Portable electronic device and method of controlling same
US20110148786A1 (en) * 2009-12-18 2011-06-23 Synaptics Incorporated Method and apparatus for changing operating modes
US20110320978A1 (en) * 2010-06-29 2011-12-29 Horodezky Samuel J Method and apparatus for touchscreen gesture recognition overlay
US20120192117A1 (en) * 2011-01-24 2012-07-26 Migos Charles J Device, Method, and Graphical User Interface with a Dynamic Gesture Disambiguation Threshold
US20120293528A1 (en) * 2011-05-18 2012-11-22 Larsen Eric J Method and apparatus for rendering a paper representation on an electronic display
US20120311438A1 (en) * 2010-01-11 2012-12-06 Apple Inc. Electronic text manipulation and display
US20130047115A1 (en) * 2011-08-19 2013-02-21 Apple Inc. Creating and viewing digital note cards

Family Cites Families (47)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6259438B1 (en) 1998-06-04 2001-07-10 Wacom Co., Ltd. Coordinate input stylus
US4896543A (en) 1988-11-15 1990-01-30 Sri International, Inc. Three-axis force measurement stylus
WO1996012222A1 (en) 1994-10-14 1996-04-25 Ast Research, Inc. A system and method for detecting screen hotspots
JP2717774B2 (en) 1995-01-13 1998-02-25 株式会社ワコム Pressure sensitive element and stylus pen with pressure sensitive function
GB9722766D0 (en) 1997-10-28 1997-12-24 British Telecomm Portable computers
US7840912B2 (en) 2006-01-30 2010-11-23 Apple Inc. Multi-touch gesture dictionary
US20010025289A1 (en) 1998-09-25 2001-09-27 Jenkins Michael D. Wireless pen input device
US20020116421A1 (en) 2001-02-17 2002-08-22 Fox Harold L. Method and system for page-like display, formating and processing of computer generated information on networked computers
US7133862B2 (en) 2001-08-13 2006-11-07 Xerox Corporation System with user directed enrichment and import/export control
US7649524B2 (en) 2004-07-15 2010-01-19 N-Trig Ltd. Tracking window for a digitizer system
US8838591B2 (en) 2005-08-23 2014-09-16 Ricoh Co., Ltd. Embedding hot spots in electronic documents
US7898541B2 (en) 2004-12-17 2011-03-01 Palo Alto Research Center Incorporated Systems and methods for turning pages in a three-dimensional electronic document
US7779347B2 (en) 2005-09-02 2010-08-17 Fourteen40, Inc. Systems and methods for collaboratively annotating electronic documents
US7694231B2 (en) 2006-01-05 2010-04-06 Apple Inc. Keyboards for portable electronic devices
US7657849B2 (en) 2005-12-23 2010-02-02 Apple Inc. Unlocking a device by performing gestures on an unlock image
US8988357B2 (en) 2006-08-10 2015-03-24 Sony Corporation Stylus activated display/key-lock
US8564544B2 (en) 2006-09-06 2013-10-22 Apple Inc. Touch screen device, method, and graphical user interface for customizing display of content category icons
US8059101B2 (en) 2007-06-22 2011-11-15 Apple Inc. Swipe gestures for touch screen keyboards
US20100100854A1 (en) 2008-10-16 2010-04-22 Dell Products L.P. Gesture operation input system
US8631354B2 (en) 2009-03-06 2014-01-14 Microsoft Corporation Focal-control user interface
US8274536B2 (en) 2009-03-16 2012-09-25 Apple Inc. Smart keyboard management for a multifunction device with a touch screen display
US20100259482A1 (en) 2009-04-10 2010-10-14 Microsoft Corporation Keyboard gesturing
US9886936B2 (en) 2009-05-14 2018-02-06 Amazon Technologies, Inc. Presenting panels and sub-panels of a document
US20100295782A1 (en) 2009-05-21 2010-11-25 Yehuda Binder System and method for control based on face ore hand gesture detection
US9285988B2 (en) 2010-04-20 2016-03-15 Blackberry Limited Portable electronic device having touch-sensitive display with variable repeat rate
USD668674S1 (en) 2010-07-26 2012-10-09 Apple Inc. Display screen or portion thereof with icon
US8754858B2 (en) 2010-09-07 2014-06-17 STMicroelectronics Aisa Pacific Pte Method to parameterize and recognize circular gestures on touch sensitive surfaces
US8963836B2 (en) 2010-09-17 2015-02-24 Tencent Technology (Shenzhen) Company Limited Method and system for gesture-based human-machine interaction and computer-readable medium thereof
US9639178B2 (en) 2010-11-19 2017-05-02 Apple Inc. Optical stylus
KR101787750B1 (en) 2010-12-01 2017-10-19 삼성전자주식회사 Capacitive stylus pen
US9645986B2 (en) 2011-02-24 2017-05-09 Google Inc. Method, medium, and system for creating an electronic book with an umbrella policy
US20120242584A1 (en) 2011-03-22 2012-09-27 Nokia Corporation Method and apparatus for providing sight independent activity reports responsive to a touch gesture
US8922489B2 (en) 2011-03-24 2014-12-30 Microsoft Corporation Text input using key and gesture information
US20120280947A1 (en) 2011-05-06 2012-11-08 3M Innovative Properties Company Stylus with pressure sensitive membrane
US8508494B2 (en) 2011-06-01 2013-08-13 Motorola Mobility Llc Using pressure differences with a touch-sensitive display screen
US20120329529A1 (en) 2011-06-21 2012-12-27 GreatCall, Inc. Gesture activate help process and system
US8913019B2 (en) 2011-07-14 2014-12-16 Microsoft Corporation Multi-finger detection and component resolution
US9256361B2 (en) 2011-08-03 2016-02-09 Ebay Inc. Control of search results with multipoint pinch gestures
US8976128B2 (en) 2011-09-12 2015-03-10 Google Technology Holdings LLC Using pressure differences with a touch-sensitive display screen
US9495012B2 (en) 2011-09-27 2016-11-15 Z124 Secondary single screen mode activation through user interface activation
US8286104B1 (en) 2011-10-06 2012-10-09 Google Inc. Input method application for a touch-sensitive user interface
KR101799408B1 (en) 2011-11-03 2017-11-20 삼성전자주식회사 Apparatus and method for controlling controllable device in portable terminal
KR20130052151A (en) 2011-11-11 2013-05-22 삼성전자주식회사 Data input method and device in portable terminal having touchscreen
CN102520847A (en) 2011-11-25 2012-06-27 鸿富锦精密工业(深圳)有限公司 Electronic reading device and page processing method thereof
JP6194167B2 (en) 2011-11-25 2017-09-06 京セラ株式会社 Apparatus, method, and program
JP6159078B2 (en) 2011-11-28 2017-07-05 京セラ株式会社 Apparatus, method, and program
US8860763B2 (en) 2012-01-31 2014-10-14 Xerox Corporation Reversible user interface component

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030068088A1 (en) * 2001-10-04 2003-04-10 International Business Machines Corporation Magnification of information with user controlled look ahead and look behind contextual information
US20030193481A1 (en) * 2002-04-12 2003-10-16 Alexander Sokolsky Touch-sensitive input overlay for graphical user interface
US20060026521A1 (en) * 2004-07-30 2006-02-02 Apple Computer, Inc. Gestures for touch sensitive input devices
US20100088653A1 (en) * 2008-10-07 2010-04-08 Research In Motion Limited Portable electronic device and method of controlling same
US20110148786A1 (en) * 2009-12-18 2011-06-23 Synaptics Incorporated Method and apparatus for changing operating modes
US20120311438A1 (en) * 2010-01-11 2012-12-06 Apple Inc. Electronic text manipulation and display
US20110320978A1 (en) * 2010-06-29 2011-12-29 Horodezky Samuel J Method and apparatus for touchscreen gesture recognition overlay
US20120192117A1 (en) * 2011-01-24 2012-07-26 Migos Charles J Device, Method, and Graphical User Interface with a Dynamic Gesture Disambiguation Threshold
US20120293528A1 (en) * 2011-05-18 2012-11-22 Larsen Eric J Method and apparatus for rendering a paper representation on an electronic display
US20130047115A1 (en) * 2011-08-19 2013-02-21 Apple Inc. Creating and viewing digital note cards

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140184531A1 (en) * 2012-12-27 2014-07-03 Kabushiki Kaisha Toshiba Electronic device, display method, and display program
US9310998B2 (en) * 2012-12-27 2016-04-12 Kabushiki Kaisha Toshiba Electronic device, display method, and display program
US20140237427A1 (en) * 2013-02-20 2014-08-21 Fuji Xerox Co., Ltd. Browsing device, browsing system, and non-transitory computer readable medium
US11150775B2 (en) * 2013-03-14 2021-10-19 Samsung Electronics Co., Ltd. Electronic device and method for controlling screen display using temperature and humidity
US20180300019A1 (en) * 2013-03-14 2018-10-18 Samsung Electronics Co., Ltd. Electronic device and method for controlling screen display using temperature and humidity
US20150095835A1 (en) * 2013-09-30 2015-04-02 Kobo Incorporated Providing a user specific reader mode on an electronic personal display
US20160034090A1 (en) * 2013-11-15 2016-02-04 Boe Technology Group Co., Ltd. Touch system and display device
US9916028B2 (en) * 2013-11-15 2018-03-13 Boe Technology Group Co., Ltd. Touch system and display device for preventing misoperation on edge area
US20170245017A1 (en) * 2014-06-12 2017-08-24 Apple Inc. Systems and Methods for Presenting and Interacting with a Picture-in-Picture Representation of Video Content on an Electronic Device with a Touch-Sensitive Display
US10732820B2 (en) 2014-06-12 2020-08-04 Apple Inc. Systems and methods for efficiently navigating between applications with linked content on an electronic device with a touch-sensitive display
US10795490B2 (en) * 2014-06-12 2020-10-06 Apple Inc. Systems and methods for presenting and interacting with a picture-in-picture representation of video content on an electronic device with a touch-sensitive display
US11592923B2 (en) 2014-06-12 2023-02-28 Apple Inc. Systems and methods for resizing applications in a multitasking view on an electronic device with a touch-sensitive display
US20160012029A1 (en) * 2014-07-09 2016-01-14 Lg Electronics Inc. Mobile terminal and method of controlling the same
US20170124622A1 (en) * 2014-11-14 2017-05-04 The Joan and Irwin Jacobs Technion-Cornell Institute System and method for intuitive content browsing
US10825069B2 (en) * 2014-11-14 2020-11-03 The Joan and Irwin Jacobs Technion-Cornell Institute System and method for intuitive content browsing
US10824987B2 (en) 2014-11-14 2020-11-03 The Joan and Irwin Jacobs Technion-Cornell Institute Techniques for embedding virtual points of sale in electronic media content
US20170123624A1 (en) * 2015-11-02 2017-05-04 Guangzhou Ucweb Computer Technology Co., Ltd. Methods, apparatuses, and devices for processing interface displays
US10416874B2 (en) * 2015-11-02 2019-09-17 Guangzhou Ucweb Computer Technology Co., Ltd. Methods, apparatuses, and devices for processing interface displays
US11966578B2 (en) 2019-05-16 2024-04-23 Apple Inc. Devices and methods for integrating video with user interface navigation

Also Published As

Publication number Publication date
US8963865B2 (en) 2015-02-24

Similar Documents

Publication Publication Date Title
US11320931B2 (en) Swipe-based confirmation for touch sensitive devices
US11204687B2 (en) Visual thumbnail, scrubber for digital content
US10585563B2 (en) Accessible reading mode techniques for electronic devices
US8963865B2 (en) Touch sensitive device with concentration mode
US9030430B2 (en) Multi-touch navigation mode
US9448719B2 (en) Touch sensitive device with pinch-based expand/collapse function
US9146672B2 (en) Multidirectional swipe key for virtual keyboard
US9766723B2 (en) Stylus sensitive device with hover over stylus control functionality
US9134892B2 (en) Drag-based content selection technique for touch screen UI
US9477382B2 (en) Multi-page content selection technique
US9448643B2 (en) Stylus sensitive device with stylus angle detection functionality
US9134893B2 (en) Block-based content selecting technique for touch screen UI
US9367208B2 (en) Move icon to reveal textual information
US9152321B2 (en) Touch sensitive UI technique for duplicating content
US9588979B2 (en) UI techniques for navigating a file manager of an electronic computing device
US20140173529A1 (en) Circular gesture for touch sensitive ui control feature
US20140306897A1 (en) Virtual keyboard swipe gestures for cursor movement
US9423932B2 (en) Zoom view mode for digital content including multiple regions of interest
US10019153B2 (en) Scrapbooking digital content in computing devices using a swiping gesture
US9134903B2 (en) Content selecting technique for touch screen UI
US20140218343A1 (en) Stylus sensitive device with hover over stylus gesture functionality
US20140331187A1 (en) Grouping objects on a computing device
US20140223382A1 (en) Z-shaped gesture for touch sensitive ui undo, delete, and clear functions
US20140380244A1 (en) Visual table of contents for touch sensitive devices

Legal Events

Date Code Title Description
AS Assignment

Owner name: BARNESANDNOBLE.COM LLC, NEW YORK

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HICKS, KOURTNY M.;BREWER, DALE J.;CUETO, GERALD B.;AND OTHERS;REEL/FRAME:029711/0588

Effective date: 20130109

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: NOOK DIGITAL LLC, NEW YORK

Free format text: CHANGE OF NAME;ASSIGNOR:BARNESANDNOBLE.COM LLC;REEL/FRAME:035187/0469

Effective date: 20150225

Owner name: NOOK DIGITAL, LLC, NEW YORK

Free format text: CHANGE OF NAME;ASSIGNOR:NOOK DIGITAL LLC;REEL/FRAME:035187/0476

Effective date: 20150303

AS Assignment

Owner name: BARNES & NOBLE COLLEGE BOOKSELLERS, LLC, NEW JERSEY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NOOK DIGITAL, LLC;REEL/FRAME:035399/0325

Effective date: 20150407

Owner name: BARNES & NOBLE COLLEGE BOOKSELLERS, LLC, NEW JERSE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NOOK DIGITAL, LLC;REEL/FRAME:035399/0325

Effective date: 20150407

AS Assignment

Owner name: NOOK DIGITAL, LLC, NEW YORK

Free format text: CORRECTIVE ASSIGNMENT TO REMOVE APPLICATION NUMBERS 13924129 AND 13924362 PREVIOUSLY RECORDED ON REEL 035187 FRAME 0476. ASSIGNOR(S) HEREBY CONFIRMS THE CHANGE OF NAME;ASSIGNOR:NOOK DIGITAL LLC;REEL/FRAME:036131/0801

Effective date: 20150303

Owner name: NOOK DIGITAL LLC, NEW YORK

Free format text: CORRECTIVE ASSIGNMENT TO REMOVE APPLICATION NUMBERS 13924129 AND 13924362 PREVIOUSLY RECORDED ON REEL 035187 FRAME 0469. ASSIGNOR(S) HEREBY CONFIRMS THE CHANGE OF NAME;ASSIGNOR:BARNESANDNOBLE.COM LLC;REEL/FRAME:036131/0409

Effective date: 20150225

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

AS Assignment

Owner name: BANK OF AMERICA, N.A., AS COLLATERAL AGENT, NORTH CAROLINA

Free format text: SECURITY INTEREST;ASSIGNOR:BARNES & NOBLE COLLEGE BOOKSELELRS, LLC;REEL/FRAME:056784/0426

Effective date: 20210528

AS Assignment

Owner name: TOPLIDS LENDCO, LLC AS COLLATERAL AGENT, MARYLAND

Free format text: SECURITY INTEREST;ASSIGNORS:BARNES & NOBLE COLLEGE BOOKSELLERS, LLC;BNED LOUDCLOUD, LLC;REEL/FRAME:060306/0191

Effective date: 20220607

AS Assignment

Owner name: BANK OF AMERICA, N.A., AS COLLATERAL AGENT, NORTH CAROLINA

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REPLACE THE GRANT OF SECURITY INTEREST IN PATENTS AND SCHEDULE I ATTACHED TO FILING PREVIOUSLY RECORDED AT REEL: 056784 FRAME: 0426. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY INTEREST;ASSIGNOR:BARNES & NOBLE COLLEGE BOOKSELLERS, LLC;REEL/FRAME:060312/0169

Effective date: 20210528

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8