WO2018185716A1 - Method and device for proofreading text - Google Patents

Method and device for proofreading text Download PDF

Info

Publication number
WO2018185716A1
WO2018185716A1 PCT/IB2018/052393 IB2018052393W WO2018185716A1 WO 2018185716 A1 WO2018185716 A1 WO 2018185716A1 IB 2018052393 W IB2018052393 W IB 2018052393W WO 2018185716 A1 WO2018185716 A1 WO 2018185716A1
Authority
WO
WIPO (PCT)
Prior art keywords
smartphone
text
signal
determined
computer
Prior art date
Application number
PCT/IB2018/052393
Other languages
French (fr)
Inventor
Martin Rainer Gabriel Schweiger
Original Assignee
Martin Rainer Gabriel Schweiger
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Martin Rainer Gabriel Schweiger filed Critical Martin Rainer Gabriel Schweiger
Publication of WO2018185716A1 publication Critical patent/WO2018185716A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2200/00Indexing scheme relating to G06F1/04 - G06F1/32
    • G06F2200/16Indexing scheme relating to G06F1/16 - G06F1/18
    • G06F2200/163Indexing scheme relating to constructional details of the computer
    • G06F2200/1636Sensing arrangement for detection of a tap gesture on the housing
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/28Constructional details of speech recognition systems

Definitions

  • This application relates to a method and device for
  • Proofreading text traditionally involves putting two versions of a text side by side and visually checking the respective documents against each other. Another way is to have one person read aloud one version of the text while another person verifies the other version of the text.
  • the present application provides a method for proofreading text through a computer-implemented program.
  • the text that is being proofread is data that contains words and at least one end-of-sentence character.
  • the end-of- sentence character can be a full-stop. This makes for
  • proofreading text can be used.
  • the text data that is being proofread also contains at least one end-of-paragraph character.
  • the end-of-paragraph character can be a carriage return/line feed (CR/LF) sequence. This helps to make the implementation of the computer program more convenient as most of the text that needs proofreading often comes in the form of a word-processed data file with
  • the method for proofreading text includes a step where a computer program provides the text as data which is loaded on a computer.
  • the computer comes with a display, a sensor, and with an audio output.
  • the computer further provides a current cursor position in the text data.
  • the sensor detects one or more signals on or in the vicinity of the computer. Since most of the texts that need to be proofread are usually stored or loaded on a computer, this makes the implementation of the computer program for proofreading text easy and convenient.
  • the cursor position provides a visual indication as to the specific word of the text which is being read or about to be read if the program is not in read mode, This helps the user to identify the part of the text that is being read when he looks up onto the text on the display of the computer.
  • the signal detected by the sensor can be a tapping on the housing of the computer.
  • the tapping signal can be produced by a user with a pointer finger or with a body part or with other objects such as a pen. Recognizing tapping signals creates a simple and versatile way for a user to start using the computer program for proofreading text, as tapping can be easily done with any objects and on any surface, as long as the signal produced can be detected by the sensor.
  • the method for proofreading text also includes a step where, if the computer is in a "not read mode" and if the signal detected on the sensor corresponds with a first pre-determined Morse signal, e.g. "e", the program turns the computer into a read mode, where it outputs the text data through the audio output.
  • the computer program If the computer is in a "read mode", and the first pre-determined Morse signal, e.g. "e” is detected on the sensor, the computer program turns the computer into a "not read mode", where it stops outputting the text data through the audio output .
  • the first pre-determined Morse signal e.g. "e”
  • the first pre-determined Morse signal e.g. "e”
  • the first pre-determined Morse signal includes a single pre-determined low frequency signal.
  • Morse signals are used as they are simple, short, yet very defined signals which can be easily used by the users of the method for proofreading text. With Morse signals, the user can remain focused on proofreading his document while
  • the user can do away with the hassle of looking up to the screen and using the mouse to click on a start/stop button of the proofreader program to start of stop reading the text.
  • the method for proofreading text also includes a step where if the signal corresponds with a second pre-determined Morse signal e.g. "i", the computer program positions the current cursor position to a position immediately after the preceding end-of-sentence character and turns the computer into a read mode, where it outputs the text data through the audio output.
  • a second pre-determined Morse signal e.g. "i”
  • the second pre-determined Morse signal e.g. "i" includes two pre-determined low frequency signals that are spaced within a pre-determined range that is more or less equal to a predetermined characteristic time apart from each other.
  • the preceding end-of-sentence character for the first sentence of the text data is the beginning of the text data. This avoids the situation where the program is unable to repeat the first sentence of the text since there is no preceding sentence for the first sentence for the command e.g. "i", to be executed.
  • the method for proofreading text allows for simple Morse code commands to control the reading of the text. By simply tapping the different commands on the computer, the user can
  • the method for proofreading text can also include a step where if the signal detected on the sensor corresponds with a third pre-determined Morse signal, e.g. "s", the computer program positions the current cursor position to a position
  • the third pre-determined Morse signal includes three pre-determined low frequency signals, wherein the three pre-determined low frequency signal peaks are spaced within a pre-determined range that is more or less equal to a predetermined characteristic time apart from one another.
  • the preceding end-of-paragraph character for the first paragraph of the text data is the beginning of the text data. This avoids the situation where the program is unable to repeat the first paragraph of the text since there is no preceding paragraph for the first paragraph for the command, e.g. "s", to be executed.
  • the text data that is being proofread by the program can refer to text data that is stored on the memory on a mainboard of a computer or stored elsewhere. This allows for easy access to the text files that needs to be proofread.
  • the method of proofreading the text data can also include a step where the computer display highlights the sentence of the text data that is being output through the audio output as the text data is being proofread by the program. This serves to visually direct the user to the sentence that is currently being read by the computer program. Once the user looks to the text on the display, he can immediately know which sentence is being read without having to search through the whole text.
  • the computer display can also further highlight each single word of the text data that is being output through the audio output with a box. This serves to visually direct the user to the word that is currently being read by the computer program. Once the user looks to the text on the display, he can immediately know which specific word is being read without having to search through the whole text.
  • the present application also provides a device that is adapted to perform the method of proofreading text data as described above .
  • the device can include a computer with a display, with a sensor and with an audio output.
  • the computer often has a memory store that is located on a mainboard or located elsewhere. This simple computer device setup can be found on almost every household or workplace. This makes implementation of the method of proofreading text convenient for the user.
  • the audio output on the computer can be a loudspeaker, an earphone jack for connecting with a loudspeaker, or other audio output means .
  • the sensor on the computer can be a key of a keyboard, e.g. a space bar key, a touch pad, a microphone, an acceleration sensor, a switch, a touch-sensitive display or a mouse button
  • a keyboard e.g. a space bar key, a touch pad, a microphone, an acceleration sensor, a switch, a touch-sensitive display or a mouse button
  • the sensor can also be adapte ⁇ . to detect signals such as the tapping on the housing of the computer by the user with a pointed finger or with a body part or with other objects such as a pen.
  • the device that is adapted to perform the method of
  • proofreading text data as described above can also include a smartphone .
  • the smartphone often has a memory on a mainboard or elsewhere, a smartphone display, a smartphone acceleration sensor, a smartphone microphone and a smartphone audio output
  • the smartphone audio output can be a loudspeaker, an earphone jack or other audio output means.
  • the senor can be the smartphone display.
  • the sensor can be a part of the smartphone display, for example the area over an icon displayed on the smartphone display, or a designated area on the smartphone display.
  • the sensor is the entire smartphone display, a user can easily actuate the sensor without diverting his attention towards the device.
  • the present application also provides a method for
  • the smartphone is a computer that has a smartphone display, a smartphone
  • the smartphone display is a touch-sensitive display and acts both as a display and as a sensor.
  • the text is provided as data which is loaded on a smartphone.
  • the smartphone further provides a current cursor position in the text.
  • the smartphone display detects one or more signals from touch input on the display surface.
  • the smartphone display can detect touch input on a specific area of the smartphone display surface only, for example above an icon displayed on the smartphone display or in a designated input area. When the smartphone display detects touch input on the entire
  • a majority of the smartphone display surface can for example be an entire smartphone display surface except for a soft-key navigation bar on the bottom of the smartphone display as used on Android smartphones.
  • a majority of the smartphone display surface can be an entire smartphone display surface except for an exit button.
  • the signal detected by the smartphone display can be a tapping on the smartphone display.
  • the tapping signal can be produced by a user with a pointer finger or with a body part or with other objects such as a pen. Recognizing tapping signals creates a simple and versatile way for a user to start using the smartphone for proofreading text, as tapping can be easily done with any objects and on any surface, as long as the signal produced can be detected by the sensor.
  • the method for proofreading text also includes a step where, if the smartphone is in a "not read mode” and if the signal detected on the sensor corresponds with a first pre-determined Morse signal, e.g. "e", the program turns the smartphone into a read mode, where it outputs the text data through the smartphone audio output. If the smartphone is in a "read mode”, and the first pre-determined Morse signal, e.g. "e” is detected on the sensor, the method turns the smartphone into a "not read mode", where it stops outputting the text data through the smartphone audio output .
  • a first pre-determined Morse signal e.g. "e”
  • the method for proofreading text also includes a step where, if the smartphone receives a character corresponding with a forth pre-determined Morse signal, e.g. "h", the smartphone switches into a speech recognition mode, if it is not in a speech recognition mode. If the smartphone is in a speech recognition mode when it receives a character corresponding with a forth pre-determined Morse signal, e.g. "h", the smartphone switches back into a proofreading mode.
  • vocal input of a user is captured by a smartphone microphone. The smartphone transforms the vocal input into textual data.
  • a user can seamlessly make annotations during the proofreading process and return quickly to the proofreading without diverting his attention.
  • the annotations are stored where they are needed.
  • the present application also provides a smartphone for proofreading text.
  • the smartphone comprises a memory on a mainboard, a touch-sensitive smartphone display, and a smartphone audio output .
  • the smartphone is adapted to perform the method according to present application.
  • the proofreading process does not provide any hurdles, the user is unobstructed and does not need any further devices besides a smartphone .
  • the smartphone can further comprise a smartphone housing and be configured for detecting tapping input on at least a portion of the smartphone housing. This further simplifies the proofreading process. The user does not need to divert his attention toward the smartphone in order to tap the smartphone display .
  • the smartphone further comprises a smartphone microphone.
  • the smartphone is configured to capture vocal input via the microphone for making speech annotations to the proofreading text.
  • the microphone may also be used for sensing taps of a user.
  • Fig. 1 shows a computer with a human user
  • Fig. 2 shows the components of the computer of Fig. 1
  • Fig. 3 shows a close-up view of the display of the computer of Fig. 1
  • Fig. 4 shows a flow diagram for causing text data to be read aloud by the computer of Fig. 1,
  • Fig. 5 shows an extended flow diagram of one routine of the flow diagram of Fig. 3,
  • Fig. 6 shows an amplitude diagram of a single first low frequency signal that is detected by the computer of
  • Fig. 7 shows an amplitude diagram of two first low
  • Fig. 8 shows an amplitude diagram of a second low frequency signal that is detected by the computer of Fig. 1,
  • Fig. 9 shows an amplitude diagram of a third low frequency signal that is detected by the computer of Fig. 1,
  • Fig. 10 shows a smartphone with its components
  • Fig. 11A shows a smartphone for proofreading
  • Fig. 11B shows a smartphone of Fig. 11A, wherein the Morse keyboard module input area is shown
  • Fig. 12 shows a block diagram of a proofreading program for a smartphone .
  • embodiments may be practiced without such details .
  • Some parts of the embodiments have similar parts.
  • the similar parts may have same names or similar part numbers. The description of one part applies by reference to another similar part, where appropriate, thereby reducing repetition of text without limiting the disclosure.
  • Fig. 1 shows a laptop computer 1 with a cover 5 and a housing
  • the laptop computer 1 is a conventional laptop type.
  • the cover 5 comprises a display 6 with the housing 10.
  • the computer 1 comprises various input devices, such as a keyboar
  • the computer 1 is operated by a human being of which only an arm 26 with a pointer finger 27 can be seen.
  • the pointer finger 27 can operate the housing 10, the display 6, the space bar key 12, and the touch pad 13 by tipping them
  • Fig. 2 shows the internal components of the laptop computer 1
  • the housing 10 of the laptop computer 1 comprises a mainboard with a memory 14.
  • the computer 1 also comprises input devices such as a sensor 15 and a microphone 17, as well as output devices such as a loudspeaker 16.
  • Fig. 3 shows a text 30 which is displayed on the display 6.
  • the text 30 starts with a first paragraph 35 which has eight sentences.
  • the first sentence of the first paragraph 35 ends with "... to receive gifts.”
  • the second sentence of the first paragraph 35 ends with "...getting is short lived”.
  • the third sentence of the first paragraph 35 ends with "...to better their lives”.
  • the fourth sentence of the first paragraph 35 ends with "...nourishes your soul”.
  • the fifth sentence of the first paragraph 35 ends with "...valuable than the gift”.
  • the sixth sentence of the first paragraph 35 ends with "...service of others".
  • the seventh sentence of the first paragraph 35 ends with "...earthly limitations”.
  • the eighth and last sentence of the paragraph 35 ends with "... something in return . "
  • a second paragraph 45 of text 30 starts with a first sentence that ends with "... being self-centered .
  • the second sentence of paragraph 45 ends with "...famous example”.
  • the third sentence of paragraph 45 ends with "...to others”.
  • the fourth sentence of paragraph 45 ends with "...herself to others”.
  • the fifth sentence of paragraph 45 ends with "...and serenity”.
  • the sixth and last sentence of paragraph 45 ends with
  • a cursor 46 indicates a position on the text 30 where the computer 1 starts reading.
  • the display 6 highlights a sentence 47, "Truly giving from the heart fills your life with joy and nourishes your soul.”, which comes after the cursor 46.
  • a word "soul” is emphasized as highlighted word 48.
  • the computer 1 reads through the text 30, it emphasizes the highlighted sentence 47 and each highlighted word 48 that is currently being read.
  • the read text is output as audible voice sound through the loudspeaker of the computer 1.
  • Fig. 4 shows a loop structure 55 for causing text 30 to be read aloud through the loudspeaker of the computer 1.
  • the loop structure 55 comprises a detection step 56 which will lead to a first decision 60 or a second decision 65 or a third decision 70 being made, depending on the type of signal detected .
  • a back loop path 75 leads back to the detection step 56 from the outcome of each of the first decision 60, the second decision 65 and the third decision 70, and completes the loop structure 55.
  • the Yes path 62 When at the first decision 60, there will be a No path 61 and a Yes path 62.
  • the Yes path 62 will be activated.
  • the Yes path 62 When activated, the Yes path 62 will trigger the computer 1 to activate a Start or stop reading step 76.
  • the Start or stop reading step 76 When the Start or stop reading step 76 is activated, the computer 1 either stop or start reading the text 30 aloud, depending on whether the computer 1 is in read mode or not, or vice versa.
  • the Yes path 62 Upon completion of the Start or stop reading step 76, the Yes path 62 will continue into a collection point which leads into the back loop path 75.
  • the pre-determined signal is a signal which is also known as the Morse signal for the letter "e", a sharp single peak, which is caused by a fast single tip of the pointer finger 27 on the housing 10.
  • the No path 61 will be activated. The No path 61 will then lead to the second decision 65 being made .
  • the two pre-determined signals that are spaced less than a pre-determined threshold time apart from each other are signals which are also known as the Morse signal for the letter "i", which comprises two consecutive sharp peaks, which are caused by a fast double tip of the pointer finger 27 on the housing 10.
  • the No path 66 When no pre-determined signal is detected by the computer 1 at the second decision 65, the No path 66 will be activated. The No path 66 will then lead to a third decision 70 being made.
  • the Yes path 72 will trigger a Go back to start of current paragraph step 78.
  • the Go back to start of current paragraph step 78 when triggered, causes the cursor 46 to go back to the start of the current paragraph and the computer 1 to start reading through from the start of the current paragraph.
  • the Yes path 72 Upon completion of the Go back to start of current paragraph step 78, the Yes path 72 will continue into a collection point which leads into the back loop path 75.
  • the three pre-determined signals that are spaced less than a pre-determined threshold time apart from one another are signals which are also known as the Morse signal for the letter "s", which comprises three consecutive sharp peaks, which are caused by a fast triple tip of the pointer finger 27 on the housing 10.
  • the No path 71 When no pre-determined signal is detected by the computer 1 at the third decision 70, the No path 71 will be activated. The No path will then continue onto the back loop path 75.
  • the time for carrying out the loop structure 55 must be less than the threshold time between two peaks (for example, at 1/lOOOth of the threshold time), otherwise the detection step 56 may be unreliable.
  • Fig. 5 shows the Start or stop reading step 76 in greater detail.
  • the computer 1 leads into a decision 90.
  • decision 90 there will be a No path 91 and a Yes path 92.
  • Yes path 62 is
  • the Yes path 92 is activated.
  • the computer 1 will initiate a Stop reading step 105, which causes the computer 1 to stop reading the text 30.
  • the Yes path 92 continues into a collection point 100, which leads to the back loop path 75.
  • the computer 1 When the No path 91 is activated, the computer 1 will initiate a Start reading step 95, which causes the computer 1 to start reading the text 30 from the position of the cursor 46. Upon completion of the Start reading step 95, the No path 91 continues into the collection point 100, which leads to the back loop path 75.
  • Fig 6. shows a general representation of a signal which is recorded by the computer 1.
  • the diagram of Fig. 6 comprises an x-axis for the time and y y-axis for the amplitude of the recorded signal.
  • the recorded signal has a 1st peak 115, which is detected by the computer 1 by evaluating the signal-to- noise level of the amplitude.
  • the computer 1 When a pre-determined signal-to-noise level is exceeded and if it immediately afterwards falls below that level, the computer 1 will detect a peak of the amplitude. In other words, there is a pre-determined threshold for discerning an amplitude peak from noise, which is not seen here.
  • the recorded signal can be from any one of the input devices of the computer 1.
  • Fig 7. shows a recorded signal has the 1st peak 115 and a 2nd peak 116.
  • the recorded signal of Fig. 7 is detected by the computer 1 by evaluating the signal-to-noise level of the amplitude .
  • the computer 1 further determines a time difference tN between detecting the 1st peak 115 and detecting the 2nd peak 116.
  • the measurement of Fig. 7 is interpreted as one single predetermined signal "i"
  • the measurement of Fig. 7 is interpreted as two consecutive pre-determined signals "e”.
  • Recorded signal can be from any one of the input devices of the computer 1, and the pre-determined time value tCHAR can be altered by a user of the computer 1, for example by training the computer 1 in a training mode which is not shown here.
  • Fig. 8 shows a recorded signal with a 1st peak 120 and a 2nd peak 121.
  • the recorded signal of Fig. 8 is detected by the computer 1 by evaluating the signal-to-noise level of the amplitude.
  • the 1st peak 120 and the 2nd peak 121 are spaced within a pre-determined range that is more or less equal to a pre-determined characteristic time tCHAR apart from each other. This is interpreted as one single pre-determined signal
  • the pre-determined time range of tolerance defined around each detected peak can also be altered by the user of the computer 1.
  • the measurement of Fig. 8 will be interpreted as an error and it will be ignored. In an alternative embodiment which is not shown here, if the time space between the 1st peak 120 and the 2nd peak 121 is longer than the predetermined range for detection of tCHAR, the measurement o Fig. 8 will be interpreted as one pre-determined signal "e
  • Fig. 9 shows a recorded signal with a 1st peak 125, a 2nd peak 126 and a 3rd peak 127.
  • the recorded signal of Fig. 7 is detected by the computer 1 by evaluating the signal-to-noise level of the amplitude.
  • the 1st peak 125, the 2nd peak 126 and the 3rd peak 127 are spaced within a pre-determined range that is more or less equal to a pre-determined characteristic time tCHAR apart from one another. This is interpreted as one single pre-determined signal "s".
  • the pre-determined range for detection of tCHAR can be altered by a user of the computer 1, for example by training the computer 1 in a training mode which is not shown here.
  • the pre-determined time range of tolerance defined around eac detected peak can also be altered by the user of the computer 1.
  • the measurement of Fig. 9 will be interpreted as two pre-determined signals "e", followed by an error which will be ignored . If the time space between the 1st peak 125 and the 2nd pea 126 is longer than the pre-determined range for detection tCHAR, but the time space between the 2nd peak 126 and 3rd peak 127 is within the pre-determined range for detection tCHAR, the measurement of Fig. 9 will be interpreted as an error and it will be ignored.
  • time space between the 1st peak 125 and the 2nd peak 126 is longer than the pre-determined range for detection of tCHAR, but the time space between the 2nd peak 126 and the 3rd peak 127 is shorter than the pre-determined range for
  • Fig. 10 shows a smartphone 128 comprising a smartphone display 130, which displays the text 30.
  • the smartphone is a
  • the smartphone display 130 is a touch-sensitive display. Therefore, the smartphone display 130 comprises both a display 6 and a sensor 15 as described before.
  • the smartphone 128 also comprises various input devices such as a smartphone microphone 129 and a smartphone acceleration sensor 131, as well as output devices such as a smartphone audio output 132.
  • the smartphone audio output 132 is a loudspeaker .
  • the text 30 is loaded into the memory on the mainboard 14 of the computer 1, and the cursor 46 sits right before the first word of the first sentence of the first paragraph 35.
  • the computer 1 is at detection step 56, waiting for a command from the user 25. Then the user 25 enters a command "e" by tapping the housing 10 of the computer 1 with his pointer finger 27 once, so that the sensor 15 in computer 1, as shown in Fig. 2, detects, in detection step 56 of Fig. 4, a single amplitude peak as shown in Fig. 6. Then the proofreader program enters the yes path 62 in Fig.
  • the proofreader program detects that the computer 1 is not in the read mode (step 90) and follows the no path 91 to initiate the "start reading" step 95.
  • the display 6 highlights the first sentence that is being read aloud, which is not shown here.
  • the display 6 also highlights each single word that is currently being read aloud in the highlighted sentence, with a box. From there, the proofreader program then enters into the collection point 100 and then collection point 79, which leads to the back loop path 75 in Fig. 4 and back into the detection step 56 in Fig. 4, where it waits for a new command.
  • Fig. 3 shows a state of the computer 1 just when the
  • proofreader program reads the last word "soul” of the fourth sentence 47, "Truly giving from the heart fills your life with joy and nourishes your soul.”
  • the fourth sentence 47 is highlighted and that the word 48 "soul” is surrounded by a box which highlights it.
  • the user 25 enters a command "i" by tapping the housing 10 of the computer 1 with his pointer finger 27 twice, so that the sensor 15 in computer 1, as shown in Fig. 2, detects, in detection step 56 of Fig. 4, a double amplitude peak as shown in Fig. 8.
  • the proofreader program enters into the decision step 60, the no path 61, the decision step 65 and then the yes path 67, as shown in Fig. 4, which moves the cursor 46 back to the start of the highlighted fourth sentence 47, "Truly giving from the heart fills your life with joy and nourishes your soul", in step 77.
  • the program then starts reading through the text 30 again starting from the word "Truly". From there, the proofreader program then enters into the collection point 80 which leads to the back loop path 75 and back into the detection step 56, where it waits for a new command.
  • the user 25 enters another command "s" by tapping the housing 10 of the computer 1 with his pointer finger 27 thrice, so that the sensor 15 in computer 1, as shown in Fig. 2, detects, in detection step 56 of Fig. 4, a triple amplitude peak as shown in Fig. 9.
  • the proofreader program enters into the decision step 60, the no path 61, the decision step 65, the no path 66, the decision step 70 and then the yes path 72, as shown in Fig. 4, which moves the cursor 46 back to the start of the first paragraph 35, "We all know how great it feels to receive gifts", in step 78.
  • the program then starts reading through the text 30 again starting from the word "We" in the first sentence of the first paragraph 35. From there, the proofreader program then enters into the collection point 81 which leads to the back loop path 75 and back into the detection step 56, where it waits for a new command.
  • the proofreader program enters into the decision step 60, the no path 61, the decision step 65, the no path 66, the decision step 70, the no step 71, and then the back loop path 75 which leads back to the detection step 56, where the loop cycle repeats again until a new command is detected in decision step 56.
  • the above operation of the proofreader program can also be performed on a smartphone 128, as shown in Fig. 10.
  • the text 30 is loaded into the memory on the mainboard of the smartphone 128, and the cursor 46 sits right before the first word of the first sentence of the first paragraph.
  • the text 30 is then displayed on the smartphone display 130.
  • the smartphone 130 is at the detection step 56 as shown in Fig. 4, waiting for a command from the user 25.
  • the smartphone sensor 131 detects, in detection step 56 of Fig. 4, a single amplitude peak as shown in Fig. 6. Then the proofreader program enters the yes path 62 in Fig.
  • the smartphone display 130 highlights e first sentence that is being read aloud, which is not own here.
  • the smartphone display 130 also highlights each ngle word that is currently being read aloud in the
  • the proofreader ogram then enters into the collection point 100 and then llection point 79, which leads to the back loop path 75 and ck into the detection step 56 in Fig. 4, where it waits for new command.
  • proofreader program on the smartphone 128 continues on to read the second sentence, the third sentence, and then the fourth sentence 47 in the same way as the first sentence: "However, the joy of getting is short-lived. Our lives are richer when we share, and that great inner joy comes from helping others to better their lives . Truly giving from the heart fills your life with joy and nourishes your soul.”
  • proofreader program then enters into the collection point 80 which leads to the back loop path 75 and back into the detection step 56, where it waits for a new command.
  • the proofreader program of the smartphone 128 reads the word "soul” and highlights the word with a box
  • the user 25 enters another command "s" on the smartphone 128 by tapping the smartphone housing 133 with his pointer finger 27 thrice, so that the smartphone sensor 131 detects, in detection step 56 of Fig. 4, a triple amplitude peak as shown in Fig. 9.
  • the proofreader program enters into the decision step 60, the no path 61, the decision step 65, the no path 66, the decision step 70 and then the yes path 72, as shown in Fig. 4, which moves the cursor 46 back to the start of the first paragraph 35, "We all know how great it feels to receive gifts", in step 78.
  • the program then starts reading through the text 30 again starting from the word "We” in the first sentence of the first paragraph 35. From there, the proofreader program then enters into the collection point 81 which leads to the back loop path 75 and back into the detection step 56, where it waits for a new command.
  • the proofreader program enters into the decision step 60, the no path 61, the decision step 65, the no path 66, the decision step 70, the no step 71, and then the back loop path 75 which leads back to the detection step 56, where the loop cycle repeats again until a new command is detected in decision step 56.
  • Fig. 11A shows a smartphone 128 comprising a touch-sensitive smartphone display 130, which displays the text 30.
  • the user is prompted to install a Morse keyboard module.
  • the user can be prompted to install a Morse keyboard module by providing a link to a Morse keyboard application in an app store.
  • the user can be prompted to install a Morse keyboard application included in the proofreader program package.
  • the user can be prompted to install a Morse keyboard module of his choice.
  • the method can comprise checking, whether a Morse keyboard module is already installed, and displaying a prompt only if no Morse keyboard module is installed yet, or only if a particular Morse keyboard module is not installed yet.
  • the proofreader program can be programmed more easily.
  • the proofreader program need not comprise a dedicated Morse signal recognition process but can rather rely on the characters provided by the Morse keyboard module.
  • the Morse keyboard module generates
  • the proofreader program can be more reliable, since the Morse signal recognition process is delegated to a specialized program.
  • the user can comfortably position a cursor in the text 30 by tapping inside the text 30 view displayed in the proofreader program.
  • Fig. 11B shows the smartphone 128 of Fig. 11A.
  • the Morse keyboard module input area 134 is shown to the user once the user taps outside the text 30 view.
  • the Morse keyboard module input area 134 can be shown to the user once the user taps a start/stop button to initiate proofreading. This can be alternatively or additionally to showing the keyboard module input area 134 once the user taps outside the text 30 view.
  • the user can profit from the entire available screen real estate for positioning a cursor in a text 30 loaded into the proofreading program, and profit from the Morse keyboard module as soon as he can make use of it.
  • the Morse keyboard module accepts a sequence of one or more taps by a user onto the smartphone display 130, identifies a Morse signal in the tap sequence and converts the Morse signal to a character.
  • the smartphone display 130 preferably is a capacitive touch display. In this case, the user can tap the display in a designated Morse keyboard module input area 134 with his finger.
  • the smartphone display 130 can be any other touch screen, for example a resistive touch screen, an optical touch screen or an
  • Fig. 12 shows a block diagram of a proofreading program 135 for a smartphone 128.
  • the smartphone 128 comprises a touch scree display 130 and housing 133.
  • the smartphone 128 is a perso 1 information handling device, which can also be a table computer, an organizer, a media player or another computing device
  • a touch screen display comprising a touch screen display, a microphone and a housing.
  • a Morse keyboard module 136 accepts touch taps from the touch screen display 130 of the smartphone 128, identifies Morse signals in the touch taps, converts the Morse signals into characters, and sends the characters to a control program 137.
  • the Morse keyboard module 136 can be a module of the
  • the Morse keyboard module 136 can accept different Morse signal input described herein, for example audio Morse signals provided by taps on a table, acceleration Morse signals provided by taps on a housing 133, or Morse signals provided by presses on a key, a switch, or a button.
  • the Morse keyboard module 136 converts a second pre-determined Morse signal comprising two pre-determined low frequency signals that are spaced within a pre-determined range that is more or less equal to a pre-determined characteristic time apart from each other, into a second character, e.g. "i".
  • the Morse keyboard module 136 converts a third pre-determined Morse signal comprising three pre-determined low frequency signals, wherein the three pre-determined low frequency signal peaks are spaced within a pre-determined range that is more or less equal to a pre-determined characteristic time apart from one another, into a third character, e.g. "s".
  • the Morse keyboard module 136 converts a forth pre-determined Morse signal comprising four pre-determined low frequency signals, wherein the four pre-determined low frequency signal peaks are spaced within a pre-determined range that is more or less equal to a pre-determined characteristic time apart from one another, into a forth character, e.g. "h” .
  • the Morse keyboard module 136 converts a fifth pre-determined Morse signal comprising five pre-determined low frequency signals, wherein the five pre-determined low frequency signal peaks are spaced within a pre-determined range that is more or less equal to a pre-determined characteristic time apart from one another, into a fifth character, e.g. "5".
  • a text module 138 loads a text 30 for proofreading.
  • the text module 138 can be a component of the proofreading application. Text from other applications or from the clipboard can be loaded into the text module 138 for proofreading.
  • the text module 138 can also comprise text editing features. By providing text editing features in a text module 138 component of the proofreading application, the user can write, dictate, or edit the text 30 from within the text module 138, thus accelerating the proofreading process .
  • the text module 138 can be a separate application.
  • the text module 138 can be a word processor, a notes application, an eBook reader, an email application, or any other application used to generate, consume, process, or exchange text.
  • the text module comprises a cursor 46, which can be positioned by a user to designate a start point for the reading output, and which is positioned by the proofreader program to indicate which words are being output by the proofreader program at any given moment.
  • the control module 137 performs the method shown in Fig. 4 in conjunction with Fig. 5. If the control module 137 receives a character corresponding with a first predetermined Morse signal, e.g. "e”, and the program is not in read mode, the control module 137 starts the audio output of reading a text, which has been loaded into the text module. If the control module 137 receives a character "e" and the program is in read mode, it stops the audio output and the cursor position.
  • a character corresponding with a first predetermined Morse signal e.g. "e”
  • the program is not in read mode
  • control module 137 If the control module 137 receives a character corresponding with a second predetermined Morse signal, e.g. "i", the control module 137 positions the current cursor position to position immediately after the preceding end-of-sentence character and turns the control module 137 into a read mode, where it outputs the text data through the audio output.
  • a character corresponding with a second predetermined Morse signal e.g. "i”
  • control module 137 If the control module 137 receives a character corresponding with a third predetermined Morse signal, e.g. "s", the control module 137 positions the current cursor position to a position immediately after the preceding end-of-paragraph character and turns the control module 137 into a read mode, where it outputs the text data through the audio output.
  • a character corresponding with a third predetermined Morse signal e.g. "s”
  • the control module 137 utilizes API functions provided by the smartphone 128 for speech synthesizing. On a smartphone running the iOS operating system, the control module 137 initiates the class ⁇ VSpeechSynthesizer' for outputting audi synthesized from text data, i.e. reading text data. The control module 137 calls the function ⁇ peak' to initiate audio output, the function ⁇ auseSpeaking' to pause the audio output, and the function ⁇ VSpeechSynthesizerDelegate' to track the audio output.
  • control module 137 If the control module 137 receives a character corresponding with a forth pre-determined Morse signal, e.g. "h", the control module 137 switches the computer, or the smartphone into a speech recognition mode, if it is not in a speech recognition mode.
  • vocal input o a user is captured by a smartphone microphone 129.
  • the smartphone transforms the vocal input into textual data.
  • the textual data is input into the proofreading text as an annotation, for example as a comment in a word file, as a comment in a pdf file, or in brackets in a text file.
  • the textual data is input into a note application, for example into an Evernote note or a Google Keep note.
  • the note can comprise a reference to the text, for example a citation of the sentence the cursor is positioned in, a citation of the paragraph the cursor is positioned in, or a link to the text data the cursor is positioned in.
  • Inputting the textual data into a note application can facilitate collection of all remarks in one location, for example when researching something across multiple texts. Further, when the textual data is input into a note
  • the control module 137 utilizes API functions provided by the smartphone 128 for speech recognition.
  • the iOS Speech API allows recognizing live and pre-recorded audio speech.
  • the control module 137 calls the function
  • the control module 137 initiates the clas 'SFSpeechAudioBufferRecognitionRequest' to recognize live audio or in-memory content.
  • the captured vocal input is sent to a speech recognition server for processing.
  • Vocal input is recognized incrementally, so the class may pass text data back to the control module 137 more than once.
  • the annotation is only closed once all text data has been received from the Speech API .
  • API functions for speech recognition makes it easy to implement speech recognition with a high quality.
  • Different speech recognition mechanisms can be used, for example in order to be independent of network availability.
  • control module 137 If the control module 137 receives a character corresponding with a forth pre-determined Morse signal, e.g. "h", the control module 137 switches the computer, or the smartphone back into a proofreading mode, if it is in a speech
  • the proofreading mode is the mode described above, wherein the smartphone 128 is either in a "read mode” outputting audio or in a “not read mode” waiting for a signal for outputting audio.
  • Switching the smartphone 128 back to proofreading mode sets the smartphone 128 to "read mode", wherein the smartphone 128 outputs audio starting at the cursor position.
  • switching the smartphone 128 back to proofreading mode sets the smartphone to "not read mode", wherein the smartphone 128 is ready to receive a further signal as described above.
  • proofreading mode sets the smartphone to "read mode” or “ton read mode", either of which was activated before the speech recognition mode has been engaged.
  • a method for proofreading comprises a different action in reaction to receiving a forth pre-determined Morse signal, e.g. "h" .
  • the method comprises positioning the current cursor position to a position immediately after the end-of-paragraph character preceding the preceding end-of-paragraph character and turning a computer, or a smartphone, into a read mode, where it outputs the text data through the audio output.
  • control module 137 For example, if the method is executed by a control module 137 on a smartphone as described in this application, it would comprise the following: if a control module 137 of a smartphone as described in this application, it would comprise the following: if a control module 137 of a smartphone as described in this application, it would comprise the following: if a control module 137 of a smartphone as described in this application, it would comprise the following: if a control module 137 of a smartphone as described in this application, it would comprise the following: if a control module 137 of a smartphone as described in this application, it would comprise the following: if a control module 137 of a smartphone as described in this application, it would comprise the following: if a control module 137 of a smartphone as described in this application, it would comprise the following: if a control module 137 of a smartphone as described in this application, it would comprise the following: if a control module 137 of a smartphone as described in this application, it would comprise the following: if
  • the control module 137 positions the current cursor position to a position immediately after the end-of-paragraph character preceding the preceding end-of-paragraph character and turns the smartphone 128 into a read mode, where it outputs the text data through the audio output.
  • the method comprises a further step.
  • the method comprises positioning the curr nt cursor position to the beginning of the text data and turning a computer, or a smartphone, into a read mode, where it outputs the text data through the audio output .
  • the method further comprises receiving, from a user, a set of preferences.
  • a user specifies which reaction is assigned to a first pre-determined Morse signal, a second pre- determined Morse signal, a third pre-determined Morse signal, a fourth pre-determined Morse signal, and a fifth pre- determined Morse signal, respectively.
  • a user assigns one of the actions described herein to each or only to some of the pre-determined Morse signals . If a user assigns one of the actions described herein only to some of the pre-determined Morse signals, the risk of mis-interpretation of a user input it reduced.
  • the smartphone audio output 132 is an interface for outputting audio via the smartphone.
  • Examples for such an interface are a 3.5 mm audio jack, an RCA audio connector, a Bluetooth audio interface, a Lightning audio interface, a USB audio interface, a TOSLINK audio interface.
  • An analog audio interface or a digital audio interface in connection with a sound transducer can serve as a smartphone audio output 132.
  • the sound transducer may be a loudspeaker, a headphone, or a bone conduction headphone.
  • the method for proofreading text also comprises presenting advertisements alongside the text view of the proofreading program to the user.
  • advertisements can be presented.
  • the proofreading program can be cheaper to the user.
  • a free version of the proofreading program can be offered featuring advertisements, in contrast to a paid version of the proofreading program without advertisements.
  • advertisements are still prominently displayed to the user each time he looks at the text.
  • the user is presented audio advertisements intermitting the proofreading process.
  • the advertisements are very effective since the user is focused on the audio output of the device and can hardly ignore the advertisements. Further, the advertisements need not disrupt the user' s workflow, since the proofreading program can seamlessly continue the proofreading after an advertisement is finished .

Abstract

The application provides a method (55) for proofreading text. The method (55) for proofreading text includes providing the text as data which is loaded on a computer with a display, a sensor, and an audio output. The computer further includes a current cursor position in the text. The method for proofreading text further includes a step for detecting a signal (56) from the sensor, where if the signal detected corresponds with a first pre-determined Morse signal "e" (60) the computer either starts or stops reading aloud (76) the text beginning from the cursor position, through the audio output, depending on whether the computer is in a read mode or not. The method (55) for proofreading also includes a step where if the signal detected corresponds with a second predetermined Morse signal "i" (65), the computer starts reading aloud text beginning from the end of a preceding sentence (77), through the audio output.

Description

METHOD AND DEVICE FOR PROOFREADING TEXT
This application relates to a method and device for
proofreading text. In particular, the method involves
controlling the reading aloud of the text with Morse signals.
Proofreading text traditionally involves putting two versions of a text side by side and visually checking the respective documents against each other. Another way is to have one person read aloud one version of the text while another person verifies the other version of the text.
There are available text-to-speech programs that can be used to read text documents aloud and aid in the proofreading process. This process is cumbersome for one single person.
It is an object of the application to provide an improved method and device for proofreading text.
The present application provides a method for proofreading text through a computer-implemented program.
The text that is being proofread is data that contains words and at least one end-of-sentence character. The end-of- sentence character can be a full-stop. This makes for
convenient implementation of the computer program, as most of the text that needs proofreading often comes in the form of a word data file with punctuations such as full-stops in between sentences. This means that there is usually no extra step needed to process the text file before the method for
proofreading text can be used.
The text data that is being proofread also contains at least one end-of-paragraph character. The end-of-paragraph character can be a carriage return/line feed (CR/LF) sequence. This helps to make the implementation of the computer program more convenient as most of the text that needs proofreading often comes in the form of a word-processed data file with
paragraphs that end with CR/LF sequence. This means that there is usually no extra step needed to process the text file before the method for proofreading text can be used.
The method for proofreading text includes a step where a computer program provides the text as data which is loaded on a computer. The computer comes with a display, a sensor, and with an audio output. The computer further provides a current cursor position in the text data. The sensor detects one or more signals on or in the vicinity of the computer. Since most of the texts that need to be proofread are usually stored or loaded on a computer, this makes the implementation of the computer program for proofreading text easy and convenient.
The cursor position provides a visual indication as to the specific word of the text which is being read or about to be read if the program is not in read mode, This helps the user to identify the part of the text that is being read when he looks up onto the text on the display of the computer.
The signal detected by the sensor can be a tapping on the housing of the computer. The tapping signal can be produced by a user with a pointer finger or with a body part or with other objects such as a pen. Recognizing tapping signals creates a simple and versatile way for a user to start using the computer program for proofreading text, as tapping can be easily done with any objects and on any surface, as long as the signal produced can be detected by the sensor. The method for proofreading text also includes a step where, if the computer is in a "not read mode" and if the signal detected on the sensor corresponds with a first pre-determined Morse signal, e.g. "e", the program turns the computer into a read mode, where it outputs the text data through the audio output. If the computer is in a "read mode", and the first pre-determined Morse signal, e.g. "e" is detected on the sensor, the computer program turns the computer into a "not read mode", where it stops outputting the text data through the audio output .
The first pre-determined Morse signal, e.g. "e", includes a single pre-determined low frequency signal.
Morse signals are used as they are simple, short, yet very defined signals which can be easily used by the users of the method for proofreading text. With Morse signals, the user can remain focused on proofreading his document while
simultaneously controlling the reading of the text by the computer. In other words, the user can do away with the hassle of looking up to the screen and using the mouse to click on a start/stop button of the proofreader program to start of stop reading the text.
The method for proofreading text also includes a step where if the signal corresponds with a second pre-determined Morse signal e.g. "i", the computer program positions the current cursor position to a position immediately after the preceding end-of-sentence character and turns the computer into a read mode, where it outputs the text data through the audio output.
The second pre-determined Morse signal e.g. "i", includes two pre-determined low frequency signals that are spaced within a pre-determined range that is more or less equal to a predetermined characteristic time apart from each other.
The preceding end-of-sentence character for the first sentence of the text data is the beginning of the text data. This avoids the situation where the program is unable to repeat the first sentence of the text since there is no preceding sentence for the first sentence for the command e.g. "i", to be executed.
The method for proofreading text allows for simple Morse code commands to control the reading of the text. By simply tapping the different commands on the computer, the user can
conveniently start or stop the reading of the text, or command the computer program to re-read sentences. The automated reading of the text by the computer program reduces the need for the user to manually look at the text on the computer screen when proofreading the text against another document. This effectively reduces the user' s time to complete the proofreading work.
The method for proofreading text can also include a step where if the signal detected on the sensor corresponds with a third pre-determined Morse signal, e.g. "s", the computer program positions the current cursor position to a position
immediately after the preceding end-of-paragraph character and turns the computer into a read mode, where it outputs the text data through the audio output.
The third pre-determined Morse signal, e.g. "s", includes three pre-determined low frequency signals, wherein the three pre-determined low frequency signal peaks are spaced within a pre-determined range that is more or less equal to a predetermined characteristic time apart from one another. The preceding end-of-paragraph character for the first paragraph of the text data is the beginning of the text data. This avoids the situation where the program is unable to repeat the first paragraph of the text since there is no preceding paragraph for the first paragraph for the command, e.g. "s", to be executed.
The text data that is being proofread by the program can refer to text data that is stored on the memory on a mainboard of a computer or stored elsewhere. This allows for easy access to the text files that needs to be proofread.
The method of proofreading the text data can also include a step where the computer display highlights the sentence of the text data that is being output through the audio output as the text data is being proofread by the program. This serves to visually direct the user to the sentence that is currently being read by the computer program. Once the user looks to the text on the display, he can immediately know which sentence is being read without having to search through the whole text.
The computer display can also further highlight each single word of the text data that is being output through the audio output with a box. This serves to visually direct the user to the word that is currently being read by the computer program. Once the user looks to the text on the display, he can immediately know which specific word is being read without having to search through the whole text.
The present application also provides a device that is adapted to perform the method of proofreading text data as described above . The device can include a computer with a display, with a sensor and with an audio output. The computer often has a memory store that is located on a mainboard or located elsewhere. This simple computer device setup can be found on almost every household or workplace. This makes implementation of the method of proofreading text convenient for the user.
The audio output on the computer can be a loudspeaker, an earphone jack for connecting with a loudspeaker, or other audio output means .
The sensor on the computer can be a key of a keyboard, e.g. a space bar key, a touch pad, a microphone, an acceleration sensor, a switch, a touch-sensitive display or a mouse button These components are common computer components . This allows the proofreading program to be implemented easily and reduces the barrier to usage the program.
The sensor can also be adapte< . to detect signals such as the tapping on the housing of the computer by the user with a pointed finger or with a body part or with other objects such as a pen.
The device that is adapted to perform the method of
proofreading text data as described above can also include a smartphone . The smartphone often has a memory on a mainboard or elsewhere, a smartphone display, a smartphone acceleration sensor, a smartphone microphone and a smartphone audio output
The smartphone audio output can be a loudspeaker, an earphone jack or other audio output means.
When the device includes a smartphone, the sensor can be the smartphone display. The sensor can be a part of the smartphone display, for example the area over an icon displayed on the smartphone display, or a designated area on the smartphone display. When the sensor is the entire smartphone display, a user can easily actuate the sensor without diverting his attention towards the device.
The present application also provides a method for
proofreading text through a smartphone. The smartphone is a computer that has a smartphone display, a smartphone
microphone and a smartphone audio output. The smartphone display is a touch-sensitive display and acts both as a display and as a sensor. The text is provided as data which is loaded on a smartphone. The smartphone further provides a current cursor position in the text.
The smartphone display detects one or more signals from touch input on the display surface. The smartphone display can detect touch input on a specific area of the smartphone display surface only, for example above an icon displayed on the smartphone display or in a designated input area. When the smartphone display detects touch input on the entire
smartphone display area or a majority of the smartphone display surface, the user can easily actuate the proofreading method without diverting his attention towards the smartphone. A majority of the smartphone display surface can for example be an entire smartphone display surface except for a soft-key navigation bar on the bottom of the smartphone display as used on Android smartphones. As another example, a majority of the smartphone display surface can be an entire smartphone display surface except for an exit button.
The signal detected by the smartphone display can be a tapping on the smartphone display. The tapping signal can be produced by a user with a pointer finger or with a body part or with other objects such as a pen. Recognizing tapping signals creates a simple and versatile way for a user to start using the smartphone for proofreading text, as tapping can be easily done with any objects and on any surface, as long as the signal produced can be detected by the sensor.
The method for proofreading text also includes a step where, if the smartphone is in a "not read mode" and if the signal detected on the sensor corresponds with a first pre-determined Morse signal, e.g. "e", the program turns the smartphone into a read mode, where it outputs the text data through the smartphone audio output. If the smartphone is in a "read mode", and the first pre-determined Morse signal, e.g. "e" is detected on the sensor, the method turns the smartphone into a "not read mode", where it stops outputting the text data through the smartphone audio output .
The method for proofreading text also includes a step where, if the smartphone receives a character corresponding with a forth pre-determined Morse signal, e.g. "h", the smartphone switches into a speech recognition mode, if it is not in a speech recognition mode. If the smartphone is in a speech recognition mode when it receives a character corresponding with a forth pre-determined Morse signal, e.g. "h", the smartphone switches back into a proofreading mode. In a speech recognition mode, vocal input of a user is captured by a smartphone microphone. The smartphone transforms the vocal input into textual data.
A user can seamlessly make annotations during the proofreading process and return quickly to the proofreading without diverting his attention. The annotations are stored where they are needed.
The present application also provides a smartphone for proofreading text. The smartphone comprises a memory on a mainboard, a touch-sensitive smartphone display, and a smartphone audio output . The smartphone is adapted to perform the method according to present application.
Since many users have a smartphone on them in many situations, no extra device is needed and proofreading can be done everywhere .
When the smartphone audio output is a loudspeaker, the proofreading process does not provide any hurdles, the user is unobstructed and does not need any further devices besides a smartphone .
The smartphone can further comprise a smartphone housing and be configured for detecting tapping input on at least a portion of the smartphone housing. This further simplifies the proofreading process. The user does not need to divert his attention toward the smartphone in order to tap the smartphone display .
The smartphone further comprises a smartphone microphone.
Therein, the smartphone is configured to capture vocal input via the microphone for making speech annotations to the proofreading text. The microphone may also be used for sensing taps of a user.
The subject matter of the application is described in greater detail in the accompanying Figures, in which
Fig. 1 shows a computer with a human user,
Fig. 2 shows the components of the computer of Fig. 1, Fig. 3 shows a close-up view of the display of the computer of Fig. 1, Fig. 4 shows a flow diagram for causing text data to be read aloud by the computer of Fig. 1,
Fig. 5 shows an extended flow diagram of one routine of the flow diagram of Fig. 3,
Fig. 6 shows an amplitude diagram of a single first low frequency signal that is detected by the computer of
Fig. 1,
Fig. 7 shows an amplitude diagram of two first low
frequency signals that is detected by the computer of Fig. 1,
Fig. 8 shows an amplitude diagram of a second low frequency signal that is detected by the computer of Fig. 1,
Fig. 9 shows an amplitude diagram of a third low frequency signal that is detected by the computer of Fig. 1,
Fig. 10 shows a smartphone with its components,
Fig. 11A shows a smartphone for proofreading,
Fig. 11B shows a smartphone of Fig. 11A, wherein the Morse keyboard module input area is shown,
Fig. 12 shows a block diagram of a proofreading program for a smartphone .
The application now will be described more fully hereinafter with reference to the accompanying drawings, in which
embodiments of the application are shown. This application may however be embodied in many different forms and should not be construed as limited to the embodiments set forth herein.
Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the application to those skilled in the art.
In the following description, details are provided to describe the embodiments of the present application. It shall be apparent to one skilled in the art, however, that the
embodiments may be practiced without such details . Some parts of the embodiments have similar parts. The similar parts may have same names or similar part numbers. The description of one part applies by reference to another similar part, where appropriate, thereby reducing repetition of text without limiting the disclosure.
It will be understood that when an element is referred to a being "on" another element, it can be directly on the other element or intervening elements may be present therebetween As used herein, the term "and/or" includes any and all combinations of one or more of the associated listed items.
It will be understood that, although the terms first, second, third etc. may be used herein to describe various elements, components, regions, layers, and/or sections, these elements, components, regions, layers, and/or sections should not be limited by these terms. These terms are only used to
distinguish one element, component, region, layer, and/or section from another element, component, region, layer, and/or section .
It will be understood that the elements, components, regions, layers and sections depicted in the figures are not
necessarily drawn to scale.
The terminology used herein is for the purpose of describing particular embodiments only, and is not intended to be limiting of the application. As used herein, the singular forms "a," "an," and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms "comprises" and/or "comprising," or "includes" and/or "including" when used in this specification, specify the presence of stated features, regions integers, steps, operations, elements, and/or components but do not preclude the presence or addition of one o more other features , regions, integers, steps, operations elements, components, and/or groups thereof .
Unless otherwise defined, all terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs . It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the relevant art and the present disclosure, and will not be interpreted in an idealized or overiy formal sense unless expressly so defined herein.
Fig. 1 shows a laptop computer 1 with a cover 5 and a housing
10 that are movably connected with each other by a hinge 20. The laptop computer 1 is a conventional laptop type.
The cover 5 comprises a display 6 with the housing 10. The computer 1 comprises various input devices, such as a keyboar
11 with a space bar key 12, a touch pad 13, and the display 6 which is a touch-sensitive display.
The computer 1 is operated by a human being of which only an arm 26 with a pointer finger 27 can be seen.
The pointer finger 27 can operate the housing 10, the display 6, the space bar key 12, and the touch pad 13 by tipping them
Fig. 2 shows the internal components of the laptop computer 1 The housing 10 of the laptop computer 1 comprises a mainboard with a memory 14. The computer 1 also comprises input devices such as a sensor 15 and a microphone 17, as well as output devices such as a loudspeaker 16.
Fig. 3 shows a text 30 which is displayed on the display 6. The text 30 starts with a first paragraph 35 which has eight sentences. The first sentence of the first paragraph 35 ends with "... to receive gifts." The second sentence of the first paragraph 35 ends with "...getting is short lived". The third sentence of the first paragraph 35 ends with "...to better their lives". The fourth sentence of the first paragraph 35 ends with "...nourishes your soul". The fifth sentence of the first paragraph 35 ends with "...valuable than the gift". The sixth sentence of the first paragraph 35 ends with "...service of others". The seventh sentence of the first paragraph 35 ends with "...earthly limitations". The eighth and last sentence of the paragraph 35 ends with "... something in return . "
A second paragraph 45 of text 30 starts with a first sentence that ends with "... being self-centered . " The second sentence of paragraph 45 ends with "...famous example". The third sentence of paragraph 45 ends with "...to others". The fourth sentence of paragraph 45 ends with "...herself to others". The fifth sentence of paragraph 45 ends with "...and serenity". The sixth and last sentence of paragraph 45 ends with
"...easier to bear".
A cursor 46 indicates a position on the text 30 where the computer 1 starts reading. When reading starts, the display 6 highlights a sentence 47, "Truly giving from the heart fills your life with joy and nourishes your soul.", which comes after the cursor 46. In Fig. 2, a word "soul" is emphasized as highlighted word 48. As the computer 1 reads through the text 30, it emphasizes the highlighted sentence 47 and each highlighted word 48 that is currently being read. The read text is output as audible voice sound through the loudspeaker of the computer 1.
Fig. 4 shows a loop structure 55 for causing text 30 to be read aloud through the loudspeaker of the computer 1. The loop structure 55 comprises a detection step 56 which will lead to a first decision 60 or a second decision 65 or a third decision 70 being made, depending on the type of signal detected .
A back loop path 75 leads back to the detection step 56 from the outcome of each of the first decision 60, the second decision 65 and the third decision 70, and completes the loop structure 55.
When at the first decision 60, there will be a No path 61 and a Yes path 62. When a pre-determined signal is detected by the computer 1 at the first decision 60, the Yes path 62 will be activated. When activated, the Yes path 62 will trigger the computer 1 to activate a Start or stop reading step 76. When the Start or stop reading step 76 is activated, the computer 1 either stop or start reading the text 30 aloud, depending on whether the computer 1 is in read mode or not, or vice versa. Upon completion of the Start or stop reading step 76, the Yes path 62 will continue into a collection point which leads into the back loop path 75. In the present embodiment, the pre-determined signal is a signal which is also known as the Morse signal for the letter "e", a sharp single peak, which is caused by a fast single tip of the pointer finger 27 on the housing 10. When no pre-determined signal is detected by the computer 1 at the first decision 60, the No path 61 will be activated. The No path 61 will then lead to the second decision 65 being made .
When at the second decision 65, there will be a No path 66 and a Yes path 67. When two pre-determined signals that are spaced less than a pre-determined threshold time apart from each other are detected by the computer 1 at the second decision 65, the Yes path 67 will be activated. When activated, the Yes path 67 will trigger a Go back to start of current sentence step 77. The Go back to start of current sentence step 77, when triggered, causes the cursor 46 to go back to the start of the current sentence and the computer 1 to start reading through the same sentence. Upon completion of the Go back to start of current sentence step 77, the Yes path 67 will continue into a collection point which leads into the back loop path 75. In the present embodiment, the two pre-determined signals that are spaced less than a pre-determined threshold time apart from each other are signals which are also known as the Morse signal for the letter "i", which comprises two consecutive sharp peaks, which are caused by a fast double tip of the pointer finger 27 on the housing 10.
When no pre-determined signal is detected by the computer 1 at the second decision 65, the No path 66 will be activated. The No path 66 will then lead to a third decision 70 being made.
When at the third decision 70, there will be a No path 71 and a Yes path 72. When three pre-determined signals that are spaced less than a pre-determined threshold time apart from one another are detected by the computer 1 at the third decision 70, the Yes path 72 will be activated. When
activated, the Yes path 72 will trigger a Go back to start of current paragraph step 78. The Go back to start of current paragraph step 78, when triggered, causes the cursor 46 to go back to the start of the current paragraph and the computer 1 to start reading through from the start of the current paragraph. Upon completion of the Go back to start of current paragraph step 78, the Yes path 72 will continue into a collection point which leads into the back loop path 75.
In the present embodiment, the three pre-determined signals that are spaced less than a pre-determined threshold time apart from one another are signals which are also known as the Morse signal for the letter "s", which comprises three consecutive sharp peaks, which are caused by a fast triple tip of the pointer finger 27 on the housing 10.
When no pre-determined signal is detected by the computer 1 at the third decision 70, the No path 71 will be activated. The No path will then continue onto the back loop path 75.
The time for carrying out the loop structure 55 must be less than the threshold time between two peaks (for example, at 1/lOOOth of the threshold time), otherwise the detection step 56 may be unreliable.
Fig. 5 shows the Start or stop reading step 76 in greater detail. When the Yes path 62 is triggered, the computer 1 leads into a decision 90. When at decision 90, there will be a No path 91 and a Yes path 92. When the Yes path 62 is
triggered - which means that a pre-determined signal "e" has been detected - and if the program is currently in the read mode, the Yes path 92 is activated. When the Yes path 92 is activated, the computer 1 will initiate a Stop reading step 105, which causes the computer 1 to stop reading the text 30. Upon completion of the Stop reading step 105, the Yes path 92 continues into a collection point 100, which leads to the back loop path 75.
When the Yes path 62 is triggered - which means that a predetermined signal "e" has been detected - and if the program is currently not in the read mode, the No path 91 is
activated. When the No path 91 is activated, the computer 1 will initiate a Start reading step 95, which causes the computer 1 to start reading the text 30 from the position of the cursor 46. Upon completion of the Start reading step 95, the No path 91 continues into the collection point 100, which leads to the back loop path 75.
Fig 6. shows a general representation of a signal which is recorded by the computer 1. The diagram of Fig. 6 comprises an x-axis for the time and y y-axis for the amplitude of the recorded signal. The recorded signal has a 1st peak 115, which is detected by the computer 1 by evaluating the signal-to- noise level of the amplitude.
When a pre-determined signal-to-noise level is exceeded and if it immediately afterwards falls below that level, the computer 1 will detect a peak of the amplitude. In other words, there is a pre-determined threshold for discerning an amplitude peak from noise, which is not seen here.
When the single 1st peak 115 is recorded, the measurement of Fig. 6 is interpreted as a single pre-determined signal "e" .
The recorded signal can be from any one of the input devices of the computer 1. Fig 7. shows a recorded signal has the 1st peak 115 and a 2nd peak 116. The recorded signal of Fig. 7 is detected by the computer 1 by evaluating the signal-to-noise level of the amplitude .
The computer 1 further determines a time difference tN between detecting the 1st peak 115 and detecting the 2nd peak 116.
If the time difference tN between the 1st peak 115 and the 2nd peak 116 is less than a pre-determined time value tCHAR, the measurement of Fig. 7 is interpreted as one single predetermined signal "i", while if the time difference tN between the 1st peak 115 and the 2nd peak is more than a predetermined time value tCHAR, the measurement of Fig. 7 is interpreted as two consecutive pre-determined signals "e".
Recorded signal can be from any one of the input devices of the computer 1, and the pre-determined time value tCHAR can be altered by a user of the computer 1, for example by training the computer 1 in a training mode which is not shown here.
Fig. 8 shows a recorded signal with a 1st peak 120 and a 2nd peak 121. The recorded signal of Fig. 8 is detected by the computer 1 by evaluating the signal-to-noise level of the amplitude. The 1st peak 120 and the 2nd peak 121 are spaced within a pre-determined range that is more or less equal to a pre-determined characteristic time tCHAR apart from each other. This is interpreted as one single pre-determined signal
d range for detect
compute:r 1, for ex
raining mode which
ermined time range around each detected peak, so that there is a little leeway for user inputs that are time-wise not absolutely constant. The pre-determined time range of tolerance defined around each detected peak can also be altered by the user of the computer 1.
If the time space between the 1st peak 120 and the 2nd peak 121 is shorter than the pre-determined range for detection of tCHAR, the measurement of Fig. 8 will be interpreted as two consecutive pre-determined signals "e".
If the time space between the 1st peak 120 and the 2nd pea 121 is longer than the pre-determined range for detection tCHAR, the measurement of Fig. 8 will be interpreted as an error and it will be ignored. In an alternative embodiment which is not shown here, if the time space between the 1st peak 120 and the 2nd peak 121 is longer than the predetermined range for detection of tCHAR, the measurement o Fig. 8 will be interpreted as one pre-determined signal "e
Fig. 9 shows a recorded signal with a 1st peak 125, a 2nd peak 126 and a 3rd peak 127. The recorded signal of Fig. 7 is detected by the computer 1 by evaluating the signal-to-noise level of the amplitude. The 1st peak 125, the 2nd peak 126 and the 3rd peak 127 are spaced within a pre-determined range that is more or less equal to a pre-determined characteristic time tCHAR apart from one another. This is interpreted as one single pre-determined signal "s".
The pre-determined range for detection of tCHAR can be altered by a user of the computer 1, for example by training the computer 1 in a training mode which is not shown here. There is also a pre-determined time range of tolerance defined around each detected peak, so that there is a little leeway for user inputs that are time-wise not absolutely constant. The pre-determined time range of tolerance defined around eac detected peak can also be altered by the user of the computer 1.
If the time space between the 1st peak 125 and the 2nd peak 126 is within the pre-determined range for detection of tCHAR but the time space between the 2nd peak 126 and the 3rd peak 127 is shorter than the pre-determined range for detection of tCHAR, the measurement of Fig. 9 will be interpreted as one pre-determined signal "i", followed by one pre-determined signal "e" .
If the time space between the 1st peak 125 and the 2nd peak 126 is within the pre-determined range for detection of tCHAR but the time space between the 2nd peak 126 and the 3rd peak 127 is longer than the pre-determined range for detection of tCHAR, the measurement of Fig. 9 will be interpreted as one pre-determined signal "i", followed by an error which will be ignored .
In an alternative embodiment which is not shown here, if the time space between the 1st peak 125 and the 2nd peak 126 is within the pre-determined range for detection of tCHAR, but the time space between the 2nd peak 126 and the 3rd peak 127 is longer than the pre-determined range for detection of tCHAR, the measurement of Fig. 9 will be interpreted as one pre-determined signal "i", followed by one pre-determined signal "e" .
If the time space between the 1st peak 125 and the 2nd peak 126 is shorter than the pre-determined range for detection of tCHAR, but the time space between the 2nd peak 126 and the 3r peak 127 are spaced within the pre-determined range for detection of tCHAR, the measurement of Fig. 9 will be interpreted as one pre-determined signal "e", followed by one pre-determined signal "i".
If the time spaces between the 1st peak 125 and the 2nd peak 126, and between the 2nd peak 126 and the 3rd peak 127 is shorter than the pre-determined range for detection of tCHAR, the measurement of Fig. 9 will be interpreted as three consecutive pre-determined signals "e".
If the time space between the 1st peak 125 and the 2nd peak 126 is shorter than the pre-determined range for detection of tCHAR, but the time space between the 2nd peak 126 and the 3rd peak 127 is longer than the pre-determined range for detection of tCHAR, the measurement of Fig. 9 will be interpreted as an error and it will be ignored.
In an alternative embodiment which is not shown here, if the time space between the 1st peak 125 and the 2nd peak 126 is shorter than the pre-determined range for detection of tCHAR, but the time space between the 2nd peak 126 and 3rd peak 127 is longer than the pre-determined range for detection of tCHAR, the measurement of Fig. 9 will be interpreted as two pre-determined signals "e", followed by one pre-determined signal "e" .
In an alternative embodiment which is not shown here, if the time space between the 1st peak 125 and the 2nd peak 126 is shorter than the pre-determined range for detection of tCHAR, but the time space between the 2nd peak 126 and 3rd peak 127 is longer than the pre-determined range for detection of tCHAR, the measurement of Fig. 9 will be interpreted as two pre-determined signals "e", followed by an error which will be ignored . If the time space between the 1st peak 125 and the 2nd pea 126 is longer than the pre-determined range for detection tCHAR, but the time space between the 2nd peak 126 and 3rd peak 127 is within the pre-determined range for detection tCHAR, the measurement of Fig. 9 will be interpreted as an error and it will be ignored.
In an alternative embodiment which is not shown here, if the time space between the 1st peak 125 and the 2nd peak 126 is longer than the pre-determined range for detection of tCHAR, but the time space between the 2nd peak 126 and 3rd peak 127 is within the pre-determined range for detection of tCHAR, the measurement of Fig. 9 will be interpreted as one predetermined signal "e", followed by one pre-determined signal
If the time space between the 1st peak 125 and the 2nd peak 126 is longer than the pre-determined range for detection of tCHAR, but the time space between the 2nd peak 126 and the 3rd peak 127 is shorter than the pre-determined range for
detection of tCHAR, the measurement of Fig. 9 will be
interpreted as an error and it will be ignored.
In an alternative embodiment which is not shown here, if the time space between the 1st peak 125 and the 2nd peak 126 is longer than the pre-determined range for detection of tCHAR, but the time space between the 2nd peak 126 and 3rd peak 127 is shorter than the pre-determined range for detection of tCHAR, the measurement of Fig. 9 will be interpreted as three pre-determined signals "e" .
In an alternative embodiment which is not shown here, if the time space between the 1st peak 125 and the 2nd peak 126 is longer than the pre-determined range for detection of tCHAR, but the time space between the 2nd peak 126 and 3rd peak 127 is shorter than the pre-determined range for detection of tCHAR, the measurement of Fig. 9 will be interpreted as two pre-determined signals "e" .
If the time spaces between the 1st peak 125 and the 2nd peak 126, and between the 2nd peak 126 and the 3rd peak 127, are longer than the pre-determined range for detection of tCHAR, the measurement of Fig. 9 will be interpreted as an error and it will be ignored.
In an alternative embodiment which is not shown here, if the time spaces between the 1st peak 125 and the 2nd peak 126, and between the 2nd peak 126 and the 3rd peak 127, are longer than the pre-determined range for detection of tCHAR, the
measurement of Fig. 9 will be interpreted as one predetermined signal "e" or two pre-determined signals "e" or three pre-determined signals "e".
Fig. 10 shows a smartphone 128 comprising a smartphone display 130, which displays the text 30. The smartphone is a
computer 1 with a display 6 and a sensor 15. The smartphone display 130 is a touch-sensitive display. Therefore, the smartphone display 130 comprises both a display 6 and a sensor 15 as described before.
The smartphone 128 also comprises various input devices such as a smartphone microphone 129 and a smartphone acceleration sensor 131, as well as output devices such as a smartphone audio output 132. The smartphone audio output 132 is a loudspeaker .
After starting the proofreader program in an initial state no shown here, the text 30 is loaded into the memory on the mainboard 14 of the computer 1, and the cursor 46 sits right before the first word of the first sentence of the first paragraph 35. The computer 1 is at detection step 56, waiting for a command from the user 25. Then the user 25 enters a command "e" by tapping the housing 10 of the computer 1 with his pointer finger 27 once, so that the sensor 15 in computer 1, as shown in Fig. 2, detects, in detection step 56 of Fig. 4, a single amplitude peak as shown in Fig. 6. Then the proofreader program enters the yes path 62 in Fig. 4 and turns the computer 1 into the read mode, where it reads aloud the first word of the first sentence after the cursor 46, through the loudspeaker 16: "We all know how great it feels to receive gifts." The turning on of the read mode is shown in Fig. 5, where the proofreader program detects that the computer 1 is not in the read mode (step 90) and follows the no path 91 to initiate the "start reading" step 95. At the same time, the display 6 highlights the first sentence that is being read aloud, which is not shown here. The display 6 also highlights each single word that is currently being read aloud in the highlighted sentence, with a box. From there, the proofreader program then enters into the collection point 100 and then collection point 79, which leads to the back loop path 75 in Fig. 4 and back into the detection step 56 in Fig. 4, where it waits for a new command.
After having read the first sentence of text 30, the
proofreader program continues on to read the second sentence, the third sentence, and then the fourth sentence 47 in the same way as the first sentence: "However, the joy of getting is short-lived. Our lives are richer when we share, and that great inner joy comes from helping others to better their lives. Truly giving from the heart fills your life with joy and nourishes your soul." Fig. 3 shows a state of the computer 1 just when the
proofreader program reads the last word "soul" of the fourth sentence 47, "Truly giving from the heart fills your life with joy and nourishes your soul." One can see that the fourth sentence 47 is highlighted and that the word 48 "soul" is surrounded by a box which highlights it.
At this time, the user 25 enters a command "i" by tapping the housing 10 of the computer 1 with his pointer finger 27 twice, so that the sensor 15 in computer 1, as shown in Fig. 2, detects, in detection step 56 of Fig. 4, a double amplitude peak as shown in Fig. 8. Then the proofreader program enters into the decision step 60, the no path 61, the decision step 65 and then the yes path 67, as shown in Fig. 4, which moves the cursor 46 back to the start of the highlighted fourth sentence 47, "Truly giving from the heart fills your life with joy and nourishes your soul", in step 77. The program then starts reading through the text 30 again starting from the word "Truly". From there, the proofreader program then enters into the collection point 80 which leads to the back loop path 75 and back into the detection step 56, where it waits for a new command.
As the proofreader program reads the word "soul" and
highlights the word with a box, the user 25 enters another command "s" by tapping the housing 10 of the computer 1 with his pointer finger 27 thrice, so that the sensor 15 in computer 1, as shown in Fig. 2, detects, in detection step 56 of Fig. 4, a triple amplitude peak as shown in Fig. 9. Then the proofreader program enters into the decision step 60, the no path 61, the decision step 65, the no path 66, the decision step 70 and then the yes path 72, as shown in Fig. 4, which moves the cursor 46 back to the start of the first paragraph 35, "We all know how great it feels to receive gifts", in step 78. The program then starts reading through the text 30 again starting from the word "We" in the first sentence of the first paragraph 35. From there, the proofreader program then enters into the collection point 81 which leads to the back loop path 75 and back into the detection step 56, where it waits for a new command.
When no new commands are given and detected at the detection step 56 in Fig. 4, the proofreader program enters into the decision step 60, the no path 61, the decision step 65, the no path 66, the decision step 70, the no step 71, and then the back loop path 75 which leads back to the detection step 56, where the loop cycle repeats again until a new command is detected in decision step 56.
In an alternative embodiment, the above operation of the proofreader program can also be performed on a smartphone 128, as shown in Fig. 10.
Similar to the operation on computer 1, after starting the proofreader program on the smartphone 128 in an initial state not shown here, the text 30 is loaded into the memory on the mainboard of the smartphone 128, and the cursor 46 sits right before the first word of the first sentence of the first paragraph. The text 30 is then displayed on the smartphone display 130. The smartphone 130 is at the detection step 56 as shown in Fig. 4, waiting for a command from the user 25. When the user 25 enters a command "e" on the smartphone 128 by tapping the smartphone housing 133 with his pointer finger 27 once, the smartphone sensor 131 detects, in detection step 56 of Fig. 4, a single amplitude peak as shown in Fig. 6. Then the proofreader program enters the yes path 62 in Fig. 4 and turns the smartphone 128 into the read mode, where it reads aloud the first word of the 1st sentence after the cursor 46, through the smartphone loudspeaker 132 : "We all know how great it feels to receive gifts." The turning on of the read mode is shown in Fig. 5, where the proof reader program detects that the smartphone 128 is not in the read mode (step 90) and follows the no path 91 to initiate the "start reading" step 95. At the same time, the smartphone display 130 highlights e first sentence that is being read aloud, which is not own here. The smartphone display 130 also highlights each ngle word that is currently being read aloud in the
ghlighted sentence, with a box. From there, the proofreader ogram then enters into the collection point 100 and then llection point 79, which leads to the back loop path 75 and ck into the detection step 56 in Fig. 4, where it waits for new command.
After having read the first sentence of text 30, the
proofreader program on the smartphone 128 continues on to read the second sentence, the third sentence, and then the fourth sentence 47 in the same way as the first sentence: "However, the joy of getting is short-lived. Our lives are richer when we share, and that great inner joy comes from helping others to better their lives . Truly giving from the heart fills your life with joy and nourishes your soul."
Similar to the operation on computer 1, when the user 25 enters a command "i" on the smartphone 128 by tapping the smartphone housing 133 with his pointer finger 27 twice, so that the smartphone sensor 131, detects, in detection step 56 of Fig. 4, a double amplitude peak as shown in Fig. 8. Then the proofreader program enters into the decision step 60, the no path 61, the decision step 65 and then the yes path 67, as shown in Fig. 4, which moves the cursor 46 back to the start of the highlighted fourth sentence 47, "Truly giving from the heart fills your life with joy and nourishes your soul", in step 77. The program then starts reading through the text 30 again starting from the word "Truly". From there, the
proofreader program then enters into the collection point 80 which leads to the back loop path 75 and back into the detection step 56, where it waits for a new command.
As the proofreader program of the smartphone 128 reads the word "soul" and highlights the word with a box, the user 25 enters another command "s" on the smartphone 128 by tapping the smartphone housing 133 with his pointer finger 27 thrice, so that the smartphone sensor 131 detects, in detection step 56 of Fig. 4, a triple amplitude peak as shown in Fig. 9. Then the proofreader program enters into the decision step 60, the no path 61, the decision step 65, the no path 66, the decision step 70 and then the yes path 72, as shown in Fig. 4, which moves the cursor 46 back to the start of the first paragraph 35, "We all know how great it feels to receive gifts", in step 78. The program then starts reading through the text 30 again starting from the word "We" in the first sentence of the first paragraph 35. From there, the proofreader program then enters into the collection point 81 which leads to the back loop path 75 and back into the detection step 56, where it waits for a new command. When no new commands are given and detected on the smartphone 128 at the detection step 56 in Fig. 4, the proofreader program enters into the decision step 60, the no path 61, the decision step 65, the no path 66, the decision step 70, the no step 71, and then the back loop path 75 which leads back to the detection step 56, where the loop cycle repeats again until a new command is detected in decision step 56.
Fig. 11A shows a smartphone 128 comprising a touch-sensitive smartphone display 130, which displays the text 30. After starting the proofreader program in an initial state not shown here, the user is prompted to install a Morse keyboard module. For example, the user can be prompted to install a Morse keyboard module by providing a link to a Morse keyboard application in an app store. Alternatively, the user can be prompted to install a Morse keyboard application included in the proofreader program package. Further alternatively, the user can be prompted to install a Morse keyboard module of his choice. The method can comprise checking, whether a Morse keyboard module is already installed, and displaying a prompt only if no Morse keyboard module is installed yet, or only if a particular Morse keyboard module is not installed yet. By prompting a user to install a Morse keyboard module on his smartphone 128, the proofreader program can be programmed more easily. In particular, the proofreader program need not comprise a dedicated Morse signal recognition process but can rather rely on the characters provided by the Morse keyboard module. Therein, the Morse keyboard module generates
characters from Morse signals provided by the user. Further, by prompting a user to install a Morse keyboard module, the proofreader program can be more reliable, since the Morse signal recognition process is delegated to a specialized program.
After a text 30 is loaded into the proofreader program, a Morse keyboard module input area is hidden from the user.
Thus, the user can comfortably position a cursor in the text 30 by tapping inside the text 30 view displayed in the proofreader program.
Fig. 11B shows the smartphone 128 of Fig. 11A. Therein, the Morse keyboard module input area 134 is shown to the user once the user taps outside the text 30 view. In another embodiment, which is not depicted in the figures, the Morse keyboard module input area 134 can be shown to the user once the user taps a start/stop button to initiate proofreading. This can be alternatively or additionally to showing the keyboard module input area 134 once the user taps outside the text 30 view.
By hiding the Morse keyboard module input area 134 initially and showing the Morse keyboard module input area 134 upon tapping outside the text 30 view or tapping a start/stop button, the user can profit from the entire available screen real estate for positioning a cursor in a text 30 loaded into the proofreading program, and profit from the Morse keyboard module as soon as he can make use of it.
The Morse keyboard module accepts a sequence of one or more taps by a user onto the smartphone display 130, identifies a Morse signal in the tap sequence and converts the Morse signal to a character. Therein, the smartphone display 130 preferably is a capacitive touch display. In this case, the user can tap the display in a designated Morse keyboard module input area 134 with his finger. Alternatively, the smartphone display 130 can be any other touch screen, for example a resistive touch screen, an optical touch screen or an
inductive touch screen.
Fig. 12 shows a block diagram of a proofreading program 135 for a smartphone 128.
The smartphone 128 comprises a touch scree display 130 and housing 133. The smartphone 128 is a perso 1 information handling device, which can also be a table computer, an organizer, a media player or another computing device
comprising a touch screen display, a microphone and a housing.
A Morse keyboard module 136 accepts touch taps from the touch screen display 130 of the smartphone 128, identifies Morse signals in the touch taps, converts the Morse signals into characters, and sends the characters to a control program 137. The Morse keyboard module 136 can be a module of the
proofreading program 135, or a separate program on the same device. Additionally, the Morse keyboard module 136 can accept different Morse signal input described herein, for example audio Morse signals provided by taps on a table, acceleration Morse signals provided by taps on a housing 133, or Morse signals provided by presses on a key, a switch, or a button.
The Morse keyboard module 136 converts a second pre-determined Morse signal comprising two pre-determined low frequency signals that are spaced within a pre-determined range that is more or less equal to a pre-determined characteristic time apart from each other, into a second character, e.g. "i".
The Morse keyboard module 136 converts a third pre-determined Morse signal comprising three pre-determined low frequency signals, wherein the three pre-determined low frequency signal peaks are spaced within a pre-determined range that is more or less equal to a pre-determined characteristic time apart from one another, into a third character, e.g. "s".
The Morse keyboard module 136 converts a forth pre-determined Morse signal comprising four pre-determined low frequency signals, wherein the four pre-determined low frequency signal peaks are spaced within a pre-determined range that is more or less equal to a pre-determined characteristic time apart from one another, into a forth character, e.g. "h" . The Morse keyboard module 136 converts a fifth pre-determined Morse signal comprising five pre-determined low frequency signals, wherein the five pre-determined low frequency signal peaks are spaced within a pre-determined range that is more or less equal to a pre-determined characteristic time apart from one another, into a fifth character, e.g. "5".
A text module 138 loads a text 30 for proofreading. The text module 138 can be a component of the proofreading application. Text from other applications or from the clipboard can be loaded into the text module 138 for proofreading. The text module 138 can also comprise text editing features. By providing text editing features in a text module 138 component of the proofreading application, the user can write, dictate, or edit the text 30 from within the text module 138, thus accelerating the proofreading process .
Alternatively or additionally, the text module 138 can be a separate application. For example, the text module 138 can be a word processor, a notes application, an eBook reader, an email application, or any other application used to generate, consume, process, or exchange text. The text module comprises a cursor 46, which can be positioned by a user to designate a start point for the reading output, and which is positioned by the proofreader program to indicate which words are being output by the proofreader program at any given moment.
The control module 137 performs the method shown in Fig. 4 in conjunction with Fig. 5. If the control module 137 receives a character corresponding with a first predetermined Morse signal, e.g. "e", and the program is not in read mode, the control module 137 starts the audio output of reading a text, which has been loaded into the text module. If the control module 137 receives a character "e" and the program is in read mode, it stops the audio output and the cursor position.
If the control module 137 receives a character corresponding with a second predetermined Morse signal, e.g. "i", the control module 137 positions the current cursor position to position immediately after the preceding end-of-sentence character and turns the control module 137 into a read mode, where it outputs the text data through the audio output.
If the control module 137 receives a character corresponding with a third predetermined Morse signal, e.g. "s", the control module 137 positions the current cursor position to a position immediately after the preceding end-of-paragraph character and turns the control module 137 into a read mode, where it outputs the text data through the audio output.
The control module 137 utilizes API functions provided by the smartphone 128 for speech synthesizing. On a smartphone running the iOS operating system, the control module 137 initiates the class ^VSpeechSynthesizer' for outputting audi synthesized from text data, i.e. reading text data. The control module 137 calls the function ^peak' to initiate audio output, the function ^auseSpeaking' to pause the audio output, and the function ^VSpeechSynthesizerDelegate' to track the audio output.
If the control module 137 receives a character corresponding with a forth pre-determined Morse signal, e.g. "h", the control module 137 switches the computer, or the smartphone into a speech recognition mode, if it is not in a speech recognition mode. In a speech recognition mode, vocal input o a user is captured by a smartphone microphone 129. The smartphone transforms the vocal input into textual data. The textual data is input into the proofreading text as an annotation, for example as a comment in a word file, as a comment in a pdf file, or in brackets in a text file.
When the textual data is input into the proofreading text, remarks which come to mind during proofreading can quickly an seamlessly be kept together with the original text for later reference. For example, such remarks can be useful when re- editing the text at a later point in time.
In another example, the textual data is input into a note application, for example into an Evernote note or a Google Keep note. The note can comprise a reference to the text, for example a citation of the sentence the cursor is positioned in, a citation of the paragraph the cursor is positioned in, or a link to the text data the cursor is positioned in.
Inputting the textual data into a note application can facilitate collection of all remarks in one location, for example when researching something across multiple texts. Further, when the textual data is input into a note
application, it will likely be available a long period of time, for example one year, after the proofreading process.
The control module 137 utilizes API functions provided by the smartphone 128 for speech recognition. For example, the iOS Speech API allows recognizing live and pre-recorded audio speech. The control module 137 calls the function
'requestAuthorization' to request the user's permission for speech recognition. The control module 137 initiates the clas 'SFSpeechAudioBufferRecognitionRequest' to recognize live audio or in-memory content. The captured vocal input is sent to a speech recognition server for processing. Vocal input is recognized incrementally, so the class may pass text data back to the control module 137 more than once. The annotation is only closed once all text data has been received from the Speech API .
Using API functions for speech recognition makes it easy to implement speech recognition with a high quality.
Different speech recognition mechanisms can be used, for example in order to be independent of network availability.
If the control module 137 receives a character corresponding with a forth pre-determined Morse signal, e.g. "h", the control module 137 switches the computer, or the smartphone back into a proofreading mode, if it is in a speech
recognition mode. The proofreading mode is the mode described above, wherein the smartphone 128 is either in a "read mode" outputting audio or in a "not read mode" waiting for a signal for outputting audio. Switching the smartphone 128 back to proofreading mode sets the smartphone 128 to "read mode", wherein the smartphone 128 outputs audio starting at the cursor position. In an alternative embodiment which is not shown here, switching the smartphone 128 back to proofreading mode sets the smartphone to "not read mode", wherein the smartphone 128 is ready to receive a further signal as described above. In another alternative embodiment which is not shown here, switching the smartphone 128 back to
proofreading mode sets the smartphone to "read mode" or "ton read mode", either of which was activated before the speech recognition mode has been engaged.
A user can seamlessly make an annotation via speech input and proceed with the proofreading process quickly. In another embodiment which is not shown here, a method for proofreading comprises a different action in reaction to receiving a forth pre-determined Morse signal, e.g. "h" . In reaction to receiving the forth pre-determined Morse signal, the method comprises positioning the current cursor position to a position immediately after the end-of-paragraph character preceding the preceding end-of-paragraph character and turning a computer, or a smartphone, into a read mode, where it outputs the text data through the audio output.
If the method is executed by a program on a computer as described in this application, it would comprise the
following: if a signal received on a sensor corresponds with a forth pre-determined Morse signal, e.g. "h", the program positions the current cursor position to a position
immediately after the end-of-paragraph character preceding the preceding end-of-paragraph character and turns the computer 1 into a read mode, where it outputs the text data through the audio output .
For example, if the method is executed by a control module 137 on a smartphone as described in this application, it would comprise the following: if a control module 137 of a
smartphone receives the forth pre-determined Morse signal, the control module 137 positions the current cursor position to a position immediately after the end-of-paragraph character preceding the preceding end-of-paragraph character and turns the smartphone 128 into a read mode, where it outputs the text data through the audio output.
The method comprises a further step. In reaction to receiving the fifth pre-determined Morse signal, e.g. "5", the method comprises positioning the curr nt cursor position to the beginning of the text data and turning a computer, or a smartphone, into a read mode, where it outputs the text data through the audio output .
The method further comprises receiving, from a user, a set of preferences. Therein, a user specifies which reaction is assigned to a first pre-determined Morse signal, a second pre- determined Morse signal, a third pre-determined Morse signal, a fourth pre-determined Morse signal, and a fifth pre- determined Morse signal, respectively. A user assigns one of the actions described herein to each or only to some of the pre-determined Morse signals . If a user assigns one of the actions described herein only to some of the pre-determined Morse signals, the risk of mis-interpretation of a user input it reduced.
In another embodiment which is not shown here, the smartphone audio output 132 is an interface for outputting audio via the smartphone. Examples for such an interface are a 3.5 mm audio jack, an RCA audio connector, a Bluetooth audio interface, a Lightning audio interface, a USB audio interface, a TOSLINK audio interface. An analog audio interface or a digital audio interface in connection with a sound transducer can serve as a smartphone audio output 132. The sound transducer may be a loudspeaker, a headphone, or a bone conduction headphone.
In an alternative embodiment which is not shown here, the method for proofreading text also comprises presenting advertisements alongside the text view of the proofreading program to the user. Therein, text advertisements, image advertisements, or animated advertisements, in particular looping video advertisements, can be presented. By presenting advertisements, the proofreading program can be cheaper to the user. For example, a free version of the proofreading program can be offered featuring advertisements, in contrast to a paid version of the proofreading program without advertisements. By presenting advertisements alongside the text view, the user is not disturbed in the proofreading process, while the
advertisements are still prominently displayed to the user each time he looks at the text.
In another embodiment not shown in the figures, the user is presented audio advertisements intermitting the proofreading process. By presenting audio advertisements to the user, the advertisements are very effective since the user is focused on the audio output of the device and can hardly ignore the advertisements. Further, the advertisements need not disrupt the user' s workflow, since the proofreading program can seamlessly continue the proofreading after an advertisement is finished .
Reference numbers
I computer
5 cover
6 display
10 housing
II keyboard
12 space bar key
13 touch pad
14 mainboard with memory
15 sensor
16 audio output
17 microphone
20 hinge
26 arm
27 pointer finger
30 text
35 first paragraph
45 second paragraph
46 cursor
47 highlighted sentence "truly giving from the heart fills your life with joy and nourishes your soul."
48 highlighted word "soul"
55 loop structure
56 detection step
60 decision step
61 no path
62 yes path
65 decision step
66 no path
67 yes path
70 decision step
75 back loop path
76 start or stop reading step 77 go back to start of current sentence step
78 go back to start of current paragraph step
79 collection point
80 collection point
81 collection point
90 decision step
91 no path
92 yes path
95 start reading step
100 collection point
105 stop reading step
115 first peak
116 second peak
120 first peak
121 second peak
125 first peak
126 second peak
127 third peak
128 smartphone
129 smartphone microphone
130 smartphone display
131 smartphone acceleration sensor
132 smartphone audio output
133 smartphone housing
134 Morse keyboard module input area
135 proofreading program
136 Morse keyboard module
137 text module
138 control module

Claims

A method for proofreading text (30), which contains words and at least one end-of-sentence character, the method comprising :
providing the text (30) as data which is loaded on a smartphone (128) with a touch-sensitive smartphone display (130), and with a smartphone audio output (132), the smartphone (128) further providing a current cursor position in the text (30) ,
detecting a signal from the smartphone
display (130), wherein
if the signal corresponds with a first predetermined Morse signal ("e"), turning the smartphone (128) into a read mode, where it outputs the text (30) through the smartphone audio output (132), if the smartphone (128) is in a "not read mode", or turning the smartphone (128) into a not read mode, where it stops outputting the text (30) through the smartphone audio output (132), if the smartphone (128) is in a "read mode",
if the signal corresponds with a second predetermined Morse signal ("i"), positioning the current cursor position to a position immediately after the preceding end-of-sentence character and turning the smartphone (128) into a read mode, where it outputs the text (30) through the smartphone audio output (132) .
The method of claim 1, wherein the text comprises at least one end-of-paragraph character, the method further comprising the following step: if the signal corresponds with a third predetermined Morse signal ("s"), positioning the current cursor position to a position immediately after the preceding end-of-paragraph character and turning the smartphone (128) into a read mode, where it outputs the text (30) through the smartphone audio output (132) .
3. The method of claim 1, where in the text (30) is stored on a memory of the smartphone (128) .
The method of claim 1, wherein the smartphone
display (130) highlights the sentence of the text (30) that is being output through the audio output.
The method of claim 1, wherein the smartphone
display (130) further highlights with a box, each single word of the text (30) that is being output through the audio output .
The method of claim 1, wherein the first pre-determined Morse signal ("e") comprises a single pre-determined low frequency signal.
The method of claim 1, wherein the second pre-determined Morse signal ("i") comprises two pre-determined low frequency signals that are spaced within a pre-determined range that is more or less equal to a pre-determined characteristic time apart from each other.
The method of claim 2, wherein the third pre-determined Morse signal ("s") comprises three pre-determined low frequency signals, wherein the three pre-determined low frequency signal peaks are spaced within a pre-determined range that is more or less equal to a pre-determined characteristic time apart from one another.
9. The method of any one of claims 1 to 8, wherein the
detecting of a signal from the smartphone display (130) comprises detecting a touch input anywhere on the smartphone display.
The method of any one of claims 1 to 9, wherein the detecting of a signal from the smartphone display (130) comprises detecting a tapping input on at least a portion of a smartphone housing (133) of the smartphone (128) .
11. The method of any one of claims 1 to 10, the method
further comprising the following step:
if the signal corresponds with a fourth predetermined Morse signal ("h"), switching the
smartphone (128) into a speech recognition mode, if it is not in a speech recognition mode, and switching the smartphone back into a proofreading mode, if it is in a speech recognition mode.
A smartphone for proofreading text (30), the smartphon comprising a memory on a mainboard, a touch-sensitive smartphone display (130), and a smartphone audio
output (132), wherein the smartphone is adapted to perform the method according to one of the preceding claims .
13. The smartphone of claim 12, wherein the smartphone audio output (132) is a loudspeaker.
The smartphone of claim 12 or claim 13, wherein the smartphone (128) further comprises a smartphone housing (133) and wherein the smartphone (128) is configured for detecting tapping input on at least a portion of the smartphone housing (133).
The smartphone of any of the claims 12 to 14, wherein th smartphone (128) further comprises a smartphone
microphone (129) and wherein the smartphone is configure to capture vocal input via the microphone (129) .
PCT/IB2018/052393 2017-04-06 2018-04-06 Method and device for proofreading text WO2018185716A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
SG10201702841T 2017-04-06
SG10201702841T 2017-04-06

Publications (1)

Publication Number Publication Date
WO2018185716A1 true WO2018185716A1 (en) 2018-10-11

Family

ID=63712109

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2018/052393 WO2018185716A1 (en) 2017-04-06 2018-04-06 Method and device for proofreading text

Country Status (1)

Country Link
WO (1) WO2018185716A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020165013A1 (en) * 2001-01-22 2002-11-07 Bright Walter G. Wireless device with vibrational communication capabilities
US20090215494A1 (en) * 2008-02-27 2009-08-27 At&T Mobility Ii Llc Blind Dialing
US20110053577A1 (en) * 2009-08-31 2011-03-03 Lee Changkee Methods and apparatus for communicating by vibrating or moving mobile devices
US8000729B1 (en) * 2008-09-04 2011-08-16 Sprint Communications Company L.P. Mobile to mobile text messaging using keyed modulation

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020165013A1 (en) * 2001-01-22 2002-11-07 Bright Walter G. Wireless device with vibrational communication capabilities
US20090215494A1 (en) * 2008-02-27 2009-08-27 At&T Mobility Ii Llc Blind Dialing
US8000729B1 (en) * 2008-09-04 2011-08-16 Sprint Communications Company L.P. Mobile to mobile text messaging using keyed modulation
US20110053577A1 (en) * 2009-08-31 2011-03-03 Lee Changkee Methods and apparatus for communicating by vibrating or moving mobile devices

Similar Documents

Publication Publication Date Title
AU2017234428B2 (en) Identification of voice inputs providing credentials
US8160881B2 (en) Human-assisted pronunciation generation
JP5166255B2 (en) Data entry system
US8150699B2 (en) Systems and methods of a structured grammar for a speech recognition command system
US20170263248A1 (en) Dictation that allows editing
US9971495B2 (en) Context based gesture delineation for user interaction in eyes-free mode
TW480474B (en) Method and apparatus for presenting proximal feedback in voice command systems
JP5829996B2 (en) Device, method and graphical user interface for setting a restricted interaction with a user interface
US8538757B2 (en) System and method of a list commands utility for a speech recognition command system
JP4667138B2 (en) Speech recognition method and speech recognition apparatus
US20140015782A1 (en) Method for transmitting and receiving data between memo layer and application and electronic device using the same
US20120299858A1 (en) System and method for audibly presenting selected text
JP2006048628A (en) Multimodal input method
EP2835798B1 (en) Interfacing device and method for supporting speech dialogue service
US20150039318A1 (en) Apparatus and method for selecting control object through voice recognition
WO2019205552A1 (en) Online document commenting method and apparatus
Ketabdar et al. MagiWrite: towards touchless digit entry using 3D space around mobile devices
WO2017176513A1 (en) Generating and rendering inflected text
CN105684012B (en) Providing contextual information
US20140180698A1 (en) Information processing apparatus, information processing method and storage medium
KR102283360B1 (en) Method, apparatus and recovering medium for guiding of text edit position
KR101447879B1 (en) Apparatus and method for selecting a control object by voice recognition
WO2018185716A1 (en) Method and device for proofreading text
WO2015156011A1 (en) Information processing device, information processing method, and program
US9965966B2 (en) Instructions on a wearable device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18780539

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18780539

Country of ref document: EP

Kind code of ref document: A1