US20130263037A1 - Song writing method and apparatus using touch screen in mobile terminal - Google Patents

Song writing method and apparatus using touch screen in mobile terminal Download PDF

Info

Publication number
US20130263037A1
US20130263037A1 US13/908,253 US201313908253A US2013263037A1 US 20130263037 A1 US20130263037 A1 US 20130263037A1 US 201313908253 A US201313908253 A US 201313908253A US 2013263037 A1 US2013263037 A1 US 2013263037A1
Authority
US
United States
Prior art keywords
musical
control unit
octave
note
display
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/908,253
Inventor
Guang Yong CHOI
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Publication of US20130263037A1 publication Critical patent/US20130263037A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04886Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04BTRANSMISSION
    • H04B1/00Details of transmission systems, not covered by a single one of groups H04B3/00 - H04B13/00; Details of transmission systems not characterised by the medium used for transmission
    • H04B1/38Transceivers, i.e. devices in which transmitter and receiver form a structural unit and in which at least one part is used for functions of transmitting and receiving
    • H04B1/40Circuits
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0008Associated control or indicating means
    • G10H1/0016Means for indicating which keys, frets or strings are to be actuated, e.g. using lights or leds
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/32Constructional details
    • G10H1/34Switch arrangements, e.g. keyboards or mechanical switches specially adapted for electrophonic musical instruments
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/101Music Composition or musical creation; Tools or processes therefor
    • G10H2210/125Medley, i.e. linking parts of different musical pieces in one single piece, e.g. sound collage, DJ mix
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2220/00Input/output interfacing specifically adapted for electrophonic musical tools or instruments
    • G10H2220/021Indicator, i.e. non-screen output user interfacing, e.g. visual or tactile instrument status or guidance information using lights, LEDs, seven segments displays
    • G10H2220/026Indicator, i.e. non-screen output user interfacing, e.g. visual or tactile instrument status or guidance information using lights, LEDs, seven segments displays associated with a key or other user input device, e.g. key indicator lights
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2220/00Input/output interfacing specifically adapted for electrophonic musical tools or instruments
    • G10H2220/091Graphical user interface [GUI] specifically adapted for electrophonic musical instruments, e.g. interactive musical displays, musical instrument icons or menus; Details of user interactions therewith
    • G10H2220/096Graphical user interface [GUI] specifically adapted for electrophonic musical instruments, e.g. interactive musical displays, musical instrument icons or menus; Details of user interactions therewith using a touch screen
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2220/00Input/output interfacing specifically adapted for electrophonic musical tools or instruments
    • G10H2220/155User input interfaces for electrophonic musical instruments
    • G10H2220/221Keyboards, i.e. configuration of several keys or key-like input devices relative to one another
    • G10H2220/241Keyboards, i.e. configuration of several keys or key-like input devices relative to one another on touchscreens, i.e. keys, frets, strings, tablature or staff displayed on a touchscreen display for note input purposes
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2230/00General physical, ergonomic or hardware implementation of electrophonic musical tools or instruments, e.g. shape or architecture
    • G10H2230/005Device type or category
    • G10H2230/021Mobile ringtone, i.e. generation, transmission, conversion or downloading of ringing tones or other sounds for mobile telephony; Special musical data formats or protocols herefor

Definitions

  • the present invention relates generally to a mobile terminal and, more particularly, to a song writing method and apparatus using the touch screen in a mobile terminal.
  • Such a mobile terminal enables conversation between a caller and recipient through a mobile communication network, and produces a bell sound, alarm sound and background music to notify reception of a call or text message by playing back a preset song file using an embedded hardware chip.
  • a mobile terminal capable of playing back music files can produce high quality sounds and store a large number of song files thanks to advanced memory chips.
  • the user of a mobile terminal may directly write a song by inputting through the keypad or by entering musical notes or drawing waveforms on the touch screen, and store the written song in a music file format for hearing or saving.
  • song writing through keypad input requires the user to be skillful in manipulation of various keys, and song writing through musical notes or waveforms requires the user to have some background knowledge of music. This process is long and cumbersome and, thus, a user is inconvenienced in developing songs.
  • the present invention provides a method that can increase utilization of song writing programs in a mobile terminal by implementing a user-friendly musical keyboard on an intuitive user interface.
  • the present invention also provides a method that can overcome the space limitation by displaying only a musical keyboard of an octave and simple setting keys.
  • the present invention further provides a method that can mix a song written through an onscreen musical keyboard with contents generated through various musical instruments.
  • a song writing method using a touch screen in a mobile terminal includes: displaying, upon activation of a song writing mode, an onscreen musical keyboard of an octave on the touch screen; identifying, when a key of the onscreen musical keyboard is touched, a musical note mapped to the touched key through computation of coordinates of the touched key; creating musical interval data on the basis of the identified note and the current octave level; and converting the musical interval data into sound data recognizable by the mobile terminal and outputting the sound data.
  • a song writing apparatus using a touch screen in a mobile terminal includes: a display unit displaying, upon activation of a song writing mode, an onscreen musical keyboard of an octave on the touch screen; a data processing section identifying, when a key of the onscreen musical keyboard is touched, a musical note mapped to the touched key through computation of coordinates of the touched key, creating musical interval data on the basis of the identified note and the current octave level, and converting the musical interval data into sound data recognizable by the mobile terminal and outputting the sound data; and a control unit outputting the sound data.
  • various types of songs can be conveniently composed through an enhanced song writing program employing an onscreen musical keyboard on the touch screen of a mobile terminal.
  • the applicability of a mobile terminal can be extended through a function that can mix a song written through an onscreen musical keyboard with contents generated through various musical instruments.
  • the song writing method can transform the user of a mobile terminal from a passive consumer only listening to songs created by others to an active creator self-producing songs that can be shared with others.
  • FIG. 1 is a block diagram illustrating a mobile terminal according to an embodiment of the present invention
  • FIG. 2 is a flow chart of a song writing procedure using an onscreen musical keyboard on the touch screen of a mobile terminal according to another embodiment of the present invention
  • FIGS. 3A , 3 B, 3 C, 3 D and 3 E are screen representations illustrating steps in the procedure of FIG. 2 ;
  • FIG. 4 is a flow chart of a procedure to mix a composed song with contents generated through another musical instrument according to another embodiment of the present invention
  • FIGS. 5A and 5B are screen representations illustrating steps in the mixing procedure of FIG. 4 ;
  • FIG. 6 is a flow chart of a song writing procedure using an onscreen musical keyboard and other musical instrument according to another embodiment of the present invention.
  • FIG. 7 is a flow chart illustrating a procedure of song writing by replacing musical notes mapped to touched keys on the onscreen musical keyboard with those of another musical instrument.
  • an ‘octave’ is a series of eight notes in a musical scale. That is, an octave is the interval between a reference note and the eighth note therefrom in a musical scale. For example, an octave includes eight notes between the middle ‘Do’ and the next ‘Do’ in the same pitch class. In a typical musical keyboard, the middle ‘Do’ corresponds to the note at the center, and the next ‘Do’ corresponds to a lower ‘Do’ or higher ‘Do’, which is the eighth note from the middle ‘Do’.
  • Musical ‘interval data’ indicates data obtained on the basis of notes corresponding to keys touched on the musical keyboard and the current octave.
  • Sound data is data that can be recognized by a mobile terminal and is obtained through conversion from at least one piece of interval data.
  • sound data may include files, usable in a mobile terminal, in an MP3 (Moving Picture Experts Group Audio Layer-3) format, a WMA (Windows Media Audio) format, a RA (Real Audio) format, an AIFF (Audio Interchange File Format) format, and a MIDI (Musical Instrument Digital Interface) format.
  • Sound data in a first file format may be converted into sound data in a second file format depending upon the particular application(s) being used.
  • ‘Mixing’ indicates a procedure that mixes a song written through an onscreen musical keyboard and contents generated through other musical instruments together into a new song.
  • Mixing may indicate a procedure that generates interval data by replacing musical notes mapped to touched keys on the onscreen musical keyboard with those of another musical instrument.
  • Mixing may indicate a procedure that generates interval data by mixing musical notes input through the onscreen musical keyboard and those of another musical instrument corresponding to the input musical notes.
  • the mobile terminal is a terminal that can be used for song writing utilizing an onscreen musical keyboard on the touch screen, and may be any terminal, such as a mobile phone, personal digital assistant (PDA), code division multiple access (CDMA) terminal, wideband CDMA (WCDMA) terminal, global system for mobile communications (GSM) terminal, international mobile telecommunications 2000 (IMT 2000) terminal, smart phone, or universal mobile telecommunications system (UMTS) terminal.
  • PDA personal digital assistant
  • CDMA code division multiple access
  • WCDMA wideband CDMA
  • GSM global system for mobile communications
  • IMT 2000 international mobile telecommunications 2000
  • smart phone smart phone
  • UMTS universal mobile telecommunications system
  • FIG. 1 is a block diagram illustrating a mobile terminal according to an exemplary embodiment of the present invention.
  • the mobile terminal includes a control unit 100 , input unit 110 , display unit 120 , storage unit 130 , audio unit 140 , and radio frequency unit 150 .
  • the control unit 100 includes a data processing section 102 .
  • the input unit 110 includes a plurality of keys for inputting alphanumeric information, and a plurality of function keys for setting various functions.
  • the function keys may include an effect sound key, direction key, side key, shortcut key, scroll wheel, and setting key.
  • Examples of a setting key may include a save key, pre-listening key, upload key, and download key.
  • the input unit 110 may be implemented using a touch screen, in which case the input unit 110 may also provide a display function.
  • the input unit 110 sends an input key signal related to user settings and controlling of the mobile terminal to the control unit 100 .
  • the input unit 110 sends a key signal corresponding to a touched key of the onscreen musical keyboard and an octave level change signal corresponding to a touch on a flick zone of the onscreen musical keyboard to the control unit 100 .
  • the flick zone is a zone that detects a touch and movement in the left or right direction to shift the onscreen musical keyboard.
  • the input unit 110 sends a signal input by an effect sound button to the control unit 100 .
  • the input unit 110 sends an input signal, which initiates generation of interval data by replacing musical notes mapped to touched keys on the onscreen musical keyboard with those of another musical instrument, to the control unit 100 .
  • the input unit 110 sends function key signals, for pre-listening, saving and uploading of a composed song, to the control unit 100 .
  • the display unit 120 displays operation states, operation results, and various information under the control of the control unit 100 .
  • the display unit 120 displays menus of the mobile terminal, information input by the user, function setting information, and information to be provided to the user.
  • the display unit 120 may include a panel that is composed of display devices such as liquid crystal display (LCD) devices or organic light emitting diodes (OLED).
  • the display unit 120 has a touch screen capability, and can act as an input means performing the function of the input unit 110 .
  • the display unit 120 receives a user command or graphic information by generating a voltage or current signal corresponding to a pressed location on a touch panel of the touch screen.
  • the touch screen includes a touch panel, LCD panel, and backlight stacked in series.
  • the touch screen may further include signal lines transmitting signals from the touch panel to the control unit 100 , and a tape carrier package (TCP) connected to the lower substrate of the LCD panel to feed electric signals to the driving part of the LCD panel.
  • TCP tape carrier package
  • the display unit 120 highlights a touched key of the onscreen musical keyboard under the control of the control unit 100 .
  • the display unit 120 displays a flick zone for changing the octave level.
  • the display unit 120 displays the octave level, which may be lowered or raised according to the direction of a touch on the flick zone.
  • the display unit 120 displays a list of contents that are generated through various musical instruments and are mixable with the song written through the onscreen musical keyboard, under the control of the control unit 100 .
  • the storage unit 130 stores application programs for embodiments of the present invention.
  • the storage unit 130 may include a program storage area and a data storage area.
  • the program storage area stores an operating system for enabling the mobile terminal to control the operation of various hardware within the mobile terminal and software (programs) operating within the mobile terminal, a program for obtaining interval data on the basis of notes corresponding to keys touched on the onscreen musical keyboard and the current octave, and a program for converting interval data of a composed song into sound data recognizable by the mobile terminal.
  • the data storage area stores data generated during the use of the mobile terminal, interval data obtained on the basis of notes corresponding to keys touched on the onscreen musical keyboard and the current octave, and sound data recognizable by the mobile terminal obtained through conversion from interval data of a composed song.
  • the audio unit 140 reproduces an audio signal from the control unit 100 , and sends an audio signal such as a voice signal from a microphone MIC to the control unit 100 . That is, the audio unit 140 converts voice and acoustic data into audible sounds for output through a speaker SPK, and converts an audio signal, such as a voice signal from a microphone MIC, into audio data for output to the control unit 100 . In particular, the audio unit 140 reproduces sound data that is obtained through conversion from interval data created on the basis of notes corresponding to keys touched on the onscreen musical keyboard and the associated octave. The audio unit 140 plays a song that is created through mixing a song material composed through the onscreen musical keyboard and contents generated through various musical instruments together, under the control of the control unit 100 .
  • the radio frequency unit 150 establishes a communication channel between the mobile terminal and a serving base station, and performs communication operations to send and receive necessary signals.
  • the radio frequency unit 150 may include a radio frequency transmitter for up-converting the frequency of a signal to be transmitted and amplifying the signal, and a radio frequency receiver for low-noise amplifying a received signal and down-converting the frequency of the signal.
  • the radio frequency unit 150 performs communication operations between the mobile terminal and base station through a wireless connection, e.g, Internet, to upload song data composed through the onscreen musical keyboard, and to download contents generated through other musical instruments.
  • the data processing section 102 controls input through the onscreen musical keyboard under the control of the control unit 100 .
  • the data processing section 102 obtains a musical note corresponding to a touched key on the onscreen musical keyboard.
  • the data processing section 102 may compute the coordinates of a touched location on the touch screen, identify the musical note mapped to the computed coordinates, create interval data on the basis of the identified musical note and associated octave, convert the interval data into sound data recognizable by the mobile terminal, and store the sound data in the storage unit 130 .
  • the data processing section 102 may create interval data by mixing a song written through the onscreen musical keyboard and the contents generated through other musical instruments together.
  • the data processing section 102 may mix a song written through the onscreen musical keyboard with the contents generated through other musical instruments to create a new song.
  • the data processing section 102 may create interval data by replacing a note corresponding to a touched key on the onscreen musical keyboard with a note of a given musical instrument.
  • the data processing section 102 may create interval data by mixing a note corresponding to a touched key on the onscreen musical keyboard with a note processed by a given musical instrument.
  • the data processing section 102 converts a piece of interval data obtained through mixing into sound data usable by the mobile terminal.
  • the control unit 100 controls the overall operation of the mobile terminal and signal exchange between internal blocks thereof.
  • the control unit 100 controls the display unit 120 to display an onscreen musical keyboard of an octave.
  • the control unit 100 can control the display unit 120 to represent a default octave composed of eight notes from the middle ‘Do’ being the reference note to the next ‘Do’.
  • the control unit 100 can control the display unit 120 to change the octave level of the onscreen musical keyboard to a higher octave or a lower octave.
  • control unit 100 controls the display unit 120 to shift the onscreen musical keyboard to the left-hand direction so that the onscreen musical keyboard represents a new octave of eight notes.
  • the control unit 100 controls the data processing section 102 to identify the note mapped to the touched key after computing the coordinates of the touched key, and to create interval data on the basis of the identified note and the set octave.
  • the control unit 100 may set a default octave to include eight notes from the middle ‘Do’ being the reference note to the next ‘Do’.
  • the control unit 100 controls the data processing section 102 to store the created interval data in the storage unit 130 , and to convert the interval data into sound data usable by the mobile terminal.
  • the created sound data may be a file in an MP3, WMA, RA, AIFF or MIDI format recognizable by the mobile terminal. Sound data in a file format may be converted into sound data in another file format according to applications.
  • the control unit 100 can control the audio unit 140 to reproduce the created sound data.
  • the control unit 100 Upon reception of the sound signal, the control unit 100 controls the display unit 120 to display a list of contents that are generated through other musical instruments and are playable together with the written song. That is, the control unit 100 displays a list of contents mixable with the written song.
  • the control unit 100 controls the data processing section 102 to mix the written song with a selected content and to generate sound data usable by the mobile terminal from the mixed result.
  • the control unit 100 may control the data processing section 102 to generate interval data by applying a musical note mapped to a touched key of the onscreen musical keyboard to a selected one of musical instruments playable together.
  • the song writing apparatus using the touch screen in a mobile terminal may receive an input musical note through the onscreen musical keyboard, create interval data using the input note and associated octave, convert at least one piece of interval data into sound data for output, change the octave level, and mix a composed song with contents generated through various musical instruments.
  • FIG. 2 is a flow chart of a song writing procedure using an onscreen musical keyboard on the touch screen of a mobile terminal according to another exemplary embodiment of the present invention.
  • FIGS. 3A to 3E are screen representations illustrating steps in the procedure of FIG. 2 .
  • the control unit 100 of the mobile terminal receives an input signal for song writing mode activation ( 201 ).
  • the control unit 100 controls the display unit 120 to display an initial screen for the song writing mode ( 203 ).
  • an onscreen musical keyboard of an octave is included in the initial screen, and the current octave level and the flick zone are also included.
  • the control unit 100 controls the display unit 120 to display an onscreen musical keyboard 307 of an octave, octave level indication 301 , effect sound button 303 , setting button 305 , and flick zone 309 .
  • the octave level indicator 301 indicates the current octave level governing the eight notes of the onscreen musical keyboard 307 on the display unit 120 .
  • the control unit 100 checks whether a touch is detected on the onscreen musical keyboard ( 205 ). When a touch is not detected at step 205 , the control unit 100 controls the display unit 120 to continuously display the initial screen for the song writing mode.
  • the data processing section 102 computes the coordinates of the touched location on the display unit 120 , and identifies the note assigned to the computed coordinates ( 207 ). For example, as shown in FIG. 3B , the data processing section 102 computes the coordinates 311 of a touched location on the touch screen, and identifies the note assigned to the key at the computed coordinates 311 .
  • control unit 100 can compute the coordinates of all the touched locations and identify all corresponding notes.
  • the control unit 100 controls the display unit 120 to highlight the touched key whose note is identified at step 209 by the data processing section 102 ( 209 ). For example, as shown in FIG. 3C , the control unit 100 controls the display unit 120 to highlight a key 313 at the touched location.
  • the control unit 100 controls the data processing section 102 to create interval data using the note corresponding to the touched key and the current octave level ( 211 ).
  • the control unit 100 may set a default octave composed of eight notes from the middle ‘Do’ to the next ‘Do’.
  • the control unit 100 may also set a default octave composed of eight notes from a note other than the middle ‘Do’ to the same note in a higher or lower octave.
  • the control unit 100 controls the data processing section 102 to store the created interval data in the storage unit 130 ( 213 ).
  • the control unit 100 checks whether a signal for song writing mode termination is input ( 215 ).
  • the control unit 100 may detect a signal for song writing mode termination generated by a setting button.
  • the control unit 100 controls the data processing section 102 to convert the interval data into sound data recognizable by the mobile terminal ( 221 ).
  • the sound data may be files in different formats usable by other applications.
  • the control unit 100 stores the sound data in the storage unit 130 ( 223 ).
  • the control unit 100 checks whether a signal for octave level change is input ( 217 ). When a signal for octave level change is input through the flick zone, the control unit 100 lowers or raises the current octave level and controls the display unit 120 to represent the new octave level ( 219 ). For example, the control unit 100 may control the octave level change according to a touch on the flick zone as shown in FIG. 3D , and control the display unit 120 to update the octave level indication 315 as shown in FIG. 3E . After the octave level change, the control unit 100 returns to step 205 for further detection of a touch on the onscreen musical keyboard.
  • control unit 100 When a signal for octave level change is not input at step 217 , the control unit 100 returns to step 205 for further detection of a touch on the onscreen musical keyboard.
  • control unit 100 After saving the sound data, the control unit 100 checks whether a signal for sound data output is input ( 225 ). The control unit 100 may detect a signal for sound data output from the pre-listening key of the input unit 110 .
  • control unit 100 controls the audio unit 140 to output the sound data through the speaker SPK ( 227 ).
  • control unit 100 When a signal for sound data output is not input, the control unit 100 performs a requested operation ( 229 ).
  • FIG. 4 is a flow chart of a procedure to mix a composed song with contents generated through another musical instrument.
  • FIGS. 5A and 5B are screen representations illustrating steps in the mixing procedure of FIG. 4 .
  • the control unit 100 of the mobile terminal detects an effect sound button signal for mixing a song composed through the song writing method with a content generated through another musical instrument ( 401 ).
  • the control unit 100 Upon detection of an effect sound button signal, the control unit 100 controls the display unit 120 to display a list of musical instruments mixable with a song written through the onscreen musical keyboard ( 403 ). For example, as shown in FIG. 5A , the control unit 100 may cause a list of various musical instruments to be displayed on the touch screen.
  • the control unit 100 checks whether a signal for selecting one of the listed musical instruments is detected ( 405 ). That is, the user may select a musical instrument for mixing. Step 404 may be repeated to select multiple musical instruments for mixing.
  • the control unit 100 controls the display unit 120 to continuously display the list of musical instruments.
  • the control unit 100 controls the display unit 120 to display a list of contents related to the selected musical instrument ( 407 ). For example, when a guitar is selected as shown in FIG. 5B , the control unit 100 can control the display unit 120 to display “G code”, “C code” and “F code” in a popup window as a content list for the guitar.
  • the control unit 100 checks whether a signal for selecting one of the listed contents is detected ( 409 ). When a signal for selecting one of the listed contents is not detected, the control unit 100 causes continued display of the list of contents.
  • control unit 100 retrieves a song written by the onscreen musical keyboard from the storage unit 130 ( 411 ).
  • the control unit 100 invokes a mixing procedure that mixes the retrieved song with the content related to the selected musical instrument ( 413 ), and controls the data processing section 102 to produce new sound data utilizing the mixing procedure ( 415 ).
  • the control unit 100 stores the newly produced sound data in the storage unit 130 ( 417 ). Later, upon request, the control unit 100 may reproduce the stored sound data through the audio unit 140 .
  • FIG. 6 is a flow chart of a song writing procedure using a combination of an onscreen musical keyboard and another musical instrument.
  • the control unit 100 of the mobile terminal detects a mixed mode signal for composing a song utilizing a combination of the onscreen musical keyboard and other musical instrument ( 601 ).
  • control unit 100 Upon detection of a mixed mode signal, the control unit 100 controls the display unit 120 to display a list of musical instruments mixable with the onscreen musical keyboard ( 603 ). The control unit 100 checks whether a signal for selecting one of the listed musical instruments is detected ( 605 ). Step 605 may be repeated to select multiple musical instruments used together for mixing.
  • control unit 100 controls the display unit 120 to continuously display the list of musical instruments.
  • the control unit 100 detects a touch on a key of the onscreen musical keyboard ( 607 ). Upon detection of a touch, the control unit 100 controls the data processing section 102 to identify the note mapped to the touched key, to extract a note of the selected musical instrument corresponding to the identified note, and to create mixed interval data on the basis of the note associated with the onscreen musical keyboard and the corresponding note of the selected musical instrument ( 609 ).
  • control unit 100 controls the data processing section 102 to store the created interval data in the storage unit 130 ( 611 ).
  • the control unit 100 checks whether a signal for song writing mode termination is input ( 613 ).
  • the control unit 100 may detect a signal for song writing mode termination generated by a setting button.
  • control unit 100 controls the data processing section 102 to convert the interval data into sound data recognizable by the mobile terminal ( 615 ), and stores the resulting sound data in the storage unit 130 ( 617 ).
  • control unit 100 When a signal for song writing mode termination is not input, the control unit 100 returns to step 607 to further detect a touch on a key of the onscreen musical keyboard.
  • FIG. 7 is a flow chart illustrating a procedure of song writing by replacing musical notes mapped to touched keys on the onscreen musical keyboard with those of another musical instrument.
  • the control unit 100 of the mobile terminal detects an effect sound button signal for writing a song by replacing a musical note given by the onscreen musical keyboard with that of another musical instrument ( 701 ).
  • control unit 100 Upon detection of an effect sound button signal, the control unit 100 controls the display unit 120 to display a list of musical instruments that may produce a note replacing a note given by the onscreen musical keyboard ( 703 ).
  • the control unit 100 checks whether a signal for selecting one of the listed musical instruments is detected ( 705 ). Step 705 may be repeated to select multiple musical instruments used together for mixing.
  • control unit 100 controls the display unit 120 to continuously display the list of musical instruments.
  • control unit 100 controls the display unit 120 to display a list of contents related to the selected musical instrument ( 707 ).
  • the control unit 100 checks whether a signal for selecting one of the listed contents is detected ( 709 ). When a signal for selecting one of the listed contents is not detected, the control unit 100 causes continued display of the list of contents.
  • control unit 100 identifies a note mapped to a touched key on the onscreen musical keyboard ( 711 ).
  • the control unit 100 creates interval data by replacing the identified note of the onscreen musical keyboard with a note extracted from the selected content ( 713 ). Thereafter, the control unit 100 checks whether a signal for song writing mode termination is input ( 715 ). Here, the control unit 100 may detect a signal for song writing mode termination generated by a setting button.
  • control unit 100 controls the data processing section 102 to convert the interval data into sound data recognizable by the mobile terminal ( 717 ), and stores the resulting sound data in the storage unit 130 ( 719 ).
  • a song writing mode termination may be explicitly determined or implicitly determined.
  • the section of a second application may be used to provide a song writing mode termination signal.
  • the above-described methods according to the present invention can be realized in hardware or as software or computer code that can be stored in a recording medium such as a CD ROM, an RAM, a floppy disk, a hard disk, or a magneto-optical disk or downloaded over a network, so that the methods described herein can be rendered in such software using a general purpose computer, or a special processor or in programmable or dedicated hardware, such as an ASIC or FPGA.
  • the computer, the processor or the programmable hardware include memory components, e.g., RAM, ROM, Flash, etc. that may store or receive software or computer code that when accessed and executed by the computer, processor or hardware implement the processing methods described herein.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • General Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Electrophonic Musical Instruments (AREA)
  • User Interface Of Digital Computer (AREA)
  • Telephone Function (AREA)

Abstract

A song writing method and apparatus using a touch screen in a mobile terminal are disclosed. The song writing method includes: displaying, upon activation of a song writing mode, an onscreen musical keyboard of an octave on the touch screen; identifying, when a key of the onscreen musical keyboard is touched, a musical note mapped to the touched key through computation of coordinates of the touched key; creating musical interval data on the basis of the identified note and the current octave level; and converting the musical interval data into sound data recognizable by the mobile terminal and outputting the sound data. Hence, various types of songs can be conveniently composed through the song writing method employing the onscreen musical keyboard. The applicability of the mobile terminal can be extended through a function that can mix a written song with contents generated through various musical instruments.

Description

    CLAIMS OF PRIORITY
  • This is a Continuation of U.S. application Ser. No. 12/547,556 filed on Aug. 26, 2009 which in turn claims the benefit, pursuant to 35 USC 119, to an earlier application filed in the Korean Intellectual Property Office on Sep. 1, 2008 and assigned Serial No. 10-2008-0085826, the contents of which are incorporated herein by reference.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates generally to a mobile terminal and, more particularly, to a song writing method and apparatus using the touch screen in a mobile terminal.
  • 2. Description of the Related Art
  • With recent advances in communication technologies, mobile terminals have been widely popularized and have become one of the necessities of life. Such a mobile terminal enables conversation between a caller and recipient through a mobile communication network, and produces a bell sound, alarm sound and background music to notify reception of a call or text message by playing back a preset song file using an embedded hardware chip.
  • In particular, a mobile terminal capable of playing back music files can produce high quality sounds and store a large number of song files thanks to advanced memory chips.
  • With proliferation of user generated contents (UGC) containing self-produced songs and moving images, users show a growing interest in creating songs or video materials utilizing mobile terminals.
  • The user of a mobile terminal may directly write a song by inputting through the keypad or by entering musical notes or drawing waveforms on the touch screen, and store the written song in a music file format for hearing or saving.
  • However, song writing through keypad input requires the user to be skillful in manipulation of various keys, and song writing through musical notes or waveforms requires the user to have some background knowledge of music. This process is long and cumbersome and, thus, a user is inconvenienced in developing songs.
  • In addition, there may be restrictions on representing information regarding the music paper for drawing musical notations or the palette for drawing waveforms, on the single touch screen.
  • SUMMARY OF THE INVENTION
  • The present invention provides a method that can increase utilization of song writing programs in a mobile terminal by implementing a user-friendly musical keyboard on an intuitive user interface.
  • The present invention also provides a method that can overcome the space limitation by displaying only a musical keyboard of an octave and simple setting keys.
  • The present invention further provides a method that can mix a song written through an onscreen musical keyboard with contents generated through various musical instruments.
  • In one embodiment of the present invention, a song writing method using a touch screen in a mobile terminal includes: displaying, upon activation of a song writing mode, an onscreen musical keyboard of an octave on the touch screen; identifying, when a key of the onscreen musical keyboard is touched, a musical note mapped to the touched key through computation of coordinates of the touched key; creating musical interval data on the basis of the identified note and the current octave level; and converting the musical interval data into sound data recognizable by the mobile terminal and outputting the sound data.
  • In another embodiment of the present invention, a song writing apparatus using a touch screen in a mobile terminal includes: a display unit displaying, upon activation of a song writing mode, an onscreen musical keyboard of an octave on the touch screen; a data processing section identifying, when a key of the onscreen musical keyboard is touched, a musical note mapped to the touched key through computation of coordinates of the touched key, creating musical interval data on the basis of the identified note and the current octave level, and converting the musical interval data into sound data recognizable by the mobile terminal and outputting the sound data; and a control unit outputting the sound data.
  • In the embodiments of the present invention, various types of songs can be conveniently composed through an enhanced song writing program employing an onscreen musical keyboard on the touch screen of a mobile terminal.
  • The applicability of a mobile terminal can be extended through a function that can mix a song written through an onscreen musical keyboard with contents generated through various musical instruments.
  • The song writing method can transform the user of a mobile terminal from a passive consumer only listening to songs created by others to an active creator self-producing songs that can be shared with others.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The features and advantages of the present invention will be more apparent from the following detailed description in conjunction with the accompanying drawings, in which:
  • FIG. 1 is a block diagram illustrating a mobile terminal according to an embodiment of the present invention;
  • FIG. 2 is a flow chart of a song writing procedure using an onscreen musical keyboard on the touch screen of a mobile terminal according to another embodiment of the present invention;
  • FIGS. 3A, 3B, 3C, 3D and 3E are screen representations illustrating steps in the procedure of FIG. 2;
  • FIG. 4 is a flow chart of a procedure to mix a composed song with contents generated through another musical instrument according to another embodiment of the present invention;
  • FIGS. 5A and 5B are screen representations illustrating steps in the mixing procedure of FIG. 4;
  • FIG. 6 is a flow chart of a song writing procedure using an onscreen musical keyboard and other musical instrument according to another embodiment of the present invention; and
  • FIG. 7 is a flow chart illustrating a procedure of song writing by replacing musical notes mapped to touched keys on the onscreen musical keyboard with those of another musical instrument.
  • DETAILED DESCRIPTION OF THE INVENTION
  • Hereinafter, embodiments of the present invention are described in detail with reference to the accompanying drawings. The same reference symbols are used throughout the drawings to refer to the same or like parts. For the purposes of clarity and simplicity, detailed descriptions of well-known functions and structures incorporated herein may be omitted to avoid obscuring the subject matter of the present invention.
  • To help understand the invention, the following definitions are used: an ‘octave’ is a series of eight notes in a musical scale. That is, an octave is the interval between a reference note and the eighth note therefrom in a musical scale. For example, an octave includes eight notes between the middle ‘Do’ and the next ‘Do’ in the same pitch class. In a typical musical keyboard, the middle ‘Do’ corresponds to the note at the center, and the next ‘Do’ corresponds to a lower ‘Do’ or higher ‘Do’, which is the eighth note from the middle ‘Do’. Musical ‘interval data’ indicates data obtained on the basis of notes corresponding to keys touched on the musical keyboard and the current octave. ‘Sound data’ is data that can be recognized by a mobile terminal and is obtained through conversion from at least one piece of interval data. For example, sound data may include files, usable in a mobile terminal, in an MP3 (Moving Picture Experts Group Audio Layer-3) format, a WMA (Windows Media Audio) format, a RA (Real Audio) format, an AIFF (Audio Interchange File Format) format, and a MIDI (Musical Instrument Digital Interface) format. Sound data in a first file format may be converted into sound data in a second file format depending upon the particular application(s) being used. ‘Mixing’ indicates a procedure that mixes a song written through an onscreen musical keyboard and contents generated through other musical instruments together into a new song. Mixing may indicate a procedure that generates interval data by replacing musical notes mapped to touched keys on the onscreen musical keyboard with those of another musical instrument. Mixing may indicate a procedure that generates interval data by mixing musical notes input through the onscreen musical keyboard and those of another musical instrument corresponding to the input musical notes.
  • Note that the mobile terminal according to the teachings of the present invention is a terminal that can be used for song writing utilizing an onscreen musical keyboard on the touch screen, and may be any terminal, such as a mobile phone, personal digital assistant (PDA), code division multiple access (CDMA) terminal, wideband CDMA (WCDMA) terminal, global system for mobile communications (GSM) terminal, international mobile telecommunications 2000 (IMT 2000) terminal, smart phone, or universal mobile telecommunications system (UMTS) terminal.
  • FIG. 1 is a block diagram illustrating a mobile terminal according to an exemplary embodiment of the present invention.
  • Referring to FIG. 1, the mobile terminal includes a control unit 100, input unit 110, display unit 120, storage unit 130, audio unit 140, and radio frequency unit 150. In particular, the control unit 100 includes a data processing section 102.
  • The input unit 110 includes a plurality of keys for inputting alphanumeric information, and a plurality of function keys for setting various functions. The function keys may include an effect sound key, direction key, side key, shortcut key, scroll wheel, and setting key. Examples of a setting key may include a save key, pre-listening key, upload key, and download key. The input unit 110 may be implemented using a touch screen, in which case the input unit 110 may also provide a display function. The input unit 110 sends an input key signal related to user settings and controlling of the mobile terminal to the control unit 100.
  • When an onscreen musical keyboard of an octave is displayed on the display unit 120, the input unit 110 sends a key signal corresponding to a touched key of the onscreen musical keyboard and an octave level change signal corresponding to a touch on a flick zone of the onscreen musical keyboard to the control unit 100. The flick zone is a zone that detects a touch and movement in the left or right direction to shift the onscreen musical keyboard. For mixing a song written through the onscreen musical keyboard and contents generated through other musical instruments together, the input unit 110 sends a signal input by an effect sound button to the control unit 100. The input unit 110 sends an input signal, which initiates generation of interval data by replacing musical notes mapped to touched keys on the onscreen musical keyboard with those of another musical instrument, to the control unit 100.
  • The input unit 110 sends function key signals, for pre-listening, saving and uploading of a composed song, to the control unit 100.
  • The display unit 120 displays operation states, operation results, and various information under the control of the control unit 100. The display unit 120 displays menus of the mobile terminal, information input by the user, function setting information, and information to be provided to the user. The display unit 120 may include a panel that is composed of display devices such as liquid crystal display (LCD) devices or organic light emitting diodes (OLED). In particular, the display unit 120 has a touch screen capability, and can act as an input means performing the function of the input unit 110. The display unit 120 receives a user command or graphic information by generating a voltage or current signal corresponding to a pressed location on a touch panel of the touch screen. The touch screen includes a touch panel, LCD panel, and backlight stacked in series. The touch screen may further include signal lines transmitting signals from the touch panel to the control unit 100, and a tape carrier package (TCP) connected to the lower substrate of the LCD panel to feed electric signals to the driving part of the LCD panel. In particular, the display unit 120 highlights a touched key of the onscreen musical keyboard under the control of the control unit 100. The display unit 120 displays a flick zone for changing the octave level. The display unit 120 displays the octave level, which may be lowered or raised according to the direction of a touch on the flick zone. The display unit 120 displays a list of contents that are generated through various musical instruments and are mixable with the song written through the onscreen musical keyboard, under the control of the control unit 100.
  • The storage unit 130 stores application programs for embodiments of the present invention. The storage unit 130 may include a program storage area and a data storage area. The program storage area stores an operating system for enabling the mobile terminal to control the operation of various hardware within the mobile terminal and software (programs) operating within the mobile terminal, a program for obtaining interval data on the basis of notes corresponding to keys touched on the onscreen musical keyboard and the current octave, and a program for converting interval data of a composed song into sound data recognizable by the mobile terminal. The data storage area stores data generated during the use of the mobile terminal, interval data obtained on the basis of notes corresponding to keys touched on the onscreen musical keyboard and the current octave, and sound data recognizable by the mobile terminal obtained through conversion from interval data of a composed song.
  • The audio unit 140 reproduces an audio signal from the control unit 100, and sends an audio signal such as a voice signal from a microphone MIC to the control unit 100. That is, the audio unit 140 converts voice and acoustic data into audible sounds for output through a speaker SPK, and converts an audio signal, such as a voice signal from a microphone MIC, into audio data for output to the control unit 100. In particular, the audio unit 140 reproduces sound data that is obtained through conversion from interval data created on the basis of notes corresponding to keys touched on the onscreen musical keyboard and the associated octave. The audio unit 140 plays a song that is created through mixing a song material composed through the onscreen musical keyboard and contents generated through various musical instruments together, under the control of the control unit 100.
  • The radio frequency unit 150 establishes a communication channel between the mobile terminal and a serving base station, and performs communication operations to send and receive necessary signals. The radio frequency unit 150 may include a radio frequency transmitter for up-converting the frequency of a signal to be transmitted and amplifying the signal, and a radio frequency receiver for low-noise amplifying a received signal and down-converting the frequency of the signal. In particular, the radio frequency unit 150 performs communication operations between the mobile terminal and base station through a wireless connection, e.g, Internet, to upload song data composed through the onscreen musical keyboard, and to download contents generated through other musical instruments.
  • The data processing section 102 controls input through the onscreen musical keyboard under the control of the control unit 100. In particular, the data processing section 102 obtains a musical note corresponding to a touched key on the onscreen musical keyboard. The data processing section 102 may compute the coordinates of a touched location on the touch screen, identify the musical note mapped to the computed coordinates, create interval data on the basis of the identified musical note and associated octave, convert the interval data into sound data recognizable by the mobile terminal, and store the sound data in the storage unit 130.
  • The data processing section 102 may create interval data by mixing a song written through the onscreen musical keyboard and the contents generated through other musical instruments together. The data processing section 102 may mix a song written through the onscreen musical keyboard with the contents generated through other musical instruments to create a new song. The data processing section 102 may create interval data by replacing a note corresponding to a touched key on the onscreen musical keyboard with a note of a given musical instrument. The data processing section 102 may create interval data by mixing a note corresponding to a touched key on the onscreen musical keyboard with a note processed by a given musical instrument.
  • The data processing section 102 converts a piece of interval data obtained through mixing into sound data usable by the mobile terminal.
  • The control unit 100 controls the overall operation of the mobile terminal and signal exchange between internal blocks thereof. In particular, upon reception of an input signal for song writing mode activation, the control unit 100 controls the display unit 120 to display an onscreen musical keyboard of an octave. Here, the control unit 100 can control the display unit 120 to represent a default octave composed of eight notes from the middle ‘Do’ being the reference note to the next ‘Do’. Upon detection of an octave level change signal through the flick zone, the control unit 100 can control the display unit 120 to change the octave level of the onscreen musical keyboard to a higher octave or a lower octave. For example, if an octave level change signal indicates the left-hand direction, the control unit 100 controls the display unit 120 to shift the onscreen musical keyboard to the left-hand direction so that the onscreen musical keyboard represents a new octave of eight notes.
  • In response to a touch on a key of the onscreen musical keyboard on the display unit 120, the control unit 100 controls the data processing section 102 to identify the note mapped to the touched key after computing the coordinates of the touched key, and to create interval data on the basis of the identified note and the set octave. The control unit 100 may set a default octave to include eight notes from the middle ‘Do’ being the reference note to the next ‘Do’. The control unit 100 controls the data processing section 102 to store the created interval data in the storage unit 130, and to convert the interval data into sound data usable by the mobile terminal. The created sound data may be a file in an MP3, WMA, RA, AIFF or MIDI format recognizable by the mobile terminal. Sound data in a file format may be converted into sound data in another file format according to applications. The control unit 100 can control the audio unit 140 to reproduce the created sound data.
  • Upon reception of the sound signal, the control unit 100 controls the display unit 120 to display a list of contents that are generated through other musical instruments and are playable together with the written song. That is, the control unit 100 displays a list of contents mixable with the written song. The control unit 100 controls the data processing section 102 to mix the written song with a selected content and to generate sound data usable by the mobile terminal from the mixed result. The control unit 100 may control the data processing section 102 to generate interval data by applying a musical note mapped to a touched key of the onscreen musical keyboard to a selected one of musical instruments playable together.
  • As described above, the song writing apparatus using the touch screen in a mobile terminal may receive an input musical note through the onscreen musical keyboard, create interval data using the input note and associated octave, convert at least one piece of interval data into sound data for output, change the octave level, and mix a composed song with contents generated through various musical instruments. Next, operations of the song writing apparatus are described in detail with reference to the drawings.
  • FIG. 2 is a flow chart of a song writing procedure using an onscreen musical keyboard on the touch screen of a mobile terminal according to another exemplary embodiment of the present invention. FIGS. 3A to 3E are screen representations illustrating steps in the procedure of FIG. 2.
  • Referring to FIG. 2, the control unit 100 of the mobile terminal receives an input signal for song writing mode activation (201). The control unit 100 controls the display unit 120 to display an initial screen for the song writing mode (203). Here, an onscreen musical keyboard of an octave is included in the initial screen, and the current octave level and the flick zone are also included. For example, as shown in FIG. 3A, upon activation of the song writing mode, the control unit 100 controls the display unit 120 to display an onscreen musical keyboard 307 of an octave, octave level indication 301, effect sound button 303, setting button 305, and flick zone 309. The octave level indicator 301 indicates the current octave level governing the eight notes of the onscreen musical keyboard 307 on the display unit 120.
  • The control unit 100 checks whether a touch is detected on the onscreen musical keyboard (205). When a touch is not detected at step 205, the control unit 100 controls the display unit 120 to continuously display the initial screen for the song writing mode.
  • When a touch is detected on the onscreen musical keyboard, the data processing section 102 computes the coordinates of the touched location on the display unit 120, and identifies the note assigned to the computed coordinates (207). For example, as shown in FIG. 3B, the data processing section 102 computes the coordinates 311 of a touched location on the touch screen, and identifies the note assigned to the key at the computed coordinates 311.
  • In the case where two or more touches are performed, the control unit 100 can compute the coordinates of all the touched locations and identify all corresponding notes.
  • The control unit 100 controls the display unit 120 to highlight the touched key whose note is identified at step 209 by the data processing section 102 (209). For example, as shown in FIG. 3C, the control unit 100 controls the display unit 120 to highlight a key 313 at the touched location.
  • The control unit 100 controls the data processing section 102 to create interval data using the note corresponding to the touched key and the current octave level (211). Here, the control unit 100 may set a default octave composed of eight notes from the middle ‘Do’ to the next ‘Do’. The control unit 100 may also set a default octave composed of eight notes from a note other than the middle ‘Do’ to the same note in a higher or lower octave.
  • The control unit 100 controls the data processing section 102 to store the created interval data in the storage unit 130 (213).
  • The control unit 100 checks whether a signal for song writing mode termination is input (215). Here, the control unit 100 may detect a signal for song writing mode termination generated by a setting button.
  • When a signal for song writing mode termination is input, the control unit 100 controls the data processing section 102 to convert the interval data into sound data recognizable by the mobile terminal (221). The sound data may be files in different formats usable by other applications. The control unit 100 stores the sound data in the storage unit 130 (223).
  • When a signal for song writing mode termination is not input at step 215, the control unit 100 checks whether a signal for octave level change is input (217). When a signal for octave level change is input through the flick zone, the control unit 100 lowers or raises the current octave level and controls the display unit 120 to represent the new octave level (219). For example, the control unit 100 may control the octave level change according to a touch on the flick zone as shown in FIG. 3D, and control the display unit 120 to update the octave level indication 315 as shown in FIG. 3E. After the octave level change, the control unit 100 returns to step 205 for further detection of a touch on the onscreen musical keyboard.
  • When a signal for octave level change is not input at step 217, the control unit 100 returns to step 205 for further detection of a touch on the onscreen musical keyboard.
  • After saving the sound data, the control unit 100 checks whether a signal for sound data output is input (225). The control unit 100 may detect a signal for sound data output from the pre-listening key of the input unit 110.
  • When a signal for sound data output is input, the control unit 100 controls the audio unit 140 to output the sound data through the speaker SPK (227).
  • When a signal for sound data output is not input, the control unit 100 performs a requested operation (229).
  • FIG. 4 is a flow chart of a procedure to mix a composed song with contents generated through another musical instrument. FIGS. 5A and 5B are screen representations illustrating steps in the mixing procedure of FIG. 4.
  • Referring to FIG. 4, the control unit 100 of the mobile terminal detects an effect sound button signal for mixing a song composed through the song writing method with a content generated through another musical instrument (401).
  • Upon detection of an effect sound button signal, the control unit 100 controls the display unit 120 to display a list of musical instruments mixable with a song written through the onscreen musical keyboard (403). For example, as shown in FIG. 5A, the control unit 100 may cause a list of various musical instruments to be displayed on the touch screen.
  • The control unit 100 checks whether a signal for selecting one of the listed musical instruments is detected (405). That is, the user may select a musical instrument for mixing. Step 404 may be repeated to select multiple musical instruments for mixing.
  • When a signal for selecting one of the listed musical instruments is not detected, the control unit 100 controls the display unit 120 to continuously display the list of musical instruments. When a signal for selecting one of the listed musical instruments is detected, the control unit 100 controls the display unit 120 to display a list of contents related to the selected musical instrument (407). For example, when a guitar is selected as shown in FIG. 5B, the control unit 100 can control the display unit 120 to display “G code”, “C code” and “F code” in a popup window as a content list for the guitar.
  • The control unit 100 checks whether a signal for selecting one of the listed contents is detected (409). When a signal for selecting one of the listed contents is not detected, the control unit 100 causes continued display of the list of contents.
  • When a signal for selecting one of the listed contents is detected, the control unit 100 retrieves a song written by the onscreen musical keyboard from the storage unit 130 (411).
  • The control unit 100 invokes a mixing procedure that mixes the retrieved song with the content related to the selected musical instrument (413), and controls the data processing section 102 to produce new sound data utilizing the mixing procedure (415).
  • The control unit 100 stores the newly produced sound data in the storage unit 130 (417). Later, upon request, the control unit 100 may reproduce the stored sound data through the audio unit 140.
  • FIG. 6 is a flow chart of a song writing procedure using a combination of an onscreen musical keyboard and another musical instrument.
  • Referring to FIG. 6, the control unit 100 of the mobile terminal detects a mixed mode signal for composing a song utilizing a combination of the onscreen musical keyboard and other musical instrument (601).
  • Upon detection of a mixed mode signal, the control unit 100 controls the display unit 120 to display a list of musical instruments mixable with the onscreen musical keyboard (603). The control unit 100 checks whether a signal for selecting one of the listed musical instruments is detected (605). Step 605 may be repeated to select multiple musical instruments used together for mixing.
  • When a signal for selecting one of the listed musical instruments is not detected, the control unit 100 controls the display unit 120 to continuously display the list of musical instruments.
  • When a signal for selecting one of the listed musical instruments is detected, the control unit 100 detects a touch on a key of the onscreen musical keyboard (607). Upon detection of a touch, the control unit 100 controls the data processing section 102 to identify the note mapped to the touched key, to extract a note of the selected musical instrument corresponding to the identified note, and to create mixed interval data on the basis of the note associated with the onscreen musical keyboard and the corresponding note of the selected musical instrument (609).
  • After creation of the mixed interval data, the control unit 100 controls the data processing section 102 to store the created interval data in the storage unit 130 (611).
  • The control unit 100 checks whether a signal for song writing mode termination is input (613). Here, the control unit 100 may detect a signal for song writing mode termination generated by a setting button.
  • When a signal for song writing mode termination is input, the control unit 100 controls the data processing section 102 to convert the interval data into sound data recognizable by the mobile terminal (615), and stores the resulting sound data in the storage unit 130 (617).
  • When a signal for song writing mode termination is not input, the control unit 100 returns to step 607 to further detect a touch on a key of the onscreen musical keyboard.
  • FIG. 7 is a flow chart illustrating a procedure of song writing by replacing musical notes mapped to touched keys on the onscreen musical keyboard with those of another musical instrument.
  • Referring to FIG. 7, the control unit 100 of the mobile terminal detects an effect sound button signal for writing a song by replacing a musical note given by the onscreen musical keyboard with that of another musical instrument (701).
  • Upon detection of an effect sound button signal, the control unit 100 controls the display unit 120 to display a list of musical instruments that may produce a note replacing a note given by the onscreen musical keyboard (703).
  • The control unit 100 checks whether a signal for selecting one of the listed musical instruments is detected (705). Step 705 may be repeated to select multiple musical instruments used together for mixing.
  • When a signal for selecting one of the listed musical instruments is not detected, the control unit 100 controls the display unit 120 to continuously display the list of musical instruments.
  • When a signal for selecting one of the listed musical instruments is detected, the control unit 100 controls the display unit 120 to display a list of contents related to the selected musical instrument (707).
  • The control unit 100 checks whether a signal for selecting one of the listed contents is detected (709). When a signal for selecting one of the listed contents is not detected, the control unit 100 causes continued display of the list of contents.
  • When a signal for selecting one of the listed contents is detected, the control unit 100 identifies a note mapped to a touched key on the onscreen musical keyboard (711).
  • The control unit 100 creates interval data by replacing the identified note of the onscreen musical keyboard with a note extracted from the selected content (713). Thereafter, the control unit 100 checks whether a signal for song writing mode termination is input (715). Here, the control unit 100 may detect a signal for song writing mode termination generated by a setting button.
  • When a signal for song writing mode termination is input, the control unit 100 controls the data processing section 102 to convert the interval data into sound data recognizable by the mobile terminal (717), and stores the resulting sound data in the storage unit 130 (719).
  • When a signal for song writing mode termination is not input, the control unit 100 returns to step 711 to further detect a touch on a key of the onscreen musical keyboard. As would be appreciated, a song writing mode termination may be explicitly determined or implicitly determined. For example, the section of a second application may be used to provide a song writing mode termination signal.
  • The above-described methods according to the present invention can be realized in hardware or as software or computer code that can be stored in a recording medium such as a CD ROM, an RAM, a floppy disk, a hard disk, or a magneto-optical disk or downloaded over a network, so that the methods described herein can be rendered in such software using a general purpose computer, or a special processor or in programmable or dedicated hardware, such as an ASIC or FPGA. As would be understood in the art, the computer, the processor or the programmable hardware include memory components, e.g., RAM, ROM, Flash, etc. that may store or receive software or computer code that when accessed and executed by the computer, processor or hardware implement the processing methods described herein.
  • Although exemplary embodiments of the present invention have been described in detail hereinabove, it should be understood that many variations and modifications of the basic inventive concept herein described, which may appear to those skilled in the art, will still fall within the spirit and scope of the exemplary embodiments of the present invention as defined in the appended claims.

Claims (21)

1-20. (canceled)
21. A method comprising:
displaying a representation of at least one portion of a musical instrument via a display;
identifying an input indicating at least a direction thereof on or in proximity of a specified area of the display; and
determining a musical property associated with the representation based at least in part on the input.
22. The method of claim 21, wherein the displaying comprises:
displaying an image of at least one portion of a musical keyboard.
23. The method of claim 21, wherein the determining comprises:
setting an octave associated with the representation as part of the musical property.
24. The method of claim 23, wherein the setting comprises:
modifying the octave from a first octave to a second octave.
25. The method of claim 21, further comprising:
displaying another representation of the at least one portion based at least in part on the musical property.
26. The method of claim 21, wherein the determining comprises:
determining a musical note corresponding to the at least one portion.
27. The method of claim 26, further comprising:
presenting sound data corresponding to the musical note.
28. The method of claim 26, wherein the determining the musical note is performed based at least in part on receiving another input on or in proximity of the specified area.
29. The method of claim 26, further comprising:
determining another musical note corresponding to the at least one portion, the another musical note associated with another musical instrument; and
combining at least a portion of the musical note and at least a portion of the other musical note to produce a mixed musical note.
30. The method of claim 29, further comprising:
presenting sound data corresponding to the mixed musical note.
31. An apparatus comprising:
a display unit to display a representation of at least one portion of a musical instrument via a display; and
a control unit to identify an input indicating at least a direction thereof on or in proximity of a specified area of the display, and to determine a musical property associated with the representation based at least in part on the input.
32. The apparatus of claim 31, wherein the display unit is configured to:
display an image of at least one portion of a musical keyboard.
33. The apparatus of claim 31, wherein the control unit is configured to:
set an octave associated with the representation as part of the musical property.
34. The apparatus of claim 33, wherein the control unit is configured to:
modify the octave from a first octave to a second octave.
35. The apparatus of claim 31, wherein the control unit is configured to:
determine a musical note corresponding to the at least one portion.
36. The apparatus of claim 35, wherein the control unit is further configured to:
present sound data corresponding to the musical note.
37. The apparatus of claim 35, wherein the control unit is further configured to:
determine another musical note corresponding to the at least one portion, the another musical note associated with another musical instrument; and
combine at least a portion of the musical note and at least a portion of the other musical note to produce a mixed musical note.
38. The apparatus of claim 37, wherein the control unit is further configured to:
present sound data corresponding to the mixed musical note.
39. The apparatus of claim 31, where in the control unit is configured to:
store at least a portion of sound data corresponding to the musical property in a musical file format.
40. A non-transitory machine-readable storage device storing instructions that, when executed by one or more processors, cause the one or more processors to perform operations comprising:
displaying a representation of at least one portion of a musical instrument via a display;
identifying an input indicating at least a direction thereof on or in proximity of a specified area of the display; and
determining a musical property associated with the representation based at least in part
US13/908,253 2008-09-01 2013-06-03 Song writing method and apparatus using touch screen in mobile terminal Abandoned US20130263037A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR10-2008-0085826 2008-09-01
KR20080085826A KR101488257B1 (en) 2008-09-01 2008-09-01 A method for composing with touch screen of mobile terminal and an apparatus thereof

Publications (1)

Publication Number Publication Date
US20130263037A1 true US20130263037A1 (en) 2013-10-03

Family

ID=41433153

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/547,556 Expired - Fee Related US8462128B2 (en) 2008-09-01 2009-08-26 Song writing method and apparatus using touch screen in mobile terminal
US13/908,253 Abandoned US20130263037A1 (en) 2008-09-01 2013-06-03 Song writing method and apparatus using touch screen in mobile terminal

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US12/547,556 Expired - Fee Related US8462128B2 (en) 2008-09-01 2009-08-26 Song writing method and apparatus using touch screen in mobile terminal

Country Status (4)

Country Link
US (2) US8462128B2 (en)
EP (1) EP2159785B1 (en)
KR (1) KR101488257B1 (en)
CN (2) CN101668058A (en)

Families Citing this family (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10438448B2 (en) * 2008-04-14 2019-10-08 Gregory A. Piccionielli Composition production with audience participation
KR101657963B1 (en) * 2009-12-08 2016-10-04 삼성전자 주식회사 Operation Method of Device based on a alteration ratio of touch area And Apparatus using the same
KR101679239B1 (en) * 2010-07-06 2016-11-24 삼성전자주식회사 Apparatus and method for playing on musical instruments by augmented reality in portable terminal
US8626324B2 (en) * 2010-09-17 2014-01-07 Apple Inc. Altering sound output on a virtual music keyboard
US8516386B2 (en) 2010-09-29 2013-08-20 Apple Inc. Scrolling virtual music keyboard
US8426716B2 (en) 2011-01-07 2013-04-23 Apple Inc. Intelligent keyboard interface for virtual musical instrument
KR101873405B1 (en) * 2011-01-18 2018-07-02 엘지전자 주식회사 Method for providing user interface using drawn patten and mobile terminal thereof
CN102800337A (en) * 2011-05-25 2012-11-28 汉王科技股份有限公司 Method and equipment for playing keytone
US9324310B2 (en) * 2011-07-07 2016-04-26 Drexel University Multi-touch piano keyboard
TW201316240A (en) * 2011-10-06 2013-04-16 Rich Ip Technology Inc Touch processing method and system using graphic user interface image
US8710344B2 (en) * 2012-06-07 2014-04-29 Gary S. Pogoda Piano keyboard with key touch point detection
CN103280133A (en) * 2013-06-14 2013-09-04 苏州旭宇升电子有限公司 Instrument device with touch function
GB2516029A (en) * 2013-07-08 2015-01-14 Ibm Touchscreen keyboard
CN103885663A (en) * 2014-03-14 2014-06-25 深圳市东方拓宇科技有限公司 Music generating and playing method and corresponding terminal thereof
KR20160017461A (en) * 2014-08-06 2016-02-16 삼성전자주식회사 Device for controlling play and method thereof
CN105607855A (en) * 2015-12-22 2016-05-25 曾旭辉 Handwritten note input system and input method thereof
US10026385B2 (en) * 2016-08-26 2018-07-17 Dean Martin Hovey Smart music device and process that allows only key correct notes and scales to be played
US10380982B1 (en) * 2016-08-26 2019-08-13 Dean Martin Hovey Smart music device and process that allows only key correct notes and scales to be played
US20210407473A1 (en) * 2017-08-04 2021-12-30 Eventide Inc. Musical Instrument Tuner
US10991349B2 (en) * 2018-07-16 2021-04-27 Samsung Electronics Co., Ltd. Method and system for musical synthesis using hand-drawn patterns/text on digital and non-digital surfaces
CN112703745A (en) * 2018-11-27 2021-04-23 深圳市柔宇科技股份有限公司 Sound box
CN109920397B (en) * 2019-01-31 2021-06-01 李奕君 System and method for making audio function in physics
CN111415643B (en) * 2020-04-26 2023-07-18 Oppo广东移动通信有限公司 Notice creation method, device, terminal equipment and storage medium
TR202018879A1 (en) * 2020-11-24 2022-06-21 Anadolu Ueniversitesi Virtual musical key instrument.

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080254824A1 (en) * 2005-02-02 2008-10-16 Aurelio Rotolo Moraes Mobile Communication Device with Musical Instrument Functions
US7453035B1 (en) * 2005-01-07 2008-11-18 Apple Inc. Methods and systems for providing musical interfaces
US20090114079A1 (en) * 2007-11-02 2009-05-07 Mark Patrick Egan Virtual Reality Composer Platform System
US20100180224A1 (en) * 2009-01-15 2010-07-15 Open Labs Universal music production system with added user functionality

Family Cites Families (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3725560A (en) * 1972-03-09 1973-04-03 Jasper Electronics Mfg Corp Chord playing organ
US4065993A (en) * 1974-12-26 1978-01-03 Nippon Gakki Seizo Kabushiki Kaisha Electronic organ with a three-finger chord and one-finger automatic chord playing mode selector
JPH0728483A (en) * 1993-07-14 1995-01-31 Pioneer Electron Corp Musical sound generating device
JP2000278391A (en) * 1999-03-26 2000-10-06 Nec Saitama Ltd Portable telephone set having back handwriting input function
JP2002073023A (en) * 2000-08-28 2002-03-12 Kenwood Corp Portable telephone and method for creating melody in portable telephone
CN1402505A (en) * 2001-08-24 2003-03-12 上海大霸实业有限公司 Method for inputting music in handset
KR20040086941A (en) * 2003-04-03 2004-10-13 엘지전자 주식회사 Performance method in mobile telecommunication terminal equipment
JP2005072809A (en) 2003-08-21 2005-03-17 Toshiba Corp Digital camera
JP4179268B2 (en) * 2004-11-25 2008-11-12 カシオ計算機株式会社 Data synthesis apparatus and data synthesis processing program
KR20060065255A (en) * 2004-12-10 2006-06-14 브이케이 주식회사 Mobile communication terminal for performing musical instruments change music play and method thereof
KR20070017803A (en) * 2005-08-08 2007-02-13 엘지전자 주식회사 Apparatus and Method for Playing Music using Mobile Terminal
JP2007057788A (en) * 2005-08-24 2007-03-08 Bridgestone Corp Manufacturing method of panel for information display and panel for information display
KR20070039692A (en) * 2005-10-10 2007-04-13 주식회사 팬택 Mobile communication terminal capable of providing song - making, accompaniment and recording function
CN2859661Y (en) * 2005-10-27 2007-01-17 余张强 Music write and display integrated computer display device
JP4192952B2 (en) * 2006-02-07 2008-12-10 ヤマハ株式会社 Mobile phone
KR100701520B1 (en) * 2006-06-26 2007-03-29 삼성전자주식회사 User Interface Method Based on Keypad Touch and Mobile Device thereof
US20080184872A1 (en) * 2006-06-30 2008-08-07 Aaron Andrew Hunt Microtonal tuner for a musical instrument using a digital interface
KR101554221B1 (en) * 2009-05-11 2015-09-21 삼성전자주식회사 Method for playing a musical instrument using potable terminal and apparatus thereof

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7453035B1 (en) * 2005-01-07 2008-11-18 Apple Inc. Methods and systems for providing musical interfaces
US20080254824A1 (en) * 2005-02-02 2008-10-16 Aurelio Rotolo Moraes Mobile Communication Device with Musical Instrument Functions
US20090114079A1 (en) * 2007-11-02 2009-05-07 Mark Patrick Egan Virtual Reality Composer Platform System
US20100180224A1 (en) * 2009-01-15 2010-07-15 Open Labs Universal music production system with added user functionality

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
iAno thru Youtube.com, "iAno for iPhone," February 18, 2008, Retrieved from https://www.youtube.com/watch?v=j-CFD_1tP9w *
iAno, "Native iPhone App - iAno turns iPhone into a Touchscreen Piano," February 17, 2008, Retrieved from http://www.iphonehacks.com/2008/02/iphone-piano.html *
MooCowMusic.com thru Youtube.com, "MooCowMusic: Band," July 2, 2008, Retrieved from https://www.youtube.com/watch?v=KXqbf1F1x9U *
MooCowMusic.com, "MooCowMusic: Band," July 29, 2008, Retrieved from https://web.archive.org/web/20080729040642/http://moocowmusic.com/Band/index.html *

Also Published As

Publication number Publication date
KR20100026711A (en) 2010-03-10
CN101668058A (en) 2010-03-10
EP2159785A2 (en) 2010-03-03
EP2159785B1 (en) 2015-04-01
KR101488257B1 (en) 2015-01-30
EP2159785A3 (en) 2010-05-05
CN105185365A (en) 2015-12-23
US8462128B2 (en) 2013-06-11
US20100053105A1 (en) 2010-03-04

Similar Documents

Publication Publication Date Title
US8462128B2 (en) Song writing method and apparatus using touch screen in mobile terminal
US8300841B2 (en) Techniques for presenting sound effects on a portable media player
US20080215980A1 (en) User interface providing method for mobile terminal having touch screen
US8548531B2 (en) Method and system of creating customized ringtones
US20060293089A1 (en) System and method for automatic creation of digitally enhanced ringtones for cellphones
US20140006948A1 (en) Method and mobile phone for capturing audio file or video file
KR101130519B1 (en) Media portion selection system and method
US20080104112A1 (en) Method and Apparatus for Music Play List Control
KR20080025772A (en) Music message service transfering/receiving method and service support sytem using the same for mobile phone
KR20090029135A (en) The method of editing playlist and the multimedia replaying apparatus thereof
KR100783113B1 (en) Method for shortened storing of music file in mobile communication terminal
US20070173291A1 (en) Information communication terminal, information processing method and information processing program
KR101507468B1 (en) Sound data generating system based on user's voice and its method
KR100506228B1 (en) Mobile terminal and method for editing and playing music
EP2933992A1 (en) Method, apparatus and computer program for selecting an audio track
US20080032676A1 (en) Mobile electronic device with telephone and music function
KR101393714B1 (en) Terminal and method for playing music thereof
US20090271395A1 (en) Media file searching system and method for a mobile phone
KR100628053B1 (en) Wireless communication terminal with auto-generation of digital music information and method of auto-generation of digital music information using same
JP2010231027A (en) Electronic musical device
JP2005338126A (en) Device, method, and program for editing musical performance data
JP2004163511A (en) Mobile terminal device
JP4093404B2 (en) Mobile communication terminal
KR100705172B1 (en) Apparatus and method audio recording regeneration
JP2002297134A (en) Portable terminal device

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION