WO2017098743A1 - Service management system, computer readable storage medium, and service management method - Google Patents

Service management system, computer readable storage medium, and service management method Download PDF

Info

Publication number
WO2017098743A1
WO2017098743A1 PCT/JP2016/067122 JP2016067122W WO2017098743A1 WO 2017098743 A1 WO2017098743 A1 WO 2017098743A1 JP 2016067122 W JP2016067122 W JP 2016067122W WO 2017098743 A1 WO2017098743 A1 WO 2017098743A1
Authority
WO
WIPO (PCT)
Prior art keywords
sound
unit
music
volume
reproduction
Prior art date
Application number
PCT/JP2016/067122
Other languages
French (fr)
Japanese (ja)
Inventor
健一郎 堀
Original Assignee
健一郎 堀
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP2015239691A external-priority patent/JP2016110154A/en
Application filed by 健一郎 堀 filed Critical 健一郎 堀
Publication of WO2017098743A1 publication Critical patent/WO2017098743A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10KSOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
    • G10K15/00Acoustics not otherwise provided for
    • G10K15/02Synthesis of acoustic waves
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10KSOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
    • G10K15/00Acoustics not otherwise provided for
    • G10K15/04Sound-producing devices
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0272Voice signal separating

Definitions

  • the present invention relates to a service management system, a computer-readable storage medium, and a service management method for managing a service that provides music.
  • a karaoke service that plays accompaniment sounds is known (see, for example, Patent Document 1).
  • the sound source used in the karaoke service is not a music master having mixed sound data including accompaniment sounds and vocals, but data consisting only of accompaniment sounds newly created for karaoke.
  • a sound including a sound to be adjusted and a sound other than the sound to be adjusted is a mixed sound
  • a master file including data for reproducing the mixed sound is obtained.
  • a stored master recording unit, a source storage unit storing a source of service for providing sound adjustment content to a user terminal, a playback record storage unit storing the number of times of sound adjustment content played by the user terminal, and the user terminal And a connected control unit. Then, the control unit specifies a range of the adjustment target sound in the master file stored in the master storage unit, and based on the specified range, adjusts the volume of the adjustment target sound in the mixed sound.
  • a mask file for individual adjustment is generated, the sound adjustment content including the mask file is provided to the user terminal, the number of reproductions is stored in the reproduction result storage unit, and the resource stored in the resource storage unit And a distribution amount to be distributed to the distribution target person of the adjustment target sound, based on the number of times of reproduction stored in the reproduction result storage unit.
  • a computer-readable storage medium for solving the above-described problem is a master file that includes a sound that includes a sound to be adjusted and a sound other than the sound to be adjusted as a mixed sound, and includes data for reproducing the mixed sound
  • a master disk storage unit a resource storage unit that stores resources for providing sound adjustment content to a user terminal, a playback record storage unit that stores the number of times the sound adjustment content is played by the user terminal, and the user terminal
  • a computer-readable storage medium storing a program for performing service management using a service management system including a control unit connected to the computer.
  • the control unit When executing the program, the control unit specifies a range of the adjustment target sound in the master file stored in the master storage unit, and based on the specified range, the control unit includes the A mask file for individually adjusting the volume of the adjustment target sound is generated, the sound adjustment content including the mask file is provided to the user terminal, the number of times of reproduction is stored in the reproduction result storage unit, and the resource Based on the resources stored in the storage unit and the number of reproductions stored in the reproduction record storage unit, the distribution amount to be distributed to the distribution target person of the adjustment target sound is calculated.
  • a sound including a sound to be adjusted and a sound other than the sound to be adjusted is a mixed sound
  • a master file including data for reproducing the mixed sound is obtained.
  • a stored master recording unit, a source storage unit storing a source of service for providing sound adjustment content to a user terminal, a playback record storage unit storing the number of times of sound adjustment content played by the user terminal, and the user terminal service management is performed using a service management system including a connected control unit.
  • the control unit specifies a range of the adjustment target sound in the master file stored in the master storage unit, and the adjustment is performed in the mixed sound based on the specified range.
  • a mask file for individually adjusting the volume of the target sound is generated, the sound adjustment content including the mask file is provided to the user terminal, the number of reproductions is stored in the reproduction result storage unit, and the resource storage A distribution amount to be distributed to the distribution target person of the adjustment target sound is calculated based on the resources stored in the section and the number of times of reproduction stored in the reproduction performance storage section.
  • the lineblock diagram showing a 1st embodiment of a service management system The block diagram of the data which an original disk memory
  • generation process The graph which shows the range of a mask file.
  • the flowchart which shows the procedure of a synchronous lyrics process.
  • regeneration of a group The figure which shows the procedure of the process in group reproduction
  • the figure which shows the procedure of the process by a service management method The graph which shows the range in fixed mode.
  • the service management system provides a master karaoke service that is an example of a content providing service.
  • the service management system uses a music master provided by a record company or the like, and provides a mixed sound in which the volume of vocals is individually adjusted with respect to the volume of other sounds.
  • Vocal is a singing voice sung by a singer who is an example of a performer, and is an example of a sound to be adjusted.
  • the music master disc contains data for playing mixed sounds including vocals and other sounds.
  • the music master may be a music master file, or may be a box format file including a music master file and a video master file, such as a promotion video.
  • the content server 10 is a computer system and provides a master file.
  • the master disk file is content for providing data for reproducing the mixed sound.
  • the content is, for example, a digital file such as a promotion video created in the mp4 format.
  • the person-in-charge terminal 30 is a computer terminal and is used for editing data for reproducing music.
  • the editing of music is adjustment of the timing at which telops such as lyrics are displayed and vocal mask adjustment.
  • the person-in-charge terminal 30 includes a control unit, an output unit, and an input unit.
  • the output unit is configured by a display, for example, and outputs various information.
  • the input unit is configured by a keyboard and a pointing device, and acquires an instruction input by a person in charge.
  • the user terminal 40 is a computer terminal and is used by a user who uses a karaoke service.
  • the user terminal 40 is a business (BtoB) device or a customer (BtoC) device.
  • Business equipment is a computer terminal installed in a karaoke box or the like. Devices for customers are smartphones and personal computers possessed by individual users.
  • the user terminal 40 downloads a karaoke file including a music file and a video file from the management server 20 as an example of sound adjustment content that is content for adjusting the volume of the adjustment target sound.
  • the karaoke file includes a vocal mask file as an example of a mask file.
  • the user terminal 40 includes a music playback unit 41 and a vocal volume adjustment unit 42.
  • the music playback unit 41 plays back music and video using the downloaded music file and video file.
  • the vocal volume adjustment unit 42 performs a process of individually adjusting the volume of the vocal in the mixed sound based on the vocal mask file.
  • the process of adjusting the volume of the vocal includes a process of deleting the vocal.
  • the vocal mask file is a file for specifying the vocal range included in the music.
  • the management server 20 is a computer system and is used for providing a karaoke file by a karaoke service provider.
  • the management server 20 includes a control unit 21, a master disk storage unit 22, a lyrics information storage unit 23, a music storage unit 24, a reproduction performance storage unit 25, a resource storage unit 26, a distribution management information storage unit 27, and a distribution result information storage unit. 28.
  • the control unit 21 includes control means including a CPU, a RAM, a ROM, and the like, and performs each process of a management stage, vocal processing stage, lyrics management stage, editor stage, reproduction management stage, and distribution management stage.
  • the control unit 21 functions as a management unit 210, a vocal processing unit 211, a lyrics management unit 212, an editor 213, a reproduction management unit 214, and a distribution management unit 215 by executing a service management program.
  • the management unit 210 executes processing for storing the master file acquired from the content server 10 in the master storage unit 22.
  • the vocal processing unit 211 executes a vocal mask file generation process for identifying the vocal range and adjusting the vocal volume.
  • the lyrics management unit 212 executes a process of adding lyrics to the file so that the timing at which the lyrics are displayed matches the timing at which the vocals are reproduced.
  • the editor 213 executes processing for adjusting the vocal mask and processing for supporting adjustment of the addition position of the lyrics.
  • the editor 213 is connected to the person-in-charge terminal 30 used by the person in charge of editing.
  • the person in charge of the editor uses the person-in-charge terminal 30 to confirm the setting of the vocal mask and the position where the lyrics are added.
  • the editing person in charge adjusts the setting position of the vocal mask and the addition position of the lyrics using the person-in-charge terminal 30.
  • the setting position of the vocal mask is a sound range masked by the vocal mask and a reproduction time for which the vocal mask is set.
  • the playback management unit 214 executes processing for managing the playback status of the music on the user terminal 40.
  • the reproduction management unit 214 holds information for calculating the reproduction determination time.
  • the reproduction determination time is a time for determining whether or not the music has been reproduced.
  • the distribution management unit 215 executes a process of distributing the resources to each distribution target person according to the reproduction status of the music.
  • the distribution target person is, for example, a copyright management organization (for example, JASRAC (registered trademark)), a performer, or a record company.
  • the distribution management unit 215 holds data regarding a distribution rate for determining a distribution amount for each party. Next, the configuration of information stored in each storage unit will be described.
  • the master storage unit 22 stores a master file 220 such as a promotion video.
  • the management server 20 acquires the master file 220 from the content server 10
  • the management server 20 stores the master file 220.
  • the master file 220 includes data related to music codes, attributes, and compressed moving images.
  • the music code data area stores data relating to an identifier for specifying a music.
  • the attribute data area stores data related to the attributes of music.
  • the attribute of the music is information relating to the record company, performer, composer, songwriter, and the like.
  • the compressed moving image data area stores compressed data for music and video.
  • the compressed data is created in the mp4 format, for example.
  • the lyrics information storage unit 23 stores a lyrics file 230 related to lyrics.
  • the lyrics file 230 is stored when the lyrics information is registered in the management server 20.
  • the lyric file 230 includes data related to music codes and lyrics.
  • the music code data area stores data relating to an identifier for specifying a music.
  • the lyric data area stores lyric information related to lyric text and ruby. Lyric text and ruby indicated by the lyric information function as a telop corresponding to vocals.
  • the music storage unit 24 stores music management data 240 to be provided to the user terminal 40.
  • the music management data 240 is stored when karaoke data is generated from the master music.
  • the music management data 240 is associated with a music code and includes an HLS file 241, an XML file 242, a vocal mask file 243, and a music management record 244.
  • the song code data area stores data relating to an identifier for specifying a song.
  • the HLS file 241 includes data obtained by transcoding the compressed video of the master disc.
  • the HLS file 241 includes data according to the HD format for high speed lines and data according to the SD format for low speed lines.
  • the XML file 242 stores song attribute information and lyrics synchronization information indicating lyrics associated with a time index that is a reproduction time.
  • the XML file 242 includes music composition information in which each of A melody, B melody, and C melody, which are musical composition elements, is associated with a time index.
  • the XML file 242 includes chapter information in which each chapter in the medley music is associated with a time index.
  • the XML file 242 includes the following description. • Lyrics: Lyrics file versions are specified. -Header: The music meta information and style definition are stored. Song meta (song_meta): The meta information of the music is stored. Style: Style definitions and references are stored.
  • -Lyrics track (lyrics_track): The lyrics body is stored.
  • ⁇ Verse The lyrics break is stored.
  • -Page Lyrics for one page and display / erase timing are defined.
  • -Line Lyric definitions for one line are stored.
  • -Ruby (ruby) Ruby definitions are stored.
  • Word Lyric text and wipe processing timing are defined.
  • the vocal mask file 243 stores the vocal range by associating the time index with the vocal range.
  • the vocal mask file 243 is used for specifying a vocal line.
  • the music management record 244 is data relating to the performer, voice quality, tone, performance time, video usage, medley attribute, status, scene, and sound range.
  • the performer data area stores data relating to an identifier such as a singer name for identifying a performer such as a singer.
  • the voice quality data area stores data related to the feature amount of the singer's voice.
  • the feature amount for the singer's voice is, for example, a statistical value calculated from the main melody.
  • Tune data area stores data related to tempo and rhythm of music.
  • the data relating to the tempo and rhythm of the music is, for example, a feature value that is a statistical value calculated from the music.
  • the scene data area stores information for identifying, for example, the season associated with the song, the scenery such as the sea, mountains, and sky associated with the song, and the emotions such as joy, sadness, and resentment associated with the song.
  • the range data area stores data related to the range required for the vocal of the music.
  • the performance time data area stores data related to the performance time, which is the time required to reproduce the music from the beginning to the end.
  • the video use data area stores a flag for identifying whether or not the music master disk uses video in the music HLS file 241.
  • the medley attribute data area stores information for identifying whether or not the music is medley.
  • the medley attribute data area stores the music code of the music constituting the medley.
  • the status data area stores a flag for specifying the creation status of the music HLS file 241 and XML file 242.
  • the status data area stores a “before editing” flag for a file immediately after registration, and a “provided” flag for a file after editing.
  • the reproduction record storage unit 25 stores reproduction record data 250 relating to the use record of the karaoke file.
  • the reproduction result data 250 is stored when the user terminal 40 uses a karaoke file.
  • the reproduction performance data 250 includes a user performance record 251 and a music performance record 252.
  • the user performance record 251 includes data related to playback date / time, user code, user code, and music code.
  • the reproduction date / time data area stores data related to the date and time when the music was reproduced in the user terminal 40.
  • the user code data area stores data relating to an identifier for specifying the user who has played the music.
  • the user code data area stores data relating to an identifier for identifying the user who has played the music.
  • the music code data area stores data relating to an identifier for specifying the reproduced music.
  • the track record 252 includes data related to the track code and the number of times of playback.
  • the music code data area stores data relating to an identifier for specifying the reproduced music.
  • the reproduction frequency data area stores data relating to the total number of times that a music piece has been reproduced.
  • the resource storage unit 26 stores a resource management file 260 relating to a karaoke usage fee acquired from the user.
  • Karaoke usage fees are an example of funding.
  • the resource management file 260 is stored when use registration of the karaoke service is performed.
  • the resource management file 260 includes data relating to a user code, a contract form, and an information fee.
  • the user code data area stores data relating to an identifier for identifying a user who pays a usage fee.
  • the contract form data area stores a flag for specifying a contract form with the user.
  • the contract form with the user is, for example, a contract form for customers such as individual billing, fixed amount payment (unlimited singing).
  • the contract form with a user is a contract form for business like the fixed amount payment for every user terminal 40, for example.
  • the information fee data area stores data related to the information fee, which is the amount to be paid when using the service, based on the contract form with the user.
  • the distribution management information storage unit 27 stores a distribution management file 270 for specifying the distribution of the funds for each distribution target person.
  • the distribution management file 270 is stored when payment conditions with each distribution target person are registered.
  • the distribution management file 270 includes data relating to music codes, payees, and distribution conditions.
  • the song code data area stores data relating to an identifier for specifying a song.
  • the payee data area stores data related to payees of distributions related to music.
  • the payee of the distribution money is, for example, a player or a record company that is an example of a distribution target person.
  • the distribution condition data area stores data relating to a calculation method of distribution paid to the payee.
  • a ratio to the charging fee is stored as a payment condition.
  • a method for calculating the distribution money based on the distribution source is stored as the payment condition.
  • the distribution record information storage unit 28 stores a distribution record record 280 related to the distribution record for the distribution destination.
  • the distribution record 280 is stored when payment is made.
  • the distribution record 280 includes data relating to distribution date, distribution destination, and distribution amount.
  • the distribution date data area stores data relating to the date of distribution of the funds.
  • the distribution destination data area stores data relating to an identifier for specifying the distribution destination of the distribution money.
  • the identifier for specifying the distribution destination is data for specifying the performer or record company as the distribution destination.
  • the distribution amount data area stores data related to the amount to be distributed to the distribution destination.
  • the master karaoke data generation process will be described with reference to FIG.
  • the control unit 21 executes a process of acquiring a promotion video that is an example of a master file (step S1-1). Specifically, the management unit 210 acquires a promotion video from the content server 10 as an mp4 file. Further, the management unit 210 acquires attribute information of the promotion video. Then, the management unit 210 registers the acquired promotion video and attribute information in the master disk storage unit 22.
  • the control unit 21 executes transcoding processing (step S1-2). Specifically, the management unit 210 converts the mp4 file registered in the master storage unit 22 into an HLS file. The management unit 210 converts the mp4 file into an HD file used for a high-speed line and an SD file used for a low-speed line. Then, the management unit 210 registers the generated HLS file in the music storage unit 24 in association with the music code.
  • control unit 21 executes main melody estimation processing (step S1-3). Specifically, the vocal processing unit 211 converts the mp4 file into a predetermined file format. Then, the vocal processing unit 211 estimates a main melody such as a vocal line in the music.
  • control unit 21 executes a vocal mask file creation process (step S1-4). Specifically, the vocal processing unit 211 identifies the vocal sound range based on the estimated main melody.
  • the vocal processing unit 211 creates a vocal mask file that stores the specified vocal range.
  • the vocal processing unit 211 stores the created vocal mask file in the music storage unit 24 in association with the music code.
  • control unit 21 executes a voice quality registration process (step S1-5). Specifically, the vocal processing unit 211 calculates a feature quantity of the voice quality of the vocal included in the music. Then, the vocal processing unit 211 registers the calculated voice quality in the music management record 244 stored in the music storage unit 24.
  • control unit 21 executes a lyrics allocation process (step S1-6). Specifically, the lyrics management unit 212 acquires the lyrics file 230 in which the song code of the processing target song is stored from the lyrics information storage unit 23.
  • the lyrics management unit 212 associates the reproduction time corresponding to the vocal range specified by the vocal processing unit 211 with the lyrics, and generates an XML file 242 that is a lyrics synchronization file in which lyrics are assigned. Then, the lyrics management unit 212 registers the XML file 242 in the music storage unit 24.
  • control unit 21 executes an editing process (step S1-7). Specifically, the management unit 210 stores the “before editing” flag in the music management record 244 after the voice quality registration process and the lyrics allocation process are completed. Then, the editor 213 provides the created HLS file 241, XML file 242, and vocal mask file 243 to the person-in-charge terminal 30.
  • the person-in-charge terminal 30 reproduces the karaoke file according to the operation of the person in charge of the editor, and confirms whether or not the volume of the vocal sound changes depending on the vocal mask, whether or not the vocal sound and the lyrics are synchronized.
  • the editor 213 corrects the XML file 242 and the vocal mask file 243 according to the operation of the editor.
  • the editor 213 that has acquired the editing end information from the person-in-charge terminal 30 stores the “provided” flag in the music management record 244.
  • the vocal processing unit 211 acquires singer information stored in the attribute data area of the master file 220 (step S2-1).
  • the vocal processing unit 211 acquires the singer information of the singer.
  • the vocal processing unit 211 acquires singer information of each singer.
  • the vocal processing unit 211 identifies the tone of the music to be processed.
  • the vocal processing unit 211 extracts, from the music storage unit 24, the music management record 244 in which the processing target singer is stored in the singer data area.
  • the vocal processing unit 211 specifies a music similar to the music tone of the music to be processed as a similar music (step S2-2).
  • the control unit 21 executes processing for specifying the voice quality of the processing target singer (step S2-3).
  • the voice quality identification process uses the following two methods.
  • control unit 21 executes a process for specifying voice quality based on similar music (step S3-1). Specifically, the vocal processing unit 211 acquires a vocal range that is the voice quality of a singer stored in the music management record 244 of similar music.
  • control unit 21 executes a process for specifying the voice quality by the singing location (step S3-2). Specifically, the vocal processing unit 211 identifies a singing area in which the sound of only the processing target singer is included among the sounds included in the HLS file 241. Then, the vocal processing unit 211 specifies the vocal sound range in the specified singing area.
  • control unit 21 executes a first mask generation process (step S2-4).
  • the control unit 21 generates a vocal mask file corresponding to the vocal range through the entire song using the specified vocal range.
  • control unit 21 determines whether or not the vocal mask is sufficient (step S2-5). Specifically, the vocal processing unit 211 uses the generated vocal mask to play a song with a reduced vocal volume, and confirms whether or not the singer's singing voice can be recognized.
  • the vocal processing unit 211 confirms whether or not the main melody of the music is distorted before and after the vocal mask. If the singing voice cannot be recognized or the main melody is not distorted, the vocal processing unit 211 determines that the vocal mask is sufficient.
  • the control unit 21 ends the vocal mask file generation processing for the processing target singer.
  • step S2-6 the control unit 21 specifies a vocal range for each vocal phrase, and generates a vocal mask file corresponding to each vocal range.
  • step S2--7 the control unit 21 executes processing for determining whether or not the vocal mask is sufficient.
  • the control unit 21 ends the vocal mask file generation processing for the processing target singer.
  • step S2-8) the control unit 21 executes a third mask generation process.
  • the control unit 21 estimates a vocal sound range for each predetermined time interval, and generates a vocal mask file that follows the dynamically specified vocal sound range.
  • the control part 21 repeats the above process about all the singers contained in a process target music.
  • the first mask generation process will be described with reference to FIG.
  • the vocal processing unit 211 estimates that the voice quality of the singer specified based on the similar music or the singing part is the vocal sound range of this music (step S4-1).
  • the vocal processing unit 211 generates a vocal mask file for masking a single vocal range (step S4-2).
  • a vocal mask 501 for masking a single vocal range is generated throughout the music.
  • the horizontal axis indicates time
  • the vertical axis indicates a sound range.
  • the upper thick line indicates the guitar sound
  • the lower broken line indicates the drum sound range.
  • the vocal processing unit 211 identifies a playback time zone in which the vocal is included based on the vocal line.
  • the vocal processing unit 211 identifies a single group of continuous singing as a phrase (step S5-1).
  • the vocal processing unit 211 identifies a vocal range for each phrase based on the vocal line (step S5-2).
  • the vocal processing unit 211 generates a vocal mask file for masking the vocal range estimated for each phrase (step S5-3).
  • vocal masks 511 and 512 created for each phrase are generated.
  • the horizontal axis indicates time
  • the vertical axis indicates a sound range.
  • the upper thick line indicates the guitar sound
  • the lower broken line indicates the drum sound range.
  • the vocal processing unit 211 divides the music for each time step that is a predetermined time (step S6-1). And the control part 21 specifies the time step of a process target sequentially, and repeats the following processes.
  • the control unit 21 executes a process for specifying a vocal range included in the specified time step (step S6-2). Next, the control unit 21 executes a process of generating a vocal mask file that masks the specified vocal range (step S6-3).
  • vocal masks 521 and 522 created for each time step are generated.
  • the mask area dynamically changes with respect to the vocal sound area.
  • the horizontal axis indicates time
  • the vertical axis indicates the sound range.
  • the upper thick line indicates the guitar sound
  • the lower broken line indicates the drum sound range.
  • the time series graph 600 shows an example including a female vocal and a male vocal.
  • a time series graph 601 shows an example in which the volume of a female vocal is individually lowered using a female vocal vocal mask file.
  • the time series graph 602 shows an example in which the volume of the male vocal is individually lowered using the vocal mask file of the male vocal.
  • a time series graph 603 shows an example in which the volume of both vocals is individually lowered using both female and male vocal mask files.
  • the lyrics management unit 212 extracts the lyrics file 230 in which the song code of the processing target song is stored from the lyrics information storage unit 23, and acquires the lyrics phrase to be assigned to the song (step S7-1).
  • the lyrics management unit 212 estimates a singing voice section that includes a singing voice and a non-singing voice section that does not include a singing voice based on the vocal line. Then, the lyrics management unit 212 uses the singing voice section to specify the start time that is the start position of the phrase and the end time that is the end position (step S7-2).
  • the lyrics management unit 212 generates a description in which characters and ruby are assigned to the start position and end position of the phrase in the XML file (step S7-3). And the control part 21 performs the determination process about whether all the lyrics were allocated (step S7-4).
  • step S7-4 When the control unit 21 determines that lyrics that have not been assigned remain (in the case of “NO” in step S7-4), the control unit 21 performs processing subsequent to the processing for specifying the assigned lyrics (step S7-1). repeat. On the other hand, if it is determined that all the lyrics have been assigned (“YES” in step S7-4), the control unit 21 ends the synchronized lyrics process.
  • the XML file 700 shown in FIG. 19 is generated by the above processing.
  • the description 701 includes song metadata relating to lyrics, composition, arrangement, etc. stored in the attribute data area of the master file 220.
  • the description 702 includes music metadata relating to the tempo stored in the attribute data area of the master file 220.
  • the description 703 includes a description for wiping “August” with “hachigatsu” added as ruby over a predetermined period.
  • the description 704 includes a description for wiping “sun” with “taiyo” added as ruby over a predetermined period.
  • the description 705 includes a description for wiping “NO” over a predetermined period
  • the description 706 includes a description for wiping “NO” over a predetermined period
  • the description 707 includes a description for changing a page to be displayed on one screen during karaoke playback.
  • an XML file for displaying lyrics in accordance with the progress of the music is generated.
  • the playback determination process will be described with reference to FIG. This process is performed when a karaoke file is reproduced on the user terminal 40.
  • the reproduction management unit 214 provides a karaoke file to the user terminal 40 and instructs the user terminal 40 to reproduce the karaoke file (step S8-1).
  • the playback management unit 214 acquires the playback status from the user terminal 40, and determines whether or not playback has ended (step S8-2). When all the melody in the music ends or when the music ends in the middle, the reproduction management unit 214 determines that the reproduction has ended.
  • the playback management unit 214 calculates the playback time from the start time of the music to the time when the playback end is detected (step S8-3).
  • the reproduction management unit 214 uses the music storage unit 24 to extract a music management record 244 in which the music code of the reproduced music is stored. Then, the reproduction management unit 214 acquires the length of the song that is the performance time from the extracted song management record 244 (step S8-4).
  • the reproduction management unit 214 specifies a reproduction determination time based on the performance time (step S8-5).
  • the playback management unit 214 uses 2/3, which is a performance reference criterion, as a playback determination time for a song whose performance time is less than 60 seconds.
  • the playback management unit 214 uses 40 seconds as a playback determination time for a musical piece having a performance time of 60 seconds or longer.
  • the playback management unit 214 executes processing for determining whether or not the playback determination time has elapsed (step S8-6).
  • the playback management unit 214 When the playback time has passed the playback determination time (in the case of “YES” in step S8-6), the playback management unit 214 generates a user result record 251 and registers it in the playback result storage unit 25. Furthermore, the reproduction management unit 214 increments the number of reproductions of the music record 252 for this music (step S8-7).
  • the playback management unit 214 newly generates a music record record 252 for the music played for the first time and registers it in the playback record storage unit 25. In this case, the reproduction management unit 214 sets an initial value “1” for the number of reproductions of the music record record 252.
  • step S8-6 playback management unit 214 skips playback count processing (step S8-7) and performs playback determination processing. finish.
  • the distribution management process is executed every predetermined period such as one month.
  • the distribution management unit 215 specifies a piece of music for individual charging in the distribution conditions of the distribution management file 270 stored in the distribution management information storage unit 27 (step S9-1).
  • the distribution management unit 215 calculates the distribution money based on the individual rules stored in the distribution management file 270 (step S9-2).
  • the distribution management unit 215 identifies a medley song based on the medley attribute stored in the song management record 244 of the song storage unit 24 (step S9-3).
  • the distribution management unit 215 obtains the music code of the medley constituent music that is an individual music constituting the medley music from the medley attribute data area in the music management record 244 of the medley music (step S9-4).
  • the distribution management unit 215 calculates the number of distributions obtained by dividing the number of times the medley music has been played by the number of medley constituent songs. Then, the distribution management unit 215 adds the calculated number of distributions to the number of reproductions of the music record record 252 in which the music code of each medley constituent music is stored (step S9-5).
  • control unit 21 executes a distribution process of the fixed amount music for the customer (step S9-6). Specifically, the distribution management unit 215 distributes resources by dividing individual billing and unlimited singing.
  • control unit 21 executes a distribution process for a fixed amount music for business (step S9-7). Specifically, the distribution management unit 215 distributes resources according to the amount of monthly information for each terminal.
  • a fixed amount music distribution process will be described with reference to FIG.
  • a process target period is determined, and the process is performed separately for the customer and the business.
  • the distribution process for the customer and the distribution process for the business differ only in the resources, and the processing procedure is the same.
  • control unit 21 executes an information fee total acquisition process (step S10-1). Specifically, the distribution management unit 215 acquires the total amount of information fee for the fixed amount payment in the resource storage unit 26. The distribution management unit 215 calculates separately for customers and for businesses.
  • the control unit 21 executes a process for calculating the total number of reproductions (step S10-2). Specifically, the distribution management unit 215 specifies the music code of the music for which the individual charge is not stored in the distribution management file 270 stored in the distribution management information storage unit 27. Next, the distribution management unit 215 obtains a music record record 252 in which the specified music code is stored from the reproduction record storage unit 25. Then, the distribution management unit 215 sums the number of reproductions stored in the acquired music record record 252.
  • control part 21 repeats the following processes for every music.
  • control unit 21 executes a process for obtaining the number of reproductions (step S10-3).
  • the distribution management unit 215 acquires a music record record 252 in which the music code to be processed is stored from the reproduction record storage unit 25.
  • control unit 21 executes a process for calculating the reproduction ratio (step S10-4). Specifically, the distribution management unit 215 calculates the reproduction ratio by dividing the number of reproductions of the music to be processed by the total number of reproductions.
  • the control unit 21 executes a process for calculating the distribution money (step S10-5). Specifically, the distribution management unit 215 acquires the resources for the processing target period from the resource storage unit 26. Here, the distribution management unit 215 calculates the total information fee for “all-you-can sing” for the process for the customer. On the other hand, the distribution management unit 215 calculates the total amount of information for each terminal for business-oriented processing. Then, the distribution management unit 215 calculates the distribution amount by multiplying the total information fee by the reproduction ratio of the music to be processed.
  • control unit 21 executes a calculation process for the copyright management organization in the distribution (step S10-6). Specifically, the distribution management unit 215 calculates the distribution amount for the copyright management organization by multiplying the distribution money by the distribution rate of the copyright management organization. Then, the distribution management unit 215 generates a distribution record record 280 and stores it in the distribution record information storage unit 28.
  • control unit 21 executes a calculation process for the performer in the distribution money (step S10-7). Specifically, the distribution management unit 215 calculates the distribution amount for the performer by multiplying the distribution amount by the distribution rate for the song. Then, the distribution management unit 215 generates a distribution record record 280 and stores it in the distribution record information storage unit 28.
  • control unit 21 executes a process for determining whether or not the processing target uses video (step S10-8). Specifically, the distribution management unit 215 confirms whether or not the video is included in the HLS file 241 stored in the music storage unit 24. When the use flag is set in the video use data area of the music management record 244, the distribution management unit 215 determines that the processing target uses video.
  • the control unit 21 executes calculation processing for the video in the distribution (step S10-9). Specifically, the distribution management unit 215 calculates the distribution amount for the video by multiplying the distribution by the distribution rate for the video. Then, the distribution management unit 215 generates a distribution record record 280 and stores it in the distribution record information storage unit 28.
  • control unit 21 skips the calculation processing for the video in the distribution (step S10-9), and creates the fixed amount music. The distribution process ends.
  • the following effects can be obtained. (1) Since a karaoke file can be generated using the master music, the service user can sing with the same accompaniment as the performer. (2) Even when a plurality of vocals are included, since a vocal mask file for each singer is generated, the volume of the vocal can be adjusted for each singer. (3) Since the second mask generation process is executed when it is determined that the vocal mask is insufficient, distortion of music due to the vocal mask can be suppressed. Further, since the third mask generation process is executed, it is possible to generate a precise vocal mask.
  • the reproduction count process is executed when the reproduction time has passed the reproduction determination time, the appropriate number of reproductions can be counted.
  • the distribution calculation process is executed based on the individual rules, the distribution can be calculated based on the contract with the distribution target person.
  • the voice quality specifying process (step S2-3) performed by the control unit 21 is any one of a voice quality specifying process (step S3-1) using a similar music piece and a voice quality specifying process using a solo location (step S3-2). Either one is acceptable.
  • the control unit 21 When the voice quality is specified by both methods, the control unit 21 generates a vocal mask file using the voice quality specified based on a predetermined rule.
  • the predetermined rule is, for example, to specify a range including both vocal ranges as a voice quality.
  • the distribution conditions included in the distribution management file 270 may be set based on the difficulty of specifying the vocal sound range. For example, when a vocal mask file is generated by the first mask process, a lower distribution rate is set for the performer than when a vocal mask file is generated by the second mask process. Furthermore, when the vocal mask file is generated by the second mask process, a lower distribution ratio is set for the performer than when the vocal mask file is generated by the third mask process. Thereby, the distribution rate can be changed according to the difficulty of generating the vocal mask file.
  • the method for determining whether or not the music has been reproduced may be, for example, determining whether or not the position of the chorus in the music has been reproduced.
  • the control unit 21 specifies the position of the rust based on the specific pattern in the process of estimating the main melody (step S1-3). Then, the specified chorus position is stored in the music management record 244 in association with the time index. Then, the playback management unit 214 determines whether or not the music has been played based on the playback time and the position of the chorus. Even if the playback time has not passed the playback determination time, the playback management unit 214 determines that playback has been performed when rust is played back. -When each user who comprises a single group uses a karaoke service using his user terminal 40, you may adjust the frequency
  • the management server 20 includes a group information storage unit.
  • the group information storage unit stores data related to the user code of each user terminal 40 used as a group with respect to the group management code.
  • the management server 20 acquires the designation of a song that uses the karaoke service
  • the management server 20 stores the parent device information in association with the user code of the user terminal 40 for which the song is designated. Then, the service management system executes a group reproduction management process shown in FIG.
  • the control unit 21 executes a process for registering a group (step S11-1). Specifically, when a plurality of users use the karaoke service using their own user terminals 40, each user terminal 40 accesses the management server 20.
  • the reproduction management unit 214 assigns a group management code to the group.
  • the reproduction management unit 214 stores the user code associated with the group management code in the group information storage unit for the user terminal 40 logged in using the group management code (group registration).
  • group registration it is assumed that the user terminals 40a to 40c log in using the same group management code.
  • control unit 21 executes a process of accepting music designation (step S11-2). Specifically, the reproduction management unit 214 obtains the music code of the music to be used from any user terminal 40 registered as a group. For example, it is assumed that a music designation is received from the user terminal 40b.
  • control unit 21 executes a process for specifying the designated person as the parent device (step S11-3). Specifically, the reproduction management unit 214 registers the user terminal 40b as a parent device for the previously specified music in the user terminals 40a to 40c to which the group management code is assigned.
  • the control unit 21 executes a process for instructing synchronization (step S11-4). Specifically, the reproduction management unit 214 transmits a synchronization instruction to the user terminals 40a to 40c registered in the group. In this case, the reproduction management unit 214 acquires the music download status at each of the user terminals 40a to 40c registered in the group. Then, the reproduction management unit 214 instructs each user terminal 40 to wait until downloading of all the user terminals 40a to 40c is completed.
  • control unit 21 executes a reproduction process (step S11-5). Specifically, when the download management information is acquired in all the user terminals 40a to 40c registered in the group, the playback management unit 214 instructs to start playback of the music at the same time.
  • control unit 21 executes a process for storing the playback results (step S11-6). Specifically, the reproduction management unit 214 registers the user performance record 251 using the user code of the user terminal 40b designated as the parent device.
  • the playback management unit 214 may use the playback of music at the same place as a condition.
  • location information is acquired from each of the user terminals 40a to 40c.
  • the location information for example, position information by a GPS function provided in the user terminals 40a to 40c can be used. Accordingly, the control unit 21 can determine that the service is being used at the same place at the same time.
  • the lyrics synchronization status may be confirmed based on information from the user.
  • the management server 20 includes a correction information storage unit, and executes lyrics synchronization adjustment processing.
  • the correction information storage unit stores a request management record relating to a correction request acquired from the user.
  • the request management record stores data related to the user code, the reproduction time, the score, and the correspondence result for the music code.
  • the user code data area stores data relating to an identifier for specifying a user who has made a correction request.
  • the reproduction time data area stores data relating to a time index for specifying a portion desired to be corrected in the music.
  • the score data area stores data related to the score given to the correction request.
  • the response result data area stores a flag for specifying the response result for the correction request. Either the “correction” flag when the correction is performed or the “correction unnecessary” flag when the correction is not performed is stored.
  • the lyrics management unit 212 stores data related to the correction confirmation reference point for confirming whether correction is necessary.
  • the lyrics management unit 212 stores weighting information to be given to the score based on the user attribute.
  • the weighting information stores a weighting value to be given to the correction ratio corrected based on the correction request.
  • the control unit 21 executes a process of acquiring user correction information (step S12-1). Specifically, when it is determined that there is a deviation in the synchronization state of lyrics during reproduction on the user terminal 40, the service user inputs a correction request on the user terminal 40. In this case, the user terminal 40 transmits a correction request to the management server 20. This correction request includes data relating to the music code and the playback time at which the correction request was input. The lyrics management unit 212 receives a correction request from the user terminal 40.
  • the control unit 21 executes processing for storing correction information (step S12-2).
  • the lyrics management unit 212 generates a request management record for the acquired correction request, and stores it in the correction information storage unit.
  • the request management record includes data related to the music code, user code, and playback time.
  • the control unit 21 executes a process of assigning points weighted according to user attributes (step S12-3). Specifically, the lyrics management unit 212 extracts, from the correction information storage unit, the record in which the processed flag is stored in the request management record in which the user code is stored. Then, the lyrics management unit 212 calculates the correction ratio of the number of records in which the correction flag is stored with respect to the number of requests. And the lyrics management part 212 gives the score weighted according to the correction ratio with respect to this correction request, and memorize
  • the control unit 21 executes a process for determining whether or not there is the same correction as the correction request (step S12-4). Specifically, the lyrics management unit 212 extracts from the correction information storage unit a correction management record in which the same music code as the correction request is stored and no flag is stored in the corresponding result data area. Further, the lyrics management unit 212 extracts a record in which the same playback time zone as the correction request is stored from the extracted correction management records.
  • the control unit 21 executes the total processing of the points (step S12-5). Specifically, the lyrics management unit 212 sums up the points stored in the extracted correction management records, and calculates it as a total score.
  • control unit 21 skips the score totaling process (step S12-5).
  • control unit 21 executes processing for determining whether or not the total score exceeds the reference point (step S12-6). Specifically, the lyrics management unit 212 compares the calculated total score with the correction confirmation reference point. If the total score does not exceed the reference point (“NO” in step S12-6), the control unit 21 ends the lyrics synchronization adjustment process.
  • step S12-7 the control unit 21 executes a process for proposing correction (step S12-7).
  • the lyrics management unit 212 transmits a confirmation request to the person-in-charge terminal 30.
  • the confirmation request includes information related to the music code and the reproduction time stored in the request management record.
  • control unit 21 executes a correction determination process (step S12-8).
  • the person-in-charge terminal 30 that has acquired the confirmation request displays the content of the confirmation request.
  • the editor in charge reproduces the music and confirms the synchronization state of the lyrics.
  • the editor in charge determines that the lyrics are not synchronized properly, the editor in charge modifies the XML file.
  • the lyrics management unit 212 stores a corrected flag in all correction management records related to the same correction request.
  • the editor in charge determines that the synchronization of lyrics is appropriate, the editor in charge inputs correction unnecessary from the person in charge terminal 30.
  • the lyrics management unit 212 stores all correction unnecessary flags related to the same correction request in the correction management record.
  • the provided content may be background music in which the volume of the vocal is individually adjusted or video in which the volume of the vocal is individually adjusted.
  • the service management system according to the second embodiment is different from the first embodiment and the modification thereof in that the degree of individually changing the vocal volume is set. Therefore, below, especially a different point from 1st Embodiment and its modification is demonstrated, and the description of the structure which mutually overlaps is omitted.
  • the user terminal 40 includes a control unit 41A, a communication unit 42A, an input unit 43, a data storage unit 44, a volume detection unit 45, and a music output unit 46.
  • the control unit 41A includes a CPU, DSP, ROM, RAM, and the like. 41 A of control parts control the process which each part, such as the input part 43, the communication part 42A, the data storage part 44, the sound volume detection part 45, and the music output part 46, performs.
  • the input unit 43 is, for example, a remote controller or a volume switch that provides operation buttons.
  • the input unit 43 inputs various types of information to the control unit 41A in a format that can be processed by the control unit 41A.
  • the information input by the input unit 43 includes a designated mode, a designated mask degree, and a designated period.
  • the degree of masking indicates the degree to which vocals are individually masked during the period when the music is being played based on the music data.
  • the degree of masking is the degree to which the volume of the vocal is relatively lowered with respect to the volume of the music that is an example of the mixed sound.
  • the vocal volume when not masked in the vocal volume relative to the volume of the music is the master volume equal to the volume being played from the master file.
  • the volume of the vocal when masked is a support volume smaller than the master volume.
  • the designated mask degree is a mask degree input by the user of the karaoke service to the user terminal 40.
  • the designated mask degree can be input one by one with respect to the vocals included in the music before the music using the music data is reproduced. For example, when the music is a solo music, it is possible to input one designated mask degree for the vocal. When the music is a duet music, it is possible to input a designated mask degree for each of the two types of vocals.
  • the designation mode is a reproduction format that is input to the user terminal 40 by the user of the karaoke service.
  • the designated mode can be input one by one with respect to the reproduction of the music before the music using the music data is reproduced.
  • the designation mode is, for example, any one of a fixed mode, a first variation mode, and a second variation mode.
  • 41 A of control parts employ
  • the control unit 41A employs a variation value as the masking degree, and masks the vocal based on the variation value when reproducing the song from the song data.
  • the variation value in the first variation mode changes according to the detected sound volume of the performance sound that is the sound volume detected by the sound volume detector 45.
  • the variation value in the first variation mode changes, for example, as the detected volume is larger than the reference volume, so that the support volume is lower.
  • the support volume is It changes as big.
  • the reference volume may be a volume obtained by masking the master volume at a certain rate, or may be a certain volume.
  • control unit 41A adopts a variation value as the mask degree, and masks the vocal with the variation value when reproducing the song from the song data.
  • the variation value in the second variation mode changes so that the total amount of the detected volume and the support volume matches the master volume.
  • the input unit 43 inputs, to the control unit 41A, an application period that is a period in which the designated mode is applied during a period in which music is played, in a format that can be processed by the control unit 41A.
  • the application period is, for example, at least one of A melody, B melody, C melody, rust, and chapter that compose the music.
  • the control unit 41A controls transmission of data executed by the communication unit 42A and reception of data executed by the communication unit 42A.
  • the data storage unit 44 includes a storage area for storing various data input by the input unit 43 to the control unit 41A and various data such as a karaoke file received by the communication unit 42A.
  • the data storage unit 44 stores a music playback program executed by the control unit 41A to change the volume of the vocal and various data used by the control unit 41A to change the volume of the vocal.
  • the data stored in the data storage unit 44 includes a standard mask degree as a default mask degree.
  • the data stored in the data storage unit 44 includes a standard period as a default application period.
  • the standard mask degree is adopted in place of the designated mask degree in the control unit 41A when the designated mask degree is not inputted for the reproduction of the music during the reproduction of the music.
  • the standard period is adopted in place of the designated period in the control unit 41A when the designated period is not input for the reproduction of the music.
  • the data stored in the data storage unit 44 includes data related to an identifier for the control unit 41A to specify the designated mode.
  • the control unit 41 ⁇ / b> A controls data storage executed by the data storage unit 44 and reading of data stored in the data storage unit 44 from the data storage unit 44.
  • the control unit 41A reads the music reproduction program stored in the data storage unit 44, and generates a music reproduction signal for reproducing the music from the music data according to the read music reproduction program.
  • the control unit 41A inputs the generated music reproduction signal to the music output unit 46, and causes the music output unit 46 to perform reproduction based on the music reproduction signal.
  • the control unit 41A and the music output unit 46 function as a volume setting unit and a playback processing unit by executing a music playback program.
  • the volume detector 45 detects the volume of the performance sound input to the user terminal 40 as a detected volume.
  • the volume detection unit 45 detects, for example, the volume of a singing voice input to a microphone by a user who plays vocals as a detected volume.
  • the control unit 41A includes a volume setting unit 411 and a reproduction signal generation unit 412.
  • the volume setting unit 411 executes a process for changing the volume of the vocal using the music data and the vocal mask included in the karaoke file, the designated mask degree stored in the data storage unit 44, and the standard mask degree. .
  • the vocal mask is data for specifying a vocal line, which is a range including the vocal scale, within the range of the reproduced sound.
  • the vocal mask includes a vocal range as data, and the vocal range is defined with respect to a time index that is a reproduction time of a song.
  • the sound included in the vocal line is specified based on the vocal mask, and the volume of the specified sound is changed based on the mask degree, thereby changing the support volume.
  • the volume setting unit 411 sets the designated mask degree as the volume setting when the designated mask degree is input in the fixed mode or the first variation mode. On the other hand, the volume setting unit 411 sets the standard mask degree as the volume setting when the designated mask degree is not input.
  • the settings in the first variation mode and the second variation mode are auxiliary settings using a varying mask degree.
  • the volume setting unit 411 sets the specified period as the volume setting when the specified period is input in the fixed mode, the first variation mode, and the second variation mode. On the other hand, the volume setting unit 411 sets the standard period when the specified period is not input.
  • the volume setting unit 411 changes the volume of the vocal.
  • the processing for setting is executed as follows.
  • the volume setting unit 411 refers to music data, vocal mask, designated mask degree, and designated period.
  • the volume setting unit 411 sets the designated masking degree to the masking degree adopted by the reproduction signal generating unit 412 so that the volume of the vocal played from the music data becomes the support volume masked by the designated masking degree.
  • the sound volume setting unit 411 sets the designated period as a period during which the reproduction signal generation unit 412 adopts the designated mask degree.
  • the volume setting unit 411 sets the volume of the vocal.
  • the setting process for changing the value is executed as follows.
  • the volume setting unit 411 refers to music data, vocal mask, designated mask degree, and designated period.
  • the volume setting unit 411 sets the specified mask degree to the mask degree adopted by the reproduction signal generation unit 412 so that the vocal volume reproduced from the music data becomes the support volume when vocal reproduction is started from the music data. Set.
  • the volume setting unit 411 decreases the masking degree adopted by the reproduction signal generation unit 412 so that the support volume is increased when the volume detected by the volume detection unit 45 is decreased.
  • the volume setting unit 411 increases the masking degree adopted by the reproduction signal generation unit 412 so that the support volume is lowered when the volume detected by the volume detection unit 45 is increased.
  • the volume setting unit 411 sets a designated period as a period during which the reproduction signal generation unit 412 adopts these mask degrees.
  • the volume setting unit 411 executes the setting process for changing the volume of the vocal as follows. To do.
  • the volume setting unit 411 sequentially performs adjustment of the mask degree adopted by the reproduction signal generation unit 412 with reference to the music data, vocal mask, and specified period. At this time, the volume setting unit 411 adjusts the mask degree so that the total volume of the volume detected by the volume detection unit 45 and the support volume becomes the master volume. Further, the volume setting unit 411 sets a designated period as a period in which the reproduction signal generation unit 412 adopts such a mask degree.
  • the reproduction signal generation unit 412 generates a music reproduction signal that is a signal for reproducing music from the music data downloaded from the management server 20.
  • the reproduction signal generation unit 412 generates a video reproduction signal that is a signal for reproducing a video from the video data downloaded from the management server 20.
  • the reproduction signal generation unit 412 uses the music data and the vocal mask, and identifies the vocal from the music reproduced from the music data. In other words, the reproduction signal generation unit 412 generates a signal for reproducing the music from the music data, and specifies a signal corresponding to vocal among the generated signals. Further, the reproduction signal generation unit 412 employs the mask degree set by the volume setting unit 411. Then, the reproduction signal generation unit 412 processes a signal for reproducing the music so that the signal specified as vocal becomes a signal masked with the mask degree, and uses it as a music reproduction signal. Note that the reproduction signal generation unit 412 adopts the period set by the volume setting unit 411 as the period in which the masking degree is adopted, and executes the above processing only during the set period.
  • the volume setting unit 411 reads the music data included in the karaoke file provided from the management server 20 from the data storage unit 44 (step S13-1).
  • the volume setting unit 411 reads the vocal mask included in the karaoke file provided from the management server 20 from the data storage unit 44 (step S13-2).
  • the volume setting unit 411 reads the designation mode input from the input unit 43 during the current music reproduction from the data storage unit 44, and sets the read designation mode in the reproduction signal generation unit 412 (step S13-). 3).
  • the volume setting unit 411 sets the designated period in the reproduction signal generating unit 412 as the adoption period of the mask degree adopted for the current music reproduction.
  • the volume setting unit 411 sets the standard period in the reproduction signal generation unit 412 as the adoption period of the mask degree employed for the current music reproduction.
  • the volume setting unit 411 sets the mask level used for the current music playback to the playback signal generation unit 412 as the specified mask level.
  • the volume setting unit 411 sets the mask level employed for the current music playback as the standard mask level in the playback signal generation unit 412.
  • control unit 41A determines whether or not the designated mode employed for the current music reproduction is the fixed mode (step S13-4). When it is determined that the designated mode employed for the current music reproduction is the first variation mode or the second variation mode (in the case of “NO” in step S13-4), the control unit 41A performs the designation mode. The reproduction of music in a certain first variation mode or second variation mode is started (step S13-5).
  • step S13-4 when it is determined that the designated mode employed in the current music reproduction is the fixed mode (in the case of “YES” in step S13-4), the control unit 41A reproduces the music in the fixed mode which is the designated mode. Is started (step S13-15).
  • control unit 41A performs a sound volume detection unit within the designated period every time the reproduction time reaches the designated period in the reproduction of the music in the first fluctuation mode or the reproduction of the music in the second fluctuation mode. The following processing is repeated based on the volume detected by 45.
  • the volume setting unit 411 of the control unit 41A when playing the music in the first variation mode, every time the volume detection unit 45 detects the volume, the volume detected by the volume detection unit 45 falls below the reference volume.
  • the masking degree adopted by the reproduction signal generation unit 412 is lowered so as to increase the support volume.
  • the volume setting unit 411 increases the masking degree adopted by the reproduction signal generation unit 412 so that the support volume is lowered when the volume detected by the volume detection unit 45 is higher than the reference volume.
  • the volume setting unit 411 when playing the music in the second variation mode, every time the volume detection unit 45 detects the volume, the total volume of the volume detected by the volume detection unit 45 and the support volume becomes the master volume.
  • the mask degree is adjusted (step S13-6).
  • the reproduction signal generation unit 412 generates a music reproduction signal for generating a reproduction sound at a new vocal volume by using the mask degree reflecting the detection result of this time (step S13-7).
  • control unit 41A sets the standard mask degree so that the volume of the vocal is masked with the current mask degree every time the reproduction time reaches the specified period in the reproduction of the music in the fixed mode (step S13). -17). Then, the reproduction signal generation unit 412 generates a music reproduction signal for generating a reproduction sound at a new vocal volume using the mask degree set this time (step S13-18).
  • the reproduction signal generation unit 412 generates a music reproduction signal for generating a reproduction sound at a new vocal volume using the mask degree set this time (step S13-18).
  • the solid line indicates the transition of the detected volume, which is the volume detected by the volume detector 45, with respect to the playback time.
  • a broken line indicates transition of the support volume, which is a vocal in the music reproduced by the user terminal 40, with respect to the reproduction time.
  • the alternate long and two short dashes line shows the transition of the master volume, which is the volume of the unmasked vocal, with respect to the playback time.
  • FIG. 28 shows the transition of the volume of each sound in the fixed mode.
  • FIG. 29 shows the transition of the volume of each sound in the first variation mode.
  • FIG. 30 shows the transition of the volume of each sound in the second variation mode.
  • the detected volume changes so as to follow the change in the master volume.
  • the playback time is the period T, for example, the user can play vocals as the user's voice range deviates from the vocal range required for vocals or the user's voice dies. Accordingly, the detected sound volume is lowered to a level where it is hardly detected.
  • the support volume masked at a certain masking degree is smaller than the master volume, and changes throughout the playback time so as to follow the change.
  • the music is composed of the support sound played by the user terminal 40 and other sounds. Therefore, playing the music is assisted by the user terminal 40.
  • the detected sound volume also changes so as to follow the change in the master disk sound volume as the reproduction time elapses.
  • the reproduction time is the period T
  • the detected sound volume is reduced to a level where it is hardly detected.
  • the support sound volume due to the varying masking degree changes following the change of the difference dV1 between the detected sound volume and the master disk sound volume with reference to the support sound volume due to the fixed masking degree.
  • the fixed mask degree is the specified mask degree.
  • the fixed mask degree is the standard mask degree.
  • the support volume generated with the varying mask degree is higher as the volume of the singing voice is lower than the master volume. Further, the support volume generated with the varying mask degree is lower as the volume of the singing voice is higher than the master volume.
  • control unit 41A changes the volume obtained by multiplying the difference dV1 between the detected volume and the master volume by a predetermined value ⁇ of 1 or less as a reference volume by masking the master volume with a fixed degree of masking. Minutes, and the total level of these is the support volume.
  • the support volume obtained in this way may be higher or lower than the volume obtained by masking the master volume with a fixed degree of masking.
  • the predetermined value ⁇ may be configured to be input by the input unit 43 together with the specified mode, the specified mask degree, the specified period, etc. It may be a configuration.
  • the music is composed of the support sound generated by the user terminal 40 and other sounds, and consequently the music is played. It becomes possible to assist.
  • the support sound is generated with a volume sufficiently lower than the support volume generated with a fixed mask degree. Therefore, it is possible to prevent the support sound from being played by the user terminal 40 from interfering with the vocal performance by the user.
  • the detected sound volume also changes so as to follow the change in the master disk sound volume as the reproduction time elapses.
  • the reproduction time is the period T
  • the detected sound volume is reduced to a level where it is hardly detected.
  • the support volume based on the varying masking degree is reproduced when the detected volume is smaller than the master volume, and the level of the volume is set to a value corresponding to the difference dV2 between the detected volume and the master volume. It changes following the change of the difference dV2 between the volume and the master volume.
  • the support volume due to the changing mask level increases in the period T to the same level as the master volume.
  • such a support volume is sufficiently suppressed during a period in which the user can sufficiently play vocals.
  • the music is composed of the support sound generated by the user terminal 40 and other sounds, and consequently the music is played. Can help. In addition, it is further suppressed that the vocal reproduction by the user terminal 40 hinders the vocal performance by the user.
  • the first variation mode and the second variation mode for example, in a duet music including female vocals and male vocals, both vocal masks are used, and the detected sound volume of either one is hardly detected. Even if it is a level, the part can be assisted by the support volume.
  • the following effects can be obtained. (1) Even when the user cannot play vocals, only the accompaniment sound flows in the space can be suppressed by generating the support sound. In addition, since the target sound such as vocals included in the master music is adopted as the support sound, it is possible to prevent a common person who listens to the playback sound together with the user from feeling uncomfortable with the support sound.
  • the third embodiment differs from the music playback device in the second embodiment in that the volume setting up to the previous time is handled as a history.
  • points different from the configurations described in the second embodiment will be mainly described, and components having the same functions as those described in the second embodiment will be denoted by the same reference numerals and description thereof will be omitted. To do.
  • the control unit 41A performs transmission of setting candidates by the communication unit 42A each time reproduction of a song using the song data is executed.
  • the management unit 210 of the management server 20 functions as a history management unit, and executes a process of storing the setting candidates received from the user terminal 40 in the reproduction result storage unit 25.
  • the playback record storage unit 25 includes the playback date and time, the user code, the music code, and the setting candidate in the playback record data 250.
  • the setting candidates included in the playback performance data 250 include data relating to an identifier for specifying the designated mode in the current playback. Further, the setting candidates included in the reproduction performance data 250 include the mask degree used for reproducing the support sound volume, such as the designated mask degree and the standard mask degree. In addition, the setting candidates included in the playback record data 250 include an application period used for playback of the support volume, such as a specified period and a standard period.
  • control unit 21 of the management server 20 When the control unit 21 of the management server 20 provides a karaoke file to the user terminal 40, the control unit 21 refers to the reproduction result data 250 stored in the reproduction result storage unit 25. The control unit 21 determines whether or not there is a setting candidate corresponding to the user code related to the current reproduction and the music code.
  • the control unit 21 of the management server 20 sets all the setting candidates corresponding to the user code and the music code related to the current reproduction as karaoke. It is included in the file and transmitted to the user terminal 40.
  • the control unit 21 of the management server 20 transmits to the user terminal 40 that the karaoke file includes no setting candidate.
  • the user terminal 40 includes a history analysis unit 413 for analyzing the setting candidates received from the management server 20.
  • the history analysis unit 413 handles each setting candidate received from the management server 20 as a setting history by the user, and generates a volume setting for the current reproduction from the setting history by the user according to a predetermined algorithm.
  • the history analysis unit 413 extracts the setting that has been adopted most frequently from a plurality of setting candidates, and sets the extracted setting as the current setting. For example, the history analysis unit 413 extracts the designated mode that has been adopted most frequently from the setting candidates, and sets the extracted designated mode as the current designated mode.
  • the history analysis unit 413 determines whether or not there is a continuous designation mode in order from the setting candidate with the latest reproduction date and time, and sets the latest continuous designation mode as the current designation mode. Further, the history analysis unit 413 uses, for example, the mask degrees included in the setting candidates, calculates the average value or mode value thereof as a representative value, and sets the calculated representative value as the current designated mask degree.
  • the control unit 41A reads music data and a vocal mask from the provided karaoke file (step S13-1). At this time, the control unit 41A determines whether or not setting candidates are included in the karaoke file (step S13-2).
  • the volume setting unit 411 sets the fixed mode as the designated mode adopted by the reproduction signal generating unit 412. Further, the volume setting unit 411 sets the standard mask degree as the mask degree adopted by the reproduction signal generation unit 412 and sets the standard period as the application period adopted by the reproduction signal generation unit 412 (step S13-3).
  • the history analysis unit 413 of the user terminal 40 uses the designated mode, mask degree, and The application period is generated from the read setting candidates (step S14-3).
  • the volume setting unit 411 temporarily sets the designation mode, the masking degree, and the designation period adopted by the reproduction signal generation unit 412 according to the analysis result by the history analysis unit 413. Further, the volume setting unit 411 prompts the user to confirm the setting in the current reproduction by outputting the temporarily set content to the display unit or the like (step S14-4).
  • the volume setting unit 411 permits the change of the designated mode, the degree of masking, and the designated period by an external operation on the input unit 43 of the user terminal 40. Then, the volume setting unit 411 sets the confirmed contents in the designation mode, the masking degree, and the application period adopted by the reproduction signal generation unit 412 (step S13-3).
  • control unit 41A reproduces the music from the music data included in the karaoke file according to the volume setting by the volume setting unit 411. Then, the control unit 41A transmits the volume setting for the current reproduction to the management server 20 as the next setting candidate (step S14-5).
  • the volume required for the current playback as the volume of the support sound and the application period thereof are the same volume and the same application period as any of the settings made up to the previous playback if the music to be played is the same.
  • the volume setting unit 411 sets the current volume using the volume setting performed up to the previous playback, the load on the user who inputs a new setting is reduced.
  • the user terminal 40 is not limited to a terminal connected to the management server 20 via a network, and may be a device that is not connected to the network, for example.
  • the user terminal 40 may include a vocal mask generation function included in the management server 20. That is, the user terminal includes a functional unit corresponding to the vocal processing unit 211 included in the management server 20, and specifies a vocal range from the music data included in the master file and generates a vocal mask for changing the vocal volume. May be executed.
  • the reproduction of the music performed by the user terminal 40 is not limited to the process using the music data stored in the music storage unit 24 or the vocal mask.
  • the user terminal 40 acquires, for example, a streaming file for music for reproducing the mixed sound and a streaming file for masking for masking the target sound in the mixed sound from the management server 20, and acquires these.
  • the music may be streamed using the file.
  • the user terminal 40 in the first embodiment may also have a similar function for performing such streaming playback.
  • the adjustment target sound may be, for example, a sound that is played by two or more musical instruments that constitute an orchestra and are different from each other.
  • the target sound may be included separately in two or more sound ranges associated with the common time index.
  • the user terminal 40 may use a plurality of vocal masks for masking different sound ranges, and a masking degree for each vocal mask.
  • the user terminal 40 may use one vocal mask defined so as to mask different sound ranges and one mask degree for each sound range. In this way, if two or more sound ranges include different adjustment target sounds, they are generated from a plurality of sounds that are played outside each having a different scale and other reproduced sounds. It is possible to assist in playing the mixed sound.
  • a plurality of vocal masks may be adopted when playing one piece of music.
  • each vocal mask and the application period in which it is adopted are associated one by one with the vocal mask in the user terminal 40 based on an operation input to the input unit 43 or the like.
  • a dynamic vocal mask such as the third vocal mask is used in an application period in which high accuracy of the mask by the vocal mask is required.
  • a static vocal mask such as the first vocal mask is used in an application period in which the accuracy of the mask by the vocal mask is not required.
  • the user terminal 40 may further include a data processing unit that generates a vocal mask from the music data stored in the data storage unit 44 according to a predetermined algorithm and causes the data storage unit 44 to store the generated vocal mask.
  • the data processing unit includes, for example, a CPU, a ROM, and a RAM, stores an algorithm for generating a vocal mask, and generates a vocal mask from music data input from the input unit 43 according to a predetermined algorithm.
  • the reproduction processing unit receives the processing result of the data processing unit, reads out the music data and the vocal mask from the data storage unit 44, and reproduces the music from the music data using the vocal mask and the mask degree.
  • the application period is not limited to A melody, B melody, rust, and chapter, for example, a period that is higher than a predetermined scale in a vocal range or a lower range than a predetermined scale in a vocal range It may be determined in advance as a period that is a sound range. In setting such a period, first, a vocal range for determining the application period is determined, and a time index in which the vocal scale is included in the range is set as the application period.
  • the human voice range is divided into, for example, a high range and a low range.
  • a user who does not have a voice range in the high range requires a support sound in the high range
  • a user who does not have a voice range in the low range has a low range.
  • Support sound is required in the range.
  • the application period is determined by the vocal range
  • the support sound is generated only during the application period in which the support sound is originally required. Therefore, it is possible to reduce the load.
  • the auxiliary setting is performed in the first variation mode and the second variation mode in such an application period, assistance can be performed at a volume that is preferable as a support volume even in an application period in which the support sound is originally required. It is also possible.
  • the user terminal 40 may further include a sound range specifying unit 47 that specifies the sound range of the sound played by the performer from the sound input by the performer to the microphone.
  • the processing for specifying the sound range performed by the sound range specifying unit 47 may be performed before the music is reproduced, or may be sequentially performed during the reproduction of the music.
  • the range specifying unit 47 specifies the range of the sound played by the performer in advance
  • the volume setting unit 411 includes the mode specifying unit 414 for performing assistance based on the designated mode in a range other than the range specified in advance. You may prepare.
  • the mode specifying unit 414 specifies, for example, a time index associated with a sound range other than the sound range specified by the sound range specifying unit 47 in the time index associated with the adjustment target sound based on the music data. And the mode specific
  • the sound range specifying unit 47 sequentially specifies the sound range of the sound played by the performer during performance
  • the volume setting unit 411 is a mode specifying unit for performing assistance based on the designated mode based on the sequentially specified sound range.
  • 414 may be provided.
  • the mode specifying unit 414 determines whether or not the difference between the player's range specified by the range specifying unit 47 and the range of the target sound based on the song file is within a predetermined range. Repeat during playback. When the difference between the player's range specified by the range specification unit 47 and the range of the target sound based on the music file is outside the predetermined range, the mode specification unit 414 determines the mask degree based on the designated mode at that time. Set.
  • the predetermined range which is an allowable range of the difference between the performer's range specified by the range specification unit 47 and the range of the adjustment target sound based on the music file, is the same as the specified mask degree and the like from the input unit 43.
  • the structure which is input may be sufficient and the structure which the data storage part 44 memorize
  • the range specifying unit 47 specifies the range of the sound played by the performer in advance, and the control unit 41A transmits the specified performer's range to the control unit 21 via the communication of the communication unit 42A.
  • the control unit 21 refers to the data related to the sound range in each piece of music management data 240, and specifies the music management data 240 in which the sound range indicated by the data related to the sound range includes the player's sound range.
  • the control unit 21 transmits the specified music management data 240 to the communication unit 42A as music management data 240 of the recommended music.
  • the control unit 41A may cause the display unit to output a list of recommended songs based on the song management data 240 of the recommended songs received by the communication unit 42A. According to the recommendation of such music, it is possible to recommend to the performer a music suitable for the sound range played by the performer.
  • control unit 41A may transmit to the control unit 21 information related to a scene requested by the performer via communication of the communication unit 42A in addition to the range of the sound played by the performer.
  • control part 21 specifies the music management data 240 in which the scene contained in the scene which a player requires is referred with reference to the data regarding a scene in each music management data 240.
  • the control unit 21 transmits the specified music management data 240 to the communication unit 42A as music management data 240 of the recommended music. According to the recommendation of such music, it is possible to recommend to the performer a music that is suitable for the sound range played by the performer and suitable for the scene that the performer desires.
  • the playback determination method executed by the playback determination process is not limited to the method based on the playback determination time.
  • the reproduction determination may be performed based on the position of rust in the music.
  • the control unit 21 of the management server 20 specifies the chorus position based on the specific pattern in the main melody estimation process (step S1-3).
  • the specified chorus position is stored in the music management data 240.
  • the playback management unit 214 performs playback determination based on the time index and the chorus position.
  • the playback determination time has not elapsed, if the rust is played back, it is determined to be played back.
  • the billing method may be changed according to the set mask degree.
  • a rule is stored in the distribution management file 270 to calculate the distribution money based on the volume of vocals output according to the masking degree of the mask file.
  • Music management data 241 ... HLS file, 242 ... XML file, 243 ... Vocal mask, 244 ... music management record, 250 ... reproduction performance data, 251 ... user performance record, 252 ... music performance record, 26 ... stock memory 411: Volume setting unit, 412: Reproduction signal generation unit, 42A: Communication unit, 500, 510, 520, 600, 601, 602, 603 ... Scale transition graph, 501: First vocal mask, 511: Second vocal mask 521 ... Third vocal mask, 700 ... XML file.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Business, Economics & Management (AREA)
  • Acoustics & Sound (AREA)
  • Tourism & Hospitality (AREA)
  • General Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Resources & Organizations (AREA)
  • General Business, Economics & Management (AREA)
  • General Physics & Mathematics (AREA)
  • Strategic Management (AREA)
  • Theoretical Computer Science (AREA)
  • Primary Health Care (AREA)
  • Marketing (AREA)
  • Economics (AREA)
  • Computational Linguistics (AREA)
  • Quality & Reliability (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Reverberation, Karaoke And Other Acoustics (AREA)

Abstract

This service management system is provided with a control unit that: specifies the range of sounds to be modified in a master file stored in a master storage unit; creates a mask file for adjusting the volume of the sounds individually within a mix of sounds on the basis of the specified range; provides a user terminal with sound modified content including the mask file and stores a number of playbacks into a playback result storage unit; and, on the basis of funds stored in a fund storage unit and the number of playbacks recorded in the playback result storage unit, calculates payments to be distributed to eligible parties for the modified sound.

Description

サービス管理システム、コンピュータ可読記憶媒体、および、サービス管理方法Service management system, computer-readable storage medium, and service management method
 本発明は、音楽を提供するサービスを管理するサービス管理システム、コンピュータ可読記憶媒体、および、サービス管理方法に関する。 The present invention relates to a service management system, a computer-readable storage medium, and a service management method for managing a service that provides music.
 音楽を提供するサービスの一例として、伴奏音を再生するカラオケサービスが知られている(例えば、特許文献1を参照)。カラオケサービスに用いられる音源は、伴奏音とボーカルとを含む混合音のデータを有した音楽原盤ではなく、カラオケ用に新たに作り出された伴奏音のみからなるデータである。 As an example of a service that provides music, a karaoke service that plays accompaniment sounds is known (see, for example, Patent Document 1). The sound source used in the karaoke service is not a music master having mixed sound data including accompaniment sounds and vocals, but data consisting only of accompaniment sounds newly created for karaoke.
特開2002-99290号公報JP 2002-99290 A
 一方、カラオケ用に新たに作り出された音と、本来の音楽原盤に基づく音とは相互に異なるため、音楽原盤に基づく音を期待するユーザを上述した技術によって満足させることは難しい。また、音楽原盤を製作した演奏者などには、コンテンツの提供サービスによる収益金が分配されないため、コンテンツの提供サービスによる音楽の普及も困難である。 On the other hand, since the sound newly created for karaoke and the sound based on the original music master are different from each other, it is difficult to satisfy the user who expects the sound based on the music master using the above-described technology. Further, since the profits from the content providing service are not distributed to the performers who have produced the original music, it is difficult to spread the music through the content providing service.
 なお、上述した課題は、伴奏音を再生するカラオケサービスに限らず、コンサートの公演や演劇の公演などのように、楽曲データから混合音を再生する技術において共通する。 Note that the above-mentioned problems are not limited to karaoke services for reproducing accompaniment sounds, but are common to technologies for reproducing mixed sounds from music data such as concert performances and theater performances.
 上記課題を解決するためのサービス管理システムは、音量調整の対象となる調整対象音とそれ以外の音とを含む音が混合音であり、前記混合音を再生するためのデータを含む原盤ファイルを記憶した原盤記憶部と、音調整コンテンツをユーザ端末に提供するサービスの原資を記憶した原資記憶部と、前記ユーザ端末による音調整コンテンツの再生回数を記憶する再生実績記憶部と、前記ユーザ端末に接続される制御部とを備える。そして、前記制御部が、前記原盤記憶部に記憶された前記原盤ファイルにおいて前記調整対象音の音域を特定し、前記特定された音域に基づき、前記混合音のなかで前記調整対象音の音量を個別に調整するためのマスクファイルを生成し、前記マスクファイルを含む前記音調整コンテンツを前記ユーザ端末に提供し、前記再生回数を前記再生実績記憶部に記憶し、前記原資記憶部に記憶した原資と、前記再生実績記憶部に記憶された前記再生回数とに基づいて、前記調整対象音の分配対象者に分配する分配金額を算出する。 In the service management system for solving the above-described problem, a sound including a sound to be adjusted and a sound other than the sound to be adjusted is a mixed sound, and a master file including data for reproducing the mixed sound is obtained. A stored master recording unit, a source storage unit storing a source of service for providing sound adjustment content to a user terminal, a playback record storage unit storing the number of times of sound adjustment content played by the user terminal, and the user terminal And a connected control unit. Then, the control unit specifies a range of the adjustment target sound in the master file stored in the master storage unit, and based on the specified range, adjusts the volume of the adjustment target sound in the mixed sound. A mask file for individual adjustment is generated, the sound adjustment content including the mask file is provided to the user terminal, the number of reproductions is stored in the reproduction result storage unit, and the resource stored in the resource storage unit And a distribution amount to be distributed to the distribution target person of the adjustment target sound, based on the number of times of reproduction stored in the reproduction result storage unit.
 上記課題を解決するためのコンピュータ可読記憶媒体は、音量調整の対象となる調整対象音とそれ以外の音とを含む音が混合音であり、前記混合音を再生するためのデータを含む原盤ファイルを記憶した原盤記憶部と、音調整コンテンツをユーザ端末に提供するサービスの原資を記憶した原資記憶部と、前記ユーザ端末による音調整コンテンツの再生回数を記憶する再生実績記憶部と、前記ユーザ端末に接続される制御部とを備えたサービス管理システムを用いて、サービス管理を行なうプログラムが記憶されたコンピュータ可読記憶媒体である。そして、前記プログラムの実行時、前記制御部は、前記原盤記憶部に記憶された前記原盤ファイルにおいて前記調整対象音の音域を特定し、前記特定された音域に基づき、前記混合音のなかで前記調整対象音の音量を個別に調整するためのマスクファイルを生成し、前記マスクファイルを含む前記音調整コンテンツを前記ユーザ端末に提供し、前記再生回数を前記再生実績記憶部に記憶し、前記原資記憶部に記憶した原資と、前記再生実績記憶部に記憶された前記再生回数とに基づいて、前記調整対象音の分配対象者に分配する分配金額を算出するように機能する。 A computer-readable storage medium for solving the above-described problem is a master file that includes a sound that includes a sound to be adjusted and a sound other than the sound to be adjusted as a mixed sound, and includes data for reproducing the mixed sound A master disk storage unit, a resource storage unit that stores resources for providing sound adjustment content to a user terminal, a playback record storage unit that stores the number of times the sound adjustment content is played by the user terminal, and the user terminal A computer-readable storage medium storing a program for performing service management using a service management system including a control unit connected to the computer. When executing the program, the control unit specifies a range of the adjustment target sound in the master file stored in the master storage unit, and based on the specified range, the control unit includes the A mask file for individually adjusting the volume of the adjustment target sound is generated, the sound adjustment content including the mask file is provided to the user terminal, the number of times of reproduction is stored in the reproduction result storage unit, and the resource Based on the resources stored in the storage unit and the number of reproductions stored in the reproduction record storage unit, the distribution amount to be distributed to the distribution target person of the adjustment target sound is calculated.
 上記課題を解決するためのサービス管理方法は、音量調整の対象となる調整対象音とそれ以外の音とを含む音が混合音であり、前記混合音を再生するためのデータを含む原盤ファイルを記憶した原盤記憶部と、音調整コンテンツをユーザ端末に提供するサービスの原資を記憶した原資記憶部と、前記ユーザ端末による音調整コンテンツの再生回数を記憶する再生実績記憶部と、前記ユーザ端末に接続される制御部とを備えたサービス管理システムを用いて、サービス管理を行なう方法である。そして、サービス管理方法は、前記制御部が、前記原盤記憶部に記憶された前記原盤ファイルにおいて前記調整対象音の音域を特定し、前記特定された音域に基づき、前記混合音のなかで前記調整対象音の音量を個別に調整するためのマスクファイルを生成し、前記マスクファイルを含む前記音調整コンテンツを前記ユーザ端末に提供し、前記再生回数を前記再生実績記憶部に記憶し、前記原資記憶部に記憶された原資と、前記再生実績記憶部に記憶された前記再生回数とに基づいて、前記調整対象音の分配対象者に分配する分配金額を算出する。 In the service management method for solving the above-described problem, a sound including a sound to be adjusted and a sound other than the sound to be adjusted is a mixed sound, and a master file including data for reproducing the mixed sound is obtained. A stored master recording unit, a source storage unit storing a source of service for providing sound adjustment content to a user terminal, a playback record storage unit storing the number of times of sound adjustment content played by the user terminal, and the user terminal In this method, service management is performed using a service management system including a connected control unit. In the service management method, the control unit specifies a range of the adjustment target sound in the master file stored in the master storage unit, and the adjustment is performed in the mixed sound based on the specified range. A mask file for individually adjusting the volume of the target sound is generated, the sound adjustment content including the mask file is provided to the user terminal, the number of reproductions is stored in the reproduction result storage unit, and the resource storage A distribution amount to be distributed to the distribution target person of the adjustment target sound is calculated based on the resources stored in the section and the number of times of reproduction stored in the reproduction performance storage section.
サービス管理システムの第1実施形態を示す構成図。The lineblock diagram showing a 1st embodiment of a service management system. 原盤記憶部が記憶するデータの構成図。The block diagram of the data which an original disk memory | storage part memorize | stores. 歌詞情報記憶部が記憶するデータの構成図。The lineblock diagram of the data which a lyric information storage part memorizes. 楽曲記憶部が記憶するデータの構成図。The block diagram of the data which a music memory | storage part memorize | stores. 再生実績記憶部が記憶するデータの構成図。The block diagram of the data which a reproduction | regeneration performance memory | storage part memorize | stores. 原資記憶部が記憶するデータの構成図。The block diagram of the data which a resource storage part memorize | stores. 分配管理情報記憶部が記憶するデータの構成図。The block diagram of the data which a distribution management information storage part memorize | stores. 分配実績情報記憶部が記憶するデータの構成図。The block diagram of the data which a distribution performance information storage part memorize | stores. サービス管理方法での処理の手順を示す図。The figure which shows the procedure of the process by a service management method. サービス管理方法での処理の手順を示す図。The figure which shows the procedure of the process by a service management method. 第1のマスク生成処理の手順を示す図。The figure which shows the procedure of a 1st mask production | generation process. マスクファイルの音域を示すグラフ。The graph which shows the range of a mask file. 第2のマスク生成処理の手順を示す図。The figure which shows the procedure of a 2nd mask production | generation process. マスクファイルの音域を示すグラフ。The graph which shows the range of a mask file. 第3のマスク生成処理の手順を示す図。The figure which shows the procedure of a 3rd mask production | generation process. マスクファイルの音域を示すグラフ。The graph which shows the range of a mask file. マスクファイルの音域を示すグラフ。The graph which shows the range of a mask file. 同期歌詞処理の手順を示すフロー図。The flowchart which shows the procedure of a synchronous lyrics process. XMLファイルの構成を示す構成図。The block diagram which shows the structure of an XML file. 再生判定処理の手順を示す図。The figure which shows the procedure of a reproduction | regeneration determination process. 分配管理処理の手順を示す図。The figure which shows the procedure of a distribution management process. 定額楽曲の分配処理における手順を示す図。The figure which shows the procedure in the distribution process of fixed amount music. グループの再生による再生回数の算出を説明する図。The figure explaining calculation of the frequency | count of reproduction | regeneration by reproduction | regeneration of a group. グループ再生管理における処理の手順を示す図。The figure which shows the procedure of the process in group reproduction | regeneration management. 歌詞同期調整における処理の手順を示す図。The figure which shows the procedure of the process in a lyrics synchronous adjustment. サービス管理システムの第2実施形態を示す構成図。The block diagram which shows 2nd Embodiment of a service management system. サービス管理方法での処理の手順を示す図。The figure which shows the procedure of the process by a service management method. 固定モードでの音域を示すグラフ。The graph which shows the range in fixed mode. 第1変動モードでの音域を示すグラフ。The graph which shows the sound range in a 1st fluctuation mode. 第2変動モードでの音域を示すグラフ。The graph which shows the sound range in 2nd fluctuation mode. サービス管理システムの第3実施形態を示す構成図。The block diagram which shows 3rd Embodiment of a service management system. サービス管理方法での処理の手順を示す図。The figure which shows the procedure of the process by a service management method. ユーザ端末の変形例における構成を示す構成図。The block diagram which shows the structure in the modification of a user terminal.
 (第1実施形態)
 以下、サービス管理システムの第1実施形態を説明する。サービス管理システムは、コンテンツの提供サービスの一例である原盤カラオケサービスを提供する。サービス管理システムは、レコード会社などが提供する音楽原盤を用い、ボーカルの音量がそれ以外の音の音量に対して個別に調整された混合音を提供する。ボーカルは、演奏者の一例である歌手が歌唱した歌声であり、調整対象音の一例である。
(First embodiment)
Hereinafter, a first embodiment of the service management system will be described. The service management system provides a master karaoke service that is an example of a content providing service. The service management system uses a music master provided by a record company or the like, and provides a mixed sound in which the volume of vocals is individually adjusted with respect to the volume of other sounds. Vocal is a singing voice sung by a singer who is an example of a performer, and is an example of a sound to be adjusted.
 音楽原盤は、ボーカルとそれ以外の音とを含む混合音を再生するためのデータを含む。音楽原盤は、音楽原盤ファイルであってもよいし、プロモーションビデオなどのように、音楽原盤ファイルと動画原盤ファイルとを含むボックス形式のファイルであってもよい。 The music master disc contains data for playing mixed sounds including vocals and other sounds. The music master may be a music master file, or may be a box format file including a music master file and a video master file, such as a promotion video.
 図1に示すように、コンテンツサーバ10は、コンピュータシステムであって、原盤ファイルを提供する。原盤ファイルは、混合音を再生するためのデータを提供するためのコンテンツである。コンテンツは、例えば、mp4フォーマットで作成されたプロモーションビデオなどのデジタルファイルである。 As shown in FIG. 1, the content server 10 is a computer system and provides a master file. The master disk file is content for providing data for reproducing the mixed sound. The content is, for example, a digital file such as a promotion video created in the mp4 format.
 担当者端末30は、コンピュータ端末であり、楽曲を再生するためのデータの編集に用いられる。楽曲の編集は、歌詞などのテロップが表示されるタイミングの調整や、ボーカルマスクの調整である。担当者端末30は、制御部、出力部、入力部を備える。出力部は、例えば、ディスプレイによって構成され、各種情報の出力を行なう。入力部は、キーボードやポインティングデバイスによって構成され、担当者の入力した指示を取得する。 The person-in-charge terminal 30 is a computer terminal and is used for editing data for reproducing music. The editing of music is adjustment of the timing at which telops such as lyrics are displayed and vocal mask adjustment. The person-in-charge terminal 30 includes a control unit, an output unit, and an input unit. The output unit is configured by a display, for example, and outputs various information. The input unit is configured by a keyboard and a pointing device, and acquires an instruction input by a person in charge.
 ユーザ端末40は、コンピュータ端末であって、カラオケサービスを利用するユーザに使用される。ユーザ端末40は、ビジネス向け(BtoB)の機器、あるいは、カスタマ向け(BtoC)の機器である。ビジネス向けの機器は、カラオケボックスなどに設置されるコンピュータ端末である。カスタマ向けの機器は、個人ユーザの所持するスマートフォンやパソコンである。 The user terminal 40 is a computer terminal and is used by a user who uses a karaoke service. The user terminal 40 is a business (BtoB) device or a customer (BtoC) device. Business equipment is a computer terminal installed in a karaoke box or the like. Devices for customers are smartphones and personal computers possessed by individual users.
 ユーザ端末40は、管理サーバ20から、楽曲ファイルや映像ファイルを含むカラオケファイルを、調整対象音の音量を調整するためのコンテンツである音調整コンテンツの一例として、ダウンロードする。カラオケファイルは、マスクファイルの一例としてボーカルマスクファイルを含む。 The user terminal 40 downloads a karaoke file including a music file and a video file from the management server 20 as an example of sound adjustment content that is content for adjusting the volume of the adjustment target sound. The karaoke file includes a vocal mask file as an example of a mask file.
 ユーザ端末40は、楽曲再生部41、および、ボーカル音量調整部42を備える。楽曲再生部41は、ダウンロードされた楽曲ファイルや映像ファイルを用い、楽曲や映像の再生を行なう。ボーカル音量調整部42は、ボーカルマスクファイルに基づいて、混合音のなかのボーカルの音量を個別に調整する処理を実行する。ボーカルの音量を調整する処理は、ボーカルを消去する処理を含む。ボーカルマスクファイルは、楽曲に含まれるボーカルの音域を特定するためのファイルである。 The user terminal 40 includes a music playback unit 41 and a vocal volume adjustment unit 42. The music playback unit 41 plays back music and video using the downloaded music file and video file. The vocal volume adjustment unit 42 performs a process of individually adjusting the volume of the vocal in the mixed sound based on the vocal mask file. The process of adjusting the volume of the vocal includes a process of deleting the vocal. The vocal mask file is a file for specifying the vocal range included in the music.
 管理サーバ20は、コンピュータシステムであり、カラオケサービス提供者によるカラオケファイルの提供に用いられる。管理サーバ20は、制御部21、原盤記憶部22、歌詞情報記憶部23、楽曲記憶部24、再生実績記憶部25、原資記憶部26、分配管理情報記憶部27、および、分配実績情報記憶部28を備える。 The management server 20 is a computer system and is used for providing a karaoke file by a karaoke service provider. The management server 20 includes a control unit 21, a master disk storage unit 22, a lyrics information storage unit 23, a music storage unit 24, a reproduction performance storage unit 25, a resource storage unit 26, a distribution management information storage unit 27, and a distribution result information storage unit. 28.
 制御部21は、CPU、RAM、ROMなどから構成される制御手段を備え、管理段階、ボーカル処理段階、歌詞管理段階、エディタ段階、再生管理段階、および、分配管理段階の各処理を行なう。制御部21は、サービス管理プログラムを実行することによって、管理部210、ボーカル処理部211、歌詞管理部212、エディタ213、再生管理部214、および、分配管理部215として機能する。
 管理部210は、コンテンツサーバ10から取得した原盤ファイルを原盤記憶部22に格納する処理を実行する。
 ボーカル処理部211は、ボーカルの音域を特定し、ボーカルの音量を調整するためのボーカルマスクファイルの生成処理を実行する。
 歌詞管理部212は、歌詞が表示されるタイミングと、ボーカルが再生されるタイミングとが合致するように、ファイルに歌詞を付加する処理を実行する。
The control unit 21 includes control means including a CPU, a RAM, a ROM, and the like, and performs each process of a management stage, vocal processing stage, lyrics management stage, editor stage, reproduction management stage, and distribution management stage. The control unit 21 functions as a management unit 210, a vocal processing unit 211, a lyrics management unit 212, an editor 213, a reproduction management unit 214, and a distribution management unit 215 by executing a service management program.
The management unit 210 executes processing for storing the master file acquired from the content server 10 in the master storage unit 22.
The vocal processing unit 211 executes a vocal mask file generation process for identifying the vocal range and adjusting the vocal volume.
The lyrics management unit 212 executes a process of adding lyrics to the file so that the timing at which the lyrics are displayed matches the timing at which the vocals are reproduced.
 エディタ213は、ボーカルマスクを調整する処理や、歌詞の付加位置の調整を支援する処理を実行する。エディタ213は、編集担当者が使用する担当者端末30に接続される。編集担当者は、担当者端末30を用いて、ボーカルマスクの設定や歌詞の付加位置を確認する。また、編集担当者は、ボーカルマスクの設定位置や歌詞の付加位置が適切でない場合には、担当者端末30を用いて、ボーカルマスクの設定位置や歌詞の付加位置を調整する。ボーカルマスクの設定位置は、ボーカルマスクがマスクする音域や、ボーカルマスクの設定される再生時間である。 The editor 213 executes processing for adjusting the vocal mask and processing for supporting adjustment of the addition position of the lyrics. The editor 213 is connected to the person-in-charge terminal 30 used by the person in charge of editing. The person in charge of the editor uses the person-in-charge terminal 30 to confirm the setting of the vocal mask and the position where the lyrics are added. In addition, when the setting position of the vocal mask and the addition position of the lyrics are not appropriate, the editing person in charge adjusts the setting position of the vocal mask and the addition position of the lyrics using the person-in-charge terminal 30. The setting position of the vocal mask is a sound range masked by the vocal mask and a reproduction time for which the vocal mask is set.
 再生管理部214は、ユーザ端末40における楽曲の再生状況を管理する処理を実行する。再生管理部214は、再生判定時間を算出するための情報を保持する。再生判定時間は、楽曲が再生されたか否かを判定するための時間である。 The playback management unit 214 executes processing for managing the playback status of the music on the user terminal 40. The reproduction management unit 214 holds information for calculating the reproduction determination time. The reproduction determination time is a time for determining whether or not the music has been reproduced.
 分配管理部215は、楽曲の再生状況に応じて、各分配対象者に原資を分配する処理を実行する。分配対象者は、例えば、著作権管理団体(例えばJASRAC(登録商標)等)、演奏者、レコード会社である。分配管理部215は、各関係者に対する分配金を決定するための分配率に関するデータを保持する。
 次に、各記憶部が記憶する情報の構成を説明する。
The distribution management unit 215 executes a process of distributing the resources to each distribution target person according to the reproduction status of the music. The distribution target person is, for example, a copyright management organization (for example, JASRAC (registered trademark)), a performer, or a record company. The distribution management unit 215 holds data regarding a distribution rate for determining a distribution amount for each party.
Next, the configuration of information stored in each storage unit will be described.
 図2に示すように、原盤記憶部22は、プロモーションビデオなどの原盤ファイル220を記憶する。管理サーバ20は、コンテンツサーバ10から原盤ファイル220を取得した場合に、その原盤ファイル220を記憶する。原盤ファイル220は、楽曲コード、属性、圧縮動画に関するデータを含む。 2, the master storage unit 22 stores a master file 220 such as a promotion video. When the management server 20 acquires the master file 220 from the content server 10, the management server 20 stores the master file 220. The master file 220 includes data related to music codes, attributes, and compressed moving images.
 楽曲コードデータ領域は、楽曲を特定するための識別子に関するデータを記憶する。属性データ領域は、楽曲の属性に関するデータを記憶する。楽曲の属性は、レコード会社、演奏者、作曲家、作詞家などに関する情報である。圧縮動画データ領域は、楽曲や映像についての圧縮データを記憶する。圧縮データは、例えば、mp4フォーマットで作成されている。 The music code data area stores data relating to an identifier for specifying a music. The attribute data area stores data related to the attributes of music. The attribute of the music is information relating to the record company, performer, composer, songwriter, and the like. The compressed moving image data area stores compressed data for music and video. The compressed data is created in the mp4 format, for example.
 図3に示すように、歌詞情報記憶部23は、歌詞に関する歌詞ファイル230を記憶する。歌詞ファイル230は、歌詞情報が管理サーバ20に登録された場合に記憶される。歌詞ファイル230は、楽曲コードや歌詞に関するデータを含む。 As shown in FIG. 3, the lyrics information storage unit 23 stores a lyrics file 230 related to lyrics. The lyrics file 230 is stored when the lyrics information is registered in the management server 20. The lyric file 230 includes data related to music codes and lyrics.
 楽曲コードデータ領域は、楽曲を特定するための識別子に関するデータを記憶する。歌詞データ領域は、歌詞テキストやルビなどに関する歌詞情報を記憶する。歌詞情報が示す歌詞テキストやルビは、ボーカルに対応したテロップとして機能する。 The music code data area stores data relating to an identifier for specifying a music. The lyric data area stores lyric information related to lyric text and ruby. Lyric text and ruby indicated by the lyric information function as a telop corresponding to vocals.
 図4に示すように、楽曲記憶部24は、ユーザ端末40に提供するための楽曲管理データ240を記憶する。楽曲管理データ240は、音楽原盤からカラオケデータが生成された場合に記憶される。 4, the music storage unit 24 stores music management data 240 to be provided to the user terminal 40. The music management data 240 is stored when karaoke data is generated from the master music.
 楽曲管理データ240は、楽曲コードに関連付けられると共に、HLSファイル241、XMLファイル242、ボーカルマスクファイル243、楽曲管理レコード244を含む。楽曲コードデータ領域は、楽曲を特定するための識別子に関するデータを記憶する。 The music management data 240 is associated with a music code and includes an HLS file 241, an XML file 242, a vocal mask file 243, and a music management record 244. The song code data area stores data relating to an identifier for specifying a song.
 HLSファイル241は、原盤の圧縮動画をトランスコードしたデータを含む。HLSファイル241は、高速回線用のHDフォーマットに従うデータと、低速回線用のSDフォーマットに従うデータとを含む。 The HLS file 241 includes data obtained by transcoding the compressed video of the master disc. The HLS file 241 includes data according to the HD format for high speed lines and data according to the SD format for low speed lines.
 XMLファイル242は、楽曲の属性情報や、再生時間であるタイムインデックスに関連付けられた歌詞を示す歌詞同期情報を記憶する。XMLファイル242は、楽曲の構成要素であるAメロ、Bメロ、Cメロの各々をタイムインデックスに関連付けた楽曲構成情報を含む。また、XMLファイル242は、メドレー楽曲における各チャプターをタイムインデックスに関連付けたチャプター情報を含む。XMLファイル242は、具体的には、以下の記述を含む。
 ・リリックス(lyrics):歌詞ファイルのバージョンが規定される。
 ・ヘッダ(header):楽曲のメタ情報やスタイル定義が記憶される。
 ・ソングメタ(song_meta):楽曲のメタ情報が記憶される。
 ・スタイル(style):スタイル定義と参照が記憶される。
 ・リリックストラック(lyrics_track):歌詞の本体が記憶される。
 ・バース(verse):歌詞の区切りが記憶される。
 ・ページ(page):1頁分の歌詞及び表示、消去タイミングが定義される。
 ・ライン(line):1行分の歌詞定義が記憶される。
 ・ルビ(ruby):ルビ定義が記憶される。
 ・ワード(word):歌詞テキスト及びワイプ処理タイミングが定義される。
The XML file 242 stores song attribute information and lyrics synchronization information indicating lyrics associated with a time index that is a reproduction time. The XML file 242 includes music composition information in which each of A melody, B melody, and C melody, which are musical composition elements, is associated with a time index. Also, the XML file 242 includes chapter information in which each chapter in the medley music is associated with a time index. Specifically, the XML file 242 includes the following description.
• Lyrics: Lyrics file versions are specified.
-Header: The music meta information and style definition are stored.
Song meta (song_meta): The meta information of the music is stored.
Style: Style definitions and references are stored.
-Lyrics track (lyrics_track): The lyrics body is stored.
・ Verse: The lyrics break is stored.
-Page: Lyrics for one page and display / erase timing are defined.
-Line: Lyric definitions for one line are stored.
-Ruby (ruby): Ruby definitions are stored.
Word: Lyric text and wipe processing timing are defined.
 ボーカルマスクファイル243は、タイムインデックスとボーカルの音域とを対応づけて、ボーカルの音域を記憶する。ボーカルマスクファイル243は、ボーカルラインの特定に用いられる。 The vocal mask file 243 stores the vocal range by associating the time index with the vocal range. The vocal mask file 243 is used for specifying a vocal line.
 ボーカル音域の音量を調整することは、ボーカルマスクファイル243を用いて、混合音からボーカルを抜くこと、ボーカルの音量を個別に小さくすることである。例えば、ユーザは、ボーカルの音量を個別に小さくすることによって、ボーカルをガイドとして利用する。
 楽曲管理レコード244は、演奏者、声質、曲調、演奏時間、映像利用、メドレー属性、ステータス、情景、音域に関するデータである。
 演奏者データ領域は、歌手などの演奏者を特定するための歌手名などの識別子に関するデータを記憶する。
 声質データ領域は、歌手の声についての特徴量に関するデータを記憶する。歌手の声についての特徴量は、例えば、主旋律から算出された統計値である。
To adjust the volume of the vocal range is to use the vocal mask file 243 to remove the vocal from the mixed sound and to individually reduce the volume of the vocal. For example, the user uses the vocal as a guide by individually reducing the volume of the vocal.
The music management record 244 is data relating to the performer, voice quality, tone, performance time, video usage, medley attribute, status, scene, and sound range.
The performer data area stores data relating to an identifier such as a singer name for identifying a performer such as a singer.
The voice quality data area stores data related to the feature amount of the singer's voice. The feature amount for the singer's voice is, for example, a statistical value calculated from the main melody.
 曲調データ領域は、楽曲のテンポやリズムに関するデータを記憶する。楽曲のテンポやリズムに関するデータは、例えば、楽曲から算出された統計値である特徴量である。 Tune data area stores data related to tempo and rhythm of music. The data relating to the tempo and rhythm of the music is, for example, a feature value that is a statistical value calculated from the music.
 情景データ領域は、例えば、楽曲から連想される季節、楽曲から連想される海や山や空などの景色、楽曲から連想される喜びや悲しみや憤りなどの感情などを識別するための情報を記憶する。
 音域データ領域は、楽曲のボーカルに求められる音域に関するデータを記憶する。
 演奏時間データ領域は、楽曲を最初から最後までの再生に要する時間である演奏時間に関するデータを記憶する。
 映像利用データ領域は、楽曲のHLSファイル241において、音楽原盤が映像を利用するか否かを識別するためのフラグを記憶する。
The scene data area stores information for identifying, for example, the season associated with the song, the scenery such as the sea, mountains, and sky associated with the song, and the emotions such as joy, sadness, and resentment associated with the song. To do.
The range data area stores data related to the range required for the vocal of the music.
The performance time data area stores data related to the performance time, which is the time required to reproduce the music from the beginning to the end.
The video use data area stores a flag for identifying whether or not the music master disk uses video in the music HLS file 241.
 メドレー属性データ領域は、楽曲がメドレーか否かを識別するための情報を記憶する。楽曲がメドレーの場合、メドレー属性データ領域は、メドレーを構成する楽曲の楽曲コードを記憶する。 The medley attribute data area stores information for identifying whether or not the music is medley. When the music is a medley, the medley attribute data area stores the music code of the music constituting the medley.
 ステータスデータ領域は、楽曲のHLSファイル241やXMLファイル242についての作成状況を特定するためのフラグを記憶する。ステータスデータ領域は、登録直後のファイルに対して「編集前」フラグを記憶し、編集後のファイルに対して「提供可能」フラグを記憶する。 The status data area stores a flag for specifying the creation status of the music HLS file 241 and XML file 242. The status data area stores a “before editing” flag for a file immediately after registration, and a “provided” flag for a file after editing.
 図5に示すように、再生実績記憶部25は、カラオケファイルの利用の実績に関する再生実績データ250を記憶する。再生実績データ250は、ユーザ端末40がカラオケファイルを利用した場合に記憶される。再生実績データ250は、ユーザ実績レコード251、および、楽曲実績レコード252を含む。
 ユーザ実績レコード251は、再生日時、ユーザコード、利用者コード、楽曲コードに関するデータを含む。
 再生日時データ領域は、ユーザ端末40において、楽曲が再生された年月日、および、時刻に関するデータを記憶する。
 ユーザコードデータ領域は、楽曲を再生したユーザを特定するための識別子に関するデータを記憶する。
 利用者コードデータ領域は、楽曲を再生した利用者を特定するための識別子に関するデータを記憶する。
 楽曲コードデータ領域は、再生された楽曲を特定するための識別子に関するデータを記憶する。
 楽曲実績レコード252は、楽曲コード、再生回数に関するデータを含む。
 楽曲コードデータ領域は、再生された楽曲を特定するための識別子に関するデータを記憶する。
 再生回数データ領域は、楽曲が再生された総回数に関するデータを記憶する。
As shown in FIG. 5, the reproduction record storage unit 25 stores reproduction record data 250 relating to the use record of the karaoke file. The reproduction result data 250 is stored when the user terminal 40 uses a karaoke file. The reproduction performance data 250 includes a user performance record 251 and a music performance record 252.
The user performance record 251 includes data related to playback date / time, user code, user code, and music code.
The reproduction date / time data area stores data related to the date and time when the music was reproduced in the user terminal 40.
The user code data area stores data relating to an identifier for specifying the user who has played the music.
The user code data area stores data relating to an identifier for identifying the user who has played the music.
The music code data area stores data relating to an identifier for specifying the reproduced music.
The track record 252 includes data related to the track code and the number of times of playback.
The music code data area stores data relating to an identifier for specifying the reproduced music.
The reproduction frequency data area stores data relating to the total number of times that a music piece has been reproduced.
 図6に示すように、原資記憶部26は、ユーザから取得するカラオケの利用料金に関する原資管理ファイル260を記憶する。カラオケの利用料金は、原資の一例である。原資管理ファイル260は、カラオケサービスの利用登録が行なわれた場合に記憶される。原資管理ファイル260は、ユーザコード、契約形態、情報料に関するデータを含む。
 ユーザコードデータ領域は、利用料金を支払うユーザを特定するための識別子に関するデータを記憶する。
As illustrated in FIG. 6, the resource storage unit 26 stores a resource management file 260 relating to a karaoke usage fee acquired from the user. Karaoke usage fees are an example of funding. The resource management file 260 is stored when use registration of the karaoke service is performed. The resource management file 260 includes data relating to a user code, a contract form, and an information fee.
The user code data area stores data relating to an identifier for identifying a user who pays a usage fee.
 契約形態データ領域は、ユーザとの契約形態を特定するためのフラグを記憶する。ユーザとの契約形態は、例えば、個別課金、定額支払(歌い放題)などのカスタマ向けの契約形態である。また、ユーザとの契約形態は、例えば、ユーザ端末40ごとの定額支払のようなビジネス向けの契約形態である。
 情報料データ領域は、ユーザとの契約形態に基づいて、サービス利用時に支払う金額である情報料に関するデータを記憶する。
The contract form data area stores a flag for specifying a contract form with the user. The contract form with the user is, for example, a contract form for customers such as individual billing, fixed amount payment (unlimited singing). Moreover, the contract form with a user is a contract form for business like the fixed amount payment for every user terminal 40, for example.
The information fee data area stores data related to the information fee, which is the amount to be paid when using the service, based on the contract form with the user.
 図7に示すように、分配管理情報記憶部27は、各分配対象者に対する原資の分配金を特定するための分配管理ファイル270を記憶する。分配管理ファイル270は、各分配対象者との支払条件が登録された場合に記憶される。分配管理ファイル270は、楽曲コード、支払先、分配条件に関するデータを含む。
 楽曲コードデータ領域は、楽曲を特定するための識別子に関するデータを記憶する。
As shown in FIG. 7, the distribution management information storage unit 27 stores a distribution management file 270 for specifying the distribution of the funds for each distribution target person. The distribution management file 270 is stored when payment conditions with each distribution target person are registered. The distribution management file 270 includes data relating to music codes, payees, and distribution conditions.
The song code data area stores data relating to an identifier for specifying a song.
 支払先データ領域は、楽曲に関する分配金の支払先に関するデータを記憶する。分配金の支払先は、例えば、分配対象者の一例である演奏者やレコード会社である。 The payee data area stores data related to payees of distributions related to music. The payee of the distribution money is, for example, a player or a record company that is an example of a distribution target person.
 分配条件データ領域は、支払先に対して支払う分配金の算出方法に関するデータを記憶する。分配条件データ領域は、契約形態が1曲ごとの個別課金である場合に、支払条件として、課金料金に対する割合を記憶する。分配条件データ領域は、契約形態が定額支払である場合、支払条件として、分配原資に基づいて、分配金を算出する方法を記憶する。 The distribution condition data area stores data relating to a calculation method of distribution paid to the payee. In the distribution condition data area, when the contract form is individual charging for each song, a ratio to the charging fee is stored as a payment condition. In the distribution condition data area, when the contract form is a fixed amount payment, a method for calculating the distribution money based on the distribution source is stored as the payment condition.
 図8に示すように、分配実績情報記憶部28は、分配先に対する分配実績に関する分配実績レコード280を記憶する。分配実績レコード280は、支払を行なう場合に記憶される。分配実績レコード280は、分配日、分配先、分配金額に関するデータを含む。
 分配日データ領域は、原資の分配を行なった年月日に関するデータを記憶する。
As shown in FIG. 8, the distribution record information storage unit 28 stores a distribution record record 280 related to the distribution record for the distribution destination. The distribution record 280 is stored when payment is made. The distribution record 280 includes data relating to distribution date, distribution destination, and distribution amount.
The distribution date data area stores data relating to the date of distribution of the funds.
 分配先データ領域は、分配金の分配先を特定するための識別子に関するデータを記憶する。分配先を特定するための識別子は、分配先である演奏者やレコード会社を特定するデータである。
 分配金額データ領域は、分配先に分配する金額に関するデータを記憶する。
The distribution destination data area stores data relating to an identifier for specifying the distribution destination of the distribution money. The identifier for specifying the distribution destination is data for specifying the performer or record company as the distribution destination.
The distribution amount data area stores data related to the amount to be distributed to the distribution destination.
 次に、サービス管理システムが行う動作を説明する。
 (原盤カラオケデータの生成処理)
 図9を用いて、原盤カラオケデータの生成処理を説明する。
 まず、制御部21は、原盤ファイルの一例であるプロモーションビデオを取得する処理を実行する(ステップS1-1)。
 具体的には、管理部210は、コンテンツサーバ10から、mp4ファイルとしてプロモーションビデオを取得する。さらに、管理部210は、プロモーションビデオの属性情報を取得する。そして、管理部210は、取得されたプロモーションビデオや属性情報を、原盤記憶部22に登録する。
Next, operations performed by the service management system will be described.
(Master karaoke data generation process)
The master karaoke data generation process will be described with reference to FIG.
First, the control unit 21 executes a process of acquiring a promotion video that is an example of a master file (step S1-1).
Specifically, the management unit 210 acquires a promotion video from the content server 10 as an mp4 file. Further, the management unit 210 acquires attribute information of the promotion video. Then, the management unit 210 registers the acquired promotion video and attribute information in the master disk storage unit 22.
 次に、制御部21は、トランスコード処理を実行する(ステップS1-2)。具体的には、管理部210は、原盤記憶部22に登録されたmp4ファイルを、HLSファイルに変換する。管理部210は、mp4ファイルを、高速回線に用いるHD用と、低速回線に用いるSD用とに変換する。そして、管理部210は、生成されたHLSファイルを、楽曲コードに関連付けて、楽曲記憶部24に登録する。 Next, the control unit 21 executes transcoding processing (step S1-2). Specifically, the management unit 210 converts the mp4 file registered in the master storage unit 22 into an HLS file. The management unit 210 converts the mp4 file into an HD file used for a high-speed line and an SD file used for a low-speed line. Then, the management unit 210 registers the generated HLS file in the music storage unit 24 in association with the music code.
 また、制御部21は、主旋律の推定処理を実行する(ステップS1-3)。具体的には、ボーカル処理部211は、mp4ファイルを所定のファイルフォーマットに変換する。そして、ボーカル処理部211は、楽曲におけるボーカルラインなどの主旋律を推定する。 Also, the control unit 21 executes main melody estimation processing (step S1-3). Specifically, the vocal processing unit 211 converts the mp4 file into a predetermined file format. Then, the vocal processing unit 211 estimates a main melody such as a vocal line in the music.
 次に、制御部21は、ボーカルマスクファイルの作成処理を実行する(ステップS1-4)。具体的には、ボーカル処理部211は、推定された主旋律に基づいて、ボーカル音域を特定する。 Next, the control unit 21 executes a vocal mask file creation process (step S1-4). Specifically, the vocal processing unit 211 identifies the vocal sound range based on the estimated main melody.
 また、ボーカル処理部211は、特定されたボーカル音域を記憶するボーカルマスクファイルを作成する。そして、ボーカル処理部211は、作成されたボーカルマスクファイルを、楽曲コードに関連付けて、楽曲記憶部24に記憶する。 Also, the vocal processing unit 211 creates a vocal mask file that stores the specified vocal range. The vocal processing unit 211 stores the created vocal mask file in the music storage unit 24 in association with the music code.
 次に、制御部21は、声質の登録処理を実行する(ステップS1-5)。具体的には、ボーカル処理部211は、楽曲に含まれるボーカルについてその声質の特徴量を算出する。そして、ボーカル処理部211は、算出された声質を、楽曲記憶部24に記憶された楽曲管理レコード244に登録する。 Next, the control unit 21 executes a voice quality registration process (step S1-5). Specifically, the vocal processing unit 211 calculates a feature quantity of the voice quality of the vocal included in the music. Then, the vocal processing unit 211 registers the calculated voice quality in the music management record 244 stored in the music storage unit 24.
 また、制御部21は、歌詞の割り当て処理を実行する(ステップS1-6)。具体的には、歌詞管理部212は、歌詞情報記憶部23から、処理対象楽曲の楽曲コードが記憶された歌詞ファイル230を取得する。 Further, the control unit 21 executes a lyrics allocation process (step S1-6). Specifically, the lyrics management unit 212 acquires the lyrics file 230 in which the song code of the processing target song is stored from the lyrics information storage unit 23.
 歌詞管理部212は、ボーカル処理部211によって特定されたボーカル音域に対応する再生時間と歌詞とを対応づけ、歌詞を割り付けた歌詞同期ファイルであるXMLファイル242を生成する。そして、歌詞管理部212は、XMLファイル242を楽曲記憶部24に登録する。 The lyrics management unit 212 associates the reproduction time corresponding to the vocal range specified by the vocal processing unit 211 with the lyrics, and generates an XML file 242 that is a lyrics synchronization file in which lyrics are assigned. Then, the lyrics management unit 212 registers the XML file 242 in the music storage unit 24.
 次に、制御部21は、編集処理を実行する(ステップS1-7)。具体的には、管理部210は、声質の登録処理、歌詞の割り当て処理の終了後、楽曲管理レコード244に「編集前」フラグを記憶する。そして、エディタ213は、作成されたHLSファイル241、XMLファイル242、および、ボーカルマスクファイル243を、担当者端末30に提供する。 Next, the control unit 21 executes an editing process (step S1-7). Specifically, the management unit 210 stores the “before editing” flag in the music management record 244 after the voice quality registration process and the lyrics allocation process are completed. Then, the editor 213 provides the created HLS file 241, XML file 242, and vocal mask file 243 to the person-in-charge terminal 30.
 担当者端末30は、編集担当者の操作に従い、カラオケファイルを再生し、ボーカル音の音量がボーカルマスクによって変わるか否か、ボーカル音と歌詞とが同期しているか否かなどを確認する。 The person-in-charge terminal 30 reproduces the karaoke file according to the operation of the person in charge of the editor, and confirms whether or not the volume of the vocal sound changes depending on the vocal mask, whether or not the vocal sound and the lyrics are synchronized.
 ボーカルマスクによるボーカルの調整が不十分な場合や、ボーカルと歌声との同期が取れていない場合には、エディタ213が、編集担当者の操作に従い、XMLファイル242やボーカルマスクファイル243を修正する。担当者端末30から編集終了情報を取得したエディタ213は、楽曲管理レコード244に「提供可能」フラグを記憶する。 When the vocal adjustment by the vocal mask is insufficient, or when the vocal and the singing voice are not synchronized, the editor 213 corrects the XML file 242 and the vocal mask file 243 according to the operation of the editor. The editor 213 that has acquired the editing end information from the person-in-charge terminal 30 stores the “provided” flag in the music management record 244.
 (ボーカルマスクファイルの生成処理)
 図10を用いて、ボーカルマスクファイルの生成処理を説明する。
 まず、ボーカル処理部211は、原盤ファイル220の属性データ領域に記憶されている歌手情報を取得する(ステップS2-1)。
 処理対象楽曲がソロ楽曲である場合、その歌手の歌手情報をボーカル処理部211は取得する。一方、処理対象楽曲がデュエット楽曲である場合、各歌手の歌手情報をボーカル処理部211は取得する。
(Vocal mask file generation process)
The vocal mask file generation process will be described with reference to FIG.
First, the vocal processing unit 211 acquires singer information stored in the attribute data area of the master file 220 (step S2-1).
When the music to be processed is a solo music, the vocal processing unit 211 acquires the singer information of the singer. On the other hand, when the music to be processed is a duet music, the vocal processing unit 211 acquires singer information of each singer.
 次に、ボーカル処理部211は、処理対象楽曲の曲調を特定する。ボーカル処理部211は、処理対象歌手を歌手データ領域に記憶した楽曲管理レコード244を楽曲記憶部24から抽出する。ボーカル処理部211は、抽出された楽曲管理レコード244において、処理対象楽曲の曲調と類似している楽曲を類似楽曲として特定する(ステップS2-2)。
 次に、制御部21は、処理対象歌手の声質を特定する処理を実行する(ステップS2-3)。声質の特定処理は、下記2つの方法を用いる。
Next, the vocal processing unit 211 identifies the tone of the music to be processed. The vocal processing unit 211 extracts, from the music storage unit 24, the music management record 244 in which the processing target singer is stored in the singer data area. In the extracted music management record 244, the vocal processing unit 211 specifies a music similar to the music tone of the music to be processed as a similar music (step S2-2).
Next, the control unit 21 executes processing for specifying the voice quality of the processing target singer (step S2-3). The voice quality identification process uses the following two methods.
 まず、制御部21は、類似楽曲によって声質を特定する処理を実行する(ステップS3-1)。具体的には、ボーカル処理部211は、類似楽曲の楽曲管理レコード244に記憶された歌手の声質であるボーカル音域を取得する。 First, the control unit 21 executes a process for specifying voice quality based on similar music (step S3-1). Specifically, the vocal processing unit 211 acquires a vocal range that is the voice quality of a singer stored in the music management record 244 of similar music.
 また、制御部21は、独唱箇所によって声質を特定する処理を実行する(ステップS3-2)。具体的には、ボーカル処理部211は、HLSファイル241に含まれる音声のなかで、処理対象歌手のみの音声が含まれる独唱領域を特定する。そして、ボーカル処理部211は、特定された独唱領域においてボーカル音域を特定する。 Also, the control unit 21 executes a process for specifying the voice quality by the singing location (step S3-2). Specifically, the vocal processing unit 211 identifies a singing area in which the sound of only the processing target singer is included among the sounds included in the HLS file 241. Then, the vocal processing unit 211 specifies the vocal sound range in the specified singing area.
 次に、制御部21は、第1のマスク生成処理を実行する(ステップS2-4)。制御部21は、特定されたボーカル音域を用い、曲全体を通して、ボーカル音域に対応したボーカルマスクファイルを生成する。 Next, the control unit 21 executes a first mask generation process (step S2-4). The control unit 21 generates a vocal mask file corresponding to the vocal range through the entire song using the specified vocal range.
 次に、制御部21は、ボーカルマスクが十分か否かの判定処理を実行する(ステップS2-5)。具体的には、ボーカル処理部211は、生成されたボーカルマスクを用いてボーカル音量を下げた楽曲を再生し、歌手の歌声を認知できるか否かを確認する。 Next, the control unit 21 determines whether or not the vocal mask is sufficient (step S2-5). Specifically, the vocal processing unit 211 uses the generated vocal mask to play a song with a reduced vocal volume, and confirms whether or not the singer's singing voice can be recognized.
 また、ボーカル処理部211は、楽曲の主旋律がボーカルマスク前後において歪んでいるか否かを確認する。歌声を認知できない場合や主旋律が歪んでいない場合、ボーカル処理部211は、ボーカルマスクが十分であると判定する。 Also, the vocal processing unit 211 confirms whether or not the main melody of the music is distorted before and after the vocal mask. If the singing voice cannot be recognized or the main melody is not distorted, the vocal processing unit 211 determines that the vocal mask is sufficient.
 ボーカルマスクが十分であるとボーカル処理部211が判定した場合(ステップS2-5において「YES」の場合)、制御部21は、処理対象歌手についてのボーカルマスクファイルの生成処理を終了する。 When the vocal processing unit 211 determines that the vocal mask is sufficient (in the case of “YES” in step S2-5), the control unit 21 ends the vocal mask file generation processing for the processing target singer.
 一方、ボーカルマスクが不十分であるとボーカル処理部211が判定した場合(ステップS2-5において「NO」の場合)、制御部21は、第2のマスク生成処理を実行する(ステップS2-6)。制御部21は、第2のマスク生成処理として、ボーカルのフレーズごとにボーカル音域を特定し、各ボーカル音域に対応したボーカルマスクファイルを生成する。
 次に、制御部21は、ステップS2-5と同様に、ボーカルマスクが十分か否かを判定する処理を実行する(ステップS2-7)。
On the other hand, when the vocal processing unit 211 determines that the vocal mask is insufficient (in the case of “NO” in step S2-5), the control unit 21 executes the second mask generation process (step S2-6). ). As the second mask generation process, the control unit 21 specifies a vocal range for each vocal phrase, and generates a vocal mask file corresponding to each vocal range.
Next, similarly to step S2-5, the control unit 21 executes processing for determining whether or not the vocal mask is sufficient (step S2-7).
 ボーカルマスクが十分であるとボーカル処理部211が判定した場合(ステップS2-7において「YES」の場合)、制御部21は、処理対象歌手についてのボーカルマスクファイルの生成処理を終了する。 When the vocal processing unit 211 determines that the vocal mask is sufficient (in the case of “YES” in step S2-7), the control unit 21 ends the vocal mask file generation processing for the processing target singer.
 一方、ボーカルマスクが不十分であるとボーカル処理部211が判定した場合(ステップS2-7において「NO」の場合)、制御部21は、第3のマスク生成処理を実行する(ステップS2-8)。制御部21は、第3のマスク生成処理として、ボーカル音域を所定の時間間隔ごとに推定し、動的に特定したボーカル音域に追従するボーカルマスクファイルを生成する。制御部21は、処理対象楽曲に含まれるすべての歌手について、以上の処理を繰り返す。 On the other hand, when the vocal processing unit 211 determines that the vocal mask is insufficient (in the case of “NO” in step S2-7), the control unit 21 executes a third mask generation process (step S2-8). ). As the third mask generation process, the control unit 21 estimates a vocal sound range for each predetermined time interval, and generates a vocal mask file that follows the dynamically specified vocal sound range. The control part 21 repeats the above process about all the singers contained in a process target music.
 <第1のマスク生成処理>
 図11を用いて、第1のマスク生成処理を説明する。
 ボーカル処理部211は、類似楽曲や独唱箇所に基づいて特定された歌手の声質を、この楽曲におけるボーカル音域であると推定する(ステップS4-1)。
 次に、ボーカル処理部211は、単一のボーカル音域をマスクするためのボーカルマスクファイルを生成する(ステップS4-2)。
 図12に示す時系列グラフ500のように、楽曲全体を通じて、単一のボーカル音域をマスクするボーカルマスク501が生成される。時系列グラフ500において、横軸は時間、および、縦軸は音域を示す。なお、時系列グラフ500において、上方の太線はギター音、下方の破線はドラム音の音域を示す。
<First mask generation process>
The first mask generation process will be described with reference to FIG.
The vocal processing unit 211 estimates that the voice quality of the singer specified based on the similar music or the singing part is the vocal sound range of this music (step S4-1).
Next, the vocal processing unit 211 generates a vocal mask file for masking a single vocal range (step S4-2).
As shown in the time series graph 500 shown in FIG. 12, a vocal mask 501 for masking a single vocal range is generated throughout the music. In the time-series graph 500, the horizontal axis indicates time, and the vertical axis indicates a sound range. In the time-series graph 500, the upper thick line indicates the guitar sound, and the lower broken line indicates the drum sound range.
 <第2のマスク生成処理>
 次に、第2のマスク生成処理を説明する。
 図13に示すように、ボーカル処理部211は、ボーカルラインに基づいて、ボーカルが含まれる再生時間帯を特定する。ボーカル処理部211は、歌唱が連続している一つのまとまりの領域をフレーズとして特定する(ステップS5-1)。
 ボーカル処理部211は、ボーカルラインに基づいてフレーズごとにボーカル音域を特定する(ステップS5-2)。
 次に、ボーカル処理部211は、フレーズごとに推定されたボーカル音域をマスクするためのボーカルマスクファイルを生成する(ステップS5-3)。
<Second Mask Generation Process>
Next, the second mask generation process will be described.
As shown in FIG. 13, the vocal processing unit 211 identifies a playback time zone in which the vocal is included based on the vocal line. The vocal processing unit 211 identifies a single group of continuous singing as a phrase (step S5-1).
The vocal processing unit 211 identifies a vocal range for each phrase based on the vocal line (step S5-2).
Next, the vocal processing unit 211 generates a vocal mask file for masking the vocal range estimated for each phrase (step S5-3).
 図14に示す時系列グラフ510のように、フレーズごとに作成されたボーカルマスク511,512が生成される。時系列グラフ510においても、横軸は時間、および、縦軸は音域を示す。なお、時系列グラフ510において、上方の太線はギター音、下方の破線はドラム音の音域を示す。 As shown in the time-series graph 510 shown in FIG. 14, vocal masks 511 and 512 created for each phrase are generated. Also in the time series graph 510, the horizontal axis indicates time, and the vertical axis indicates a sound range. In the time series graph 510, the upper thick line indicates the guitar sound, and the lower broken line indicates the drum sound range.
 <第3のマスク生成処理>
 次に、図15を用いて、第3のマスク生成処理を説明する。
 まず、ボーカル処理部211は、所定の時間であるタイムステップごとに楽曲を分割する(ステップS6-1)。そして、制御部21は、処理対象のタイムステップを順次特定し、以下の処理を繰り返す。
<Third mask generation process>
Next, the third mask generation process will be described with reference to FIG.
First, the vocal processing unit 211 divides the music for each time step that is a predetermined time (step S6-1). And the control part 21 specifies the time step of a process target sequentially, and repeats the following processes.
 制御部21は、特定されたタイムステップに含まれるボーカル音域を特定する処理を実行する(ステップS6-2)。次に、制御部21は、特定されたボーカル音域をマスクするボーカルマスクファイルを生成する処理を実行する(ステップS6-3)。 The control unit 21 executes a process for specifying a vocal range included in the specified time step (step S6-2). Next, the control unit 21 executes a process of generating a vocal mask file that masks the specified vocal range (step S6-3).
 図16に示す時系列グラフ520のように、タイムステップごとに作成されたボーカルマスク521,522が生成される。ボーカルマスク521,522は、ボーカル音域に対して動的にマスク領域が変化している。時系列グラフ520においても、横軸は時間、および、縦軸は音域を示す。なお、時系列グラフ520において、上方の太線はギター音、下方の破線はドラム音の音域を示す。 As shown in the time-series graph 520 shown in FIG. 16, vocal masks 521 and 522 created for each time step are generated. In the vocal masks 521 and 522, the mask area dynamically changes with respect to the vocal sound area. Also in the time series graph 520, the horizontal axis indicates time, and the vertical axis indicates the sound range. In the time series graph 520, the upper thick line indicates the guitar sound, and the lower broken line indicates the drum sound range.
 (デュエット対応処理)
 次に、図17を用いて、デュエット曲に対応した処理を説明する。
 時系列グラフ600は、女性ボーカル、および、男性ボーカルを含む例を示す。時系列グラフ601は、女性ボーカルのボーカルマスクファイルを用いて女性ボーカルの音量が個別に下げられた例を示す。時系列グラフ602は、男性ボーカルのボーカルマスクファイルを用いて男性ボーカルの音量が個別に下げられた例を示す。時系列グラフ603は、女性および男性の両ボーカルマスクファイルを用いて両ボーカルの音量が個別に下げられた例を示す。
(Duet compatible processing)
Next, the process corresponding to a duet music is demonstrated using FIG.
The time series graph 600 shows an example including a female vocal and a male vocal. A time series graph 601 shows an example in which the volume of a female vocal is individually lowered using a female vocal vocal mask file. The time series graph 602 shows an example in which the volume of the male vocal is individually lowered using the vocal mask file of the male vocal. A time series graph 603 shows an example in which the volume of both vocals is individually lowered using both female and male vocal mask files.
 (同期歌詞処理)
 次に、図18を用いて、同期歌詞の処理を説明する。
 まず、歌詞管理部212は、歌詞情報記憶部23から、処理対象楽曲の楽曲コードが記憶された歌詞ファイル230を抽出し、楽曲に割り付ける歌詞のフレーズを取得する(ステップS7-1)。
(Synchronized lyrics processing)
Next, synchronized lyrics processing will be described with reference to FIG.
First, the lyrics management unit 212 extracts the lyrics file 230 in which the song code of the processing target song is stored from the lyrics information storage unit 23, and acquires the lyrics phrase to be assigned to the song (step S7-1).
 次に、歌詞管理部212は、ボーカルラインに基づいて、歌声が含まれている歌声区間と、歌声が含まれていない非歌声区間とを推定する。そして、歌詞管理部212は、歌声区間を用いて、フレーズの開始位置である開始時間、および、終了位置である終了時間を特定する(ステップS7-2)。 Next, the lyrics management unit 212 estimates a singing voice section that includes a singing voice and a non-singing voice section that does not include a singing voice based on the vocal line. Then, the lyrics management unit 212 uses the singing voice section to specify the start time that is the start position of the phrase and the end time that is the end position (step S7-2).
 次に、歌詞管理部212は、XMLファイルにおいて、フレーズの開始位置、終了位置に対して、文字やルビを割り当てた記述を生成する(ステップS7-3)。
 そして、制御部21は、すべての歌詞を割り付けたか否かについての判定処理を実行する(ステップS7-4)。
Next, the lyrics management unit 212 generates a description in which characters and ruby are assigned to the start position and end position of the phrase in the XML file (step S7-3).
And the control part 21 performs the determination process about whether all the lyrics were allocated (step S7-4).
 割り付けられていない歌詞が残っていると制御部21が判定した場合(ステップS7-4において「NO」の場合)、制御部21は、割り付ける歌詞の特定処理(ステップS7-1)以降の処理を繰り返す。
 一方、すべての歌詞を割り付けたと判定した場合(ステップS7-4において「YES」の場合)、制御部21は、同期歌詞処理を終了する。
When the control unit 21 determines that lyrics that have not been assigned remain (in the case of “NO” in step S7-4), the control unit 21 performs processing subsequent to the processing for specifying the assigned lyrics (step S7-1). repeat.
On the other hand, if it is determined that all the lyrics have been assigned (“YES” in step S7-4), the control unit 21 ends the synchronized lyrics process.
 図19に示すXMLファイル700は、以上の処理によって生成される。記述701は、原盤ファイル220の属性データ領域に記憶された作詞、作曲、編曲などに関する楽曲メタデータを含む。記述702は、原盤ファイル220の属性データ領域に記憶されたテンポに関する楽曲メタデータを含む。 The XML file 700 shown in FIG. 19 is generated by the above processing. The description 701 includes song metadata relating to lyrics, composition, arrangement, etc. stored in the attribute data area of the master file 220. The description 702 includes music metadata relating to the tempo stored in the attribute data area of the master file 220.
 記述703は、ルビとして「はちがつ」が付加された「8月」を、所定期間にかけてワイプするための記述を含む。記述704は、ルビとして「たいよう」が付加された「太陽」を、所定期間にかけてワイプするための記述を含む。 The description 703 includes a description for wiping “August” with “hachigatsu” added as ruby over a predetermined period. The description 704 includes a description for wiping “sun” with “taiyo” added as ruby over a predetermined period.
 記述705は、「の」を所定期間にかけてワイプするための記述、記述706は、「ように」を所定期間にかけてワイプするための記述を含む。記述707は、カラオケ再生時において、一画面に表示させるページを変更するための記述を含む。
 このように、楽曲の進行に合わせて歌詞を表示させるためのXMLファイルが生成される。
The description 705 includes a description for wiping “NO” over a predetermined period, and the description 706 includes a description for wiping “NO” over a predetermined period. The description 707 includes a description for changing a page to be displayed on one screen during karaoke playback.
Thus, an XML file for displaying lyrics in accordance with the progress of the music is generated.
 (再生判定処理)
 図20を用いて、再生判定処理を説明する。この処理は、ユーザ端末40において、カラオケファイルが再生された場合に行なわれる。
 まず、再生管理部214は、ユーザ端末40に対してカラオケファイルを提供し、このカラオケファイルの再生を指示する(ステップS8-1)。
 次に、再生管理部214は、ユーザ端末40から、再生状況を取得し、再生が終了したか否かを判定する(ステップS8-2)。楽曲におけるすべての旋律が終了した場合や楽曲が途中で終了した場合、再生管理部214は、再生が終了したと判定する。
 次に、再生管理部214は、楽曲の開始時刻から、再生終了を検知した時刻までの再生時間を算出する(ステップS8-3)。
(Playback judgment process)
The playback determination process will be described with reference to FIG. This process is performed when a karaoke file is reproduced on the user terminal 40.
First, the reproduction management unit 214 provides a karaoke file to the user terminal 40 and instructs the user terminal 40 to reproduce the karaoke file (step S8-1).
Next, the playback management unit 214 acquires the playback status from the user terminal 40, and determines whether or not playback has ended (step S8-2). When all the melody in the music ends or when the music ends in the middle, the reproduction management unit 214 determines that the reproduction has ended.
Next, the playback management unit 214 calculates the playback time from the start time of the music to the time when the playback end is detected (step S8-3).
 また、再生管理部214は、楽曲記憶部24を用いて、再生された楽曲の楽曲コードが記憶された楽曲管理レコード244を抽出する。そして、再生管理部214は、抽出された楽曲管理レコード244から演奏時間である曲の長さを取得する(ステップS8-4)。 Also, the reproduction management unit 214 uses the music storage unit 24 to extract a music management record 244 in which the music code of the reproduced music is stored. Then, the reproduction management unit 214 acquires the length of the song that is the performance time from the extracted song management record 244 (step S8-4).
 次に、再生管理部214は、演奏時間に基づいて、再生判定時間を特定する(ステップS8-5)。再生管理部214は、演奏時間が60秒未満である楽曲に対して、再生判定時間として、演奏時間の判定基準割合である2/3を用いる。一方、再生管理部214は、演奏時間が60秒以上の楽曲に対して、再生判定時間として40秒を用いる。
 次に、再生管理部214は、再生判定時間が経過したか否かを判定する処理を実行する(ステップS8-6)。
Next, the reproduction management unit 214 specifies a reproduction determination time based on the performance time (step S8-5). The playback management unit 214 uses 2/3, which is a performance reference criterion, as a playback determination time for a song whose performance time is less than 60 seconds. On the other hand, the playback management unit 214 uses 40 seconds as a playback determination time for a musical piece having a performance time of 60 seconds or longer.
Next, the playback management unit 214 executes processing for determining whether or not the playback determination time has elapsed (step S8-6).
 再生時間が再生判定時間を経過している場合(ステップS8-6において「YES」の場合)、再生管理部214は、ユーザ実績レコード251を生成し、再生実績記憶部25に登録する。さらに、再生管理部214は、この楽曲についての楽曲実績レコード252の再生回数をインクリメントする(ステップS8-7)。 When the playback time has passed the playback determination time (in the case of “YES” in step S8-6), the playback management unit 214 generates a user result record 251 and registers it in the playback result storage unit 25. Furthermore, the reproduction management unit 214 increments the number of reproductions of the music record 252 for this music (step S8-7).
 なお、再生管理部214は、初めて再生された楽曲について、新たに楽曲実績レコード252を生成し、再生実績記憶部25に登録する。この場合、再生管理部214は、楽曲実績レコード252の再生回数に初期値「1」を設定する。 Note that the playback management unit 214 newly generates a music record record 252 for the music played for the first time and registers it in the playback record storage unit 25. In this case, the reproduction management unit 214 sets an initial value “1” for the number of reproductions of the music record record 252.
 一方、再生時間が再生判定時間を経過していない場合(ステップS8-6において「NO」の場合)、再生管理部214は、再生カウント処理(ステップS8-7)をスキップして再生判定処理を終了する。 On the other hand, if the playback time has not passed the playback determination time (“NO” in step S8-6), playback management unit 214 skips playback count processing (step S8-7) and performs playback determination processing. finish.
 (分配管理処理)
 次に、図21を用いて、分配管理処理を説明する。分割管理処理は、例えば1か月などのような所定期間ごとに実行される。
 まず、分配管理部215は、分配管理情報記憶部27に記憶された分配管理ファイル270の分配条件において、個別課金の楽曲を特定する(ステップS9-1)。
 次に、分配管理部215は、分配管理ファイル270に記憶された個別ルールに基づいて、分配金を算出する(ステップS9-2)。
 また、分配管理部215は、楽曲記憶部24の楽曲管理レコード244に記憶されたメドレー属性に基づいてメドレー曲を特定する(ステップS9-3)。
(Distribution management processing)
Next, the distribution management process will be described with reference to FIG. The division management process is executed every predetermined period such as one month.
First, the distribution management unit 215 specifies a piece of music for individual charging in the distribution conditions of the distribution management file 270 stored in the distribution management information storage unit 27 (step S9-1).
Next, the distribution management unit 215 calculates the distribution money based on the individual rules stored in the distribution management file 270 (step S9-2).
In addition, the distribution management unit 215 identifies a medley song based on the medley attribute stored in the song management record 244 of the song storage unit 24 (step S9-3).
 次に、分配管理部215は、メドレー曲の楽曲管理レコード244において、メドレー属性データ領域から、メドレー曲を構成する個別楽曲であるメドレー構成曲の楽曲コードを取得する(ステップS9-4)。 Next, the distribution management unit 215 obtains the music code of the medley constituent music that is an individual music constituting the medley music from the medley attribute data area in the music management record 244 of the medley music (step S9-4).
 次に、分配管理部215は、メドレー曲の再生回数をメドレー構成曲数で除算した分配回数を算出する。そして、分配管理部215は、算出された分配回数を、各メドレー構成曲の楽曲コードが記憶された楽曲実績レコード252の再生回数に加算する(ステップS9-5)。 Next, the distribution management unit 215 calculates the number of distributions obtained by dividing the number of times the medley music has been played by the number of medley constituent songs. Then, the distribution management unit 215 adds the calculated number of distributions to the number of reproductions of the music record record 252 in which the music code of each medley constituent music is stored (step S9-5).
 次に、制御部21は、カスタマ向け定額楽曲の分配処理を実行する(ステップS9-6)。具体的には、分配管理部215は、個別課金と歌い放題とを分けて、原資の分配を行なう。 Next, the control unit 21 executes a distribution process of the fixed amount music for the customer (step S9-6). Specifically, the distribution management unit 215 distributes resources by dividing individual billing and unlimited singing.
 次に、制御部21は、ビジネス向け定額楽曲の分配処理を実行する(ステップS9-7)。具体的には、分配管理部215は、端末毎の月額情報量に応じて、原資の分配を行なう。 Next, the control unit 21 executes a distribution process for a fixed amount music for business (step S9-7). Specifically, the distribution management unit 215 distributes resources according to the amount of monthly information for each terminal.
 (定額楽曲の分配処理)
 次に、図22を用いて、定額楽曲の分配処理を説明する。この処理においては、処理対象期間を定めて、カスタマ向けとビジネス向けとを分けて実行する。カスタマ向けの分配処理とビジネス向けの分配処理では、原資が異なるだけであって、処理の手順は相互に同じである。
(Distribution of fixed amount music)
Next, a fixed amount music distribution process will be described with reference to FIG. In this process, a process target period is determined, and the process is performed separately for the customer and the business. The distribution process for the customer and the distribution process for the business differ only in the resources, and the processing procedure is the same.
 まず、制御部21は、情報料総額の取得処理を実行する(ステップS10-1)。具体的には、分配管理部215は、原資記憶部26において、定額支払の情報料総額を取得する。分配管理部215は、カスタマ向けとビジネス向けとを分けて算出する。 First, the control unit 21 executes an information fee total acquisition process (step S10-1). Specifically, the distribution management unit 215 acquires the total amount of information fee for the fixed amount payment in the resource storage unit 26. The distribution management unit 215 calculates separately for customers and for businesses.
 次に、制御部21は、再生総数を算出する処理を実行する(ステップS10-2)。具体的には、分配管理部215は、分配管理情報記憶部27に記憶された分配管理ファイル270において、個別課金が記憶されていない楽曲の楽曲コードを特定する。次に、分配管理部215は、特定された楽曲コードが記憶されている楽曲実績レコード252を、再生実績記憶部25から取得する。そして、分配管理部215は、取得された楽曲実績レコード252に記憶されている再生回数を合計する。 Next, the control unit 21 executes a process for calculating the total number of reproductions (step S10-2). Specifically, the distribution management unit 215 specifies the music code of the music for which the individual charge is not stored in the distribution management file 270 stored in the distribution management information storage unit 27. Next, the distribution management unit 215 obtains a music record record 252 in which the specified music code is stored from the reproduction record storage unit 25. Then, the distribution management unit 215 sums the number of reproductions stored in the acquired music record record 252.
 次に、制御部21は、楽曲ごとに以下の処理を繰り返す。
 まず、制御部21は、再生回数を取得する処理を実行する(ステップS10-3)。具体的には、分配管理部215は、再生実績記憶部25から、処理対象の楽曲コードが記憶された楽曲実績レコード252を取得する。
Next, the control part 21 repeats the following processes for every music.
First, the control unit 21 executes a process for obtaining the number of reproductions (step S10-3). Specifically, the distribution management unit 215 acquires a music record record 252 in which the music code to be processed is stored from the reproduction record storage unit 25.
 次に、制御部21は、再生割合を算出する処理を実行する(ステップS10-4)。具体的には、分配管理部215は、処理対象の楽曲の再生回数を、再生総数で除算することによって、再生割合を算出する。 Next, the control unit 21 executes a process for calculating the reproduction ratio (step S10-4). Specifically, the distribution management unit 215 calculates the reproduction ratio by dividing the number of reproductions of the music to be processed by the total number of reproductions.
 次に、制御部21は、分配金を算出する処理を実行する(ステップS10-5)。具体的には、分配管理部215は、原資記憶部26から、処理対象期間の原資を取得する。ここで、分配管理部215は、カスタマ向けの処理には、「歌い放題」についての情報料総額を算出する。一方、分配管理部215は、ビジネス向けの処理には、端末毎の情報量総額を算出する。そして、分配管理部215は、情報料総額に対して、処理対象楽曲の再生割合を乗算することによって、分配金額を算出する。 Next, the control unit 21 executes a process for calculating the distribution money (step S10-5). Specifically, the distribution management unit 215 acquires the resources for the processing target period from the resource storage unit 26. Here, the distribution management unit 215 calculates the total information fee for “all-you-can sing” for the process for the customer. On the other hand, the distribution management unit 215 calculates the total amount of information for each terminal for business-oriented processing. Then, the distribution management unit 215 calculates the distribution amount by multiplying the total information fee by the reproduction ratio of the music to be processed.
 次に、制御部21は、分配金における著作権管理団体分の算出処理を実行する(ステップS10-6)。具体的には、分配管理部215は、分配金に著作権管理団体の分配率を乗算することによって、著作権管理団体分の分配金額を算出する。そして、分配管理部215は、分配実績レコード280を生成し、それを分配実績情報記憶部28に記憶する。 Next, the control unit 21 executes a calculation process for the copyright management organization in the distribution (step S10-6). Specifically, the distribution management unit 215 calculates the distribution amount for the copyright management organization by multiplying the distribution money by the distribution rate of the copyright management organization. Then, the distribution management unit 215 generates a distribution record record 280 and stores it in the distribution record information storage unit 28.
 次に、制御部21は、分配金における演奏者分の算出処理を実行する(ステップS10-7)。具体的には、分配管理部215は、分配金に楽曲分の分配率を乗算することによって、演奏者分の分配金額を算出する。そして、分配管理部215は、分配実績レコード280を生成し、それを分配実績情報記憶部28に記憶する。 Next, the control unit 21 executes a calculation process for the performer in the distribution money (step S10-7). Specifically, the distribution management unit 215 calculates the distribution amount for the performer by multiplying the distribution amount by the distribution rate for the song. Then, the distribution management unit 215 generates a distribution record record 280 and stores it in the distribution record information storage unit 28.
 次に、制御部21は、処理対象が映像を利用するか否かの判定処理を実行する(ステップS10-8)。具体的には、分配管理部215は、楽曲記憶部24に記憶されたHLSファイル241に映像が含まれるか否かを確認する。楽曲管理レコード244の映像利用データ領域に利用フラグが設定されている場合、処理対象が映像を利用するものであると分配管理部215は判定する。 Next, the control unit 21 executes a process for determining whether or not the processing target uses video (step S10-8). Specifically, the distribution management unit 215 confirms whether or not the video is included in the HLS file 241 stored in the music storage unit 24. When the use flag is set in the video use data area of the music management record 244, the distribution management unit 215 determines that the processing target uses video.
 処理対象が映像を利用するとき(ステップS10-8において「YES」の場合)、分配金における映像分の算出処理を制御部21は実行する(ステップS10-9)。具体的には、分配管理部215は、分配金に映像分の分配率を乗算することによって、映像分の分配金額を算出する。そして、分配管理部215は、分配実績レコード280を生成し、それを分配実績情報記憶部28に記憶する。 When the processing target uses video (in the case of “YES” in step S10-8), the control unit 21 executes calculation processing for the video in the distribution (step S10-9). Specifically, the distribution management unit 215 calculates the distribution amount for the video by multiplying the distribution by the distribution rate for the video. Then, the distribution management unit 215 generates a distribution record record 280 and stores it in the distribution record information storage unit 28.
 一方、処理対象が映像を利用しないとき(ステップS10-8において「NO」の場合)、制御部21は、分配金における映像分の算出処理(ステップS10-9)をスキップして、定額楽曲の分配処理を終了する。 On the other hand, when the processing target does not use the video (in the case of “NO” in step S10-8), the control unit 21 skips the calculation processing for the video in the distribution (step S10-9), and creates the fixed amount music. The distribution process ends.
 以上、第1実施形態によれば以下に示す効果が得られる。
 (1)音楽原盤を用いて、カラオケファイルを生成することができるため、演奏者と同じ伴奏でサービス利用者は歌うことができる。
 (2)複数のボーカルが含まれる場合においても、歌手ごとのボーカルマスクファイルが生成されるため、ボーカルの音量を歌手ごとに調整することができる。
 (3)ボーカルマスクが不十分であると判定された場合に第2のマスク生成処理が実行されるため、ボーカルマスクによる楽曲の歪を抑制することができる。また、第3のマスク生成処理が実行されるため、精緻なボーカルマスクを生成することができる。
As described above, according to the first embodiment, the following effects can be obtained.
(1) Since a karaoke file can be generated using the master music, the service user can sing with the same accompaniment as the performer.
(2) Even when a plurality of vocals are included, since a vocal mask file for each singer is generated, the volume of the vocal can be adjusted for each singer.
(3) Since the second mask generation process is executed when it is determined that the vocal mask is insufficient, distortion of music due to the vocal mask can be suppressed. Further, since the third mask generation process is executed, it is possible to generate a precise vocal mask.
 (4)再生時間が再生判定時間を経過した場合に、再生カウント処理が実行されるため、的確な再生回数をカウントすることができる。
 (5)個別ルールに基づいて分配金の算出処理が実行されるため、分配対象者との契約に基づいて分配金を算出することができる。
 (6)複数の楽曲を含む多様なメドレー曲においても、構成曲が考慮された分配金を分配することができる。
 (7)定額料金制が適用される場合にも、再生状況に応じた分配金を算出することができる。また、分配金における映像分の算出処理も実行されるため、映像についても原資を分配することができる。
(4) Since the reproduction count process is executed when the reproduction time has passed the reproduction determination time, the appropriate number of reproductions can be counted.
(5) Since the distribution calculation process is executed based on the individual rules, the distribution can be calculated based on the contract with the distribution target person.
(6) Even in various medley songs including a plurality of music pieces, it is possible to distribute the distribution money in consideration of the constituent music pieces.
(7) Even when the flat fee system is applied, it is possible to calculate the distribution according to the reproduction status. Further, since the calculation processing for the video in the distribution money is also executed, it is possible to distribute the resources for the video.
 なお、上記第1実施形態は、以下の態様に変更して実施できる。
 ・制御部21が行う声質の特定処理(ステップS2-3)は、類似楽曲による声質の特定処理(ステップS3-1)と、独唱箇所による声質の特定処理(ステップS3-2)とのいずれか一方でもよい。また、両方の方法によって声質が特定された場合、制御部21は、所定のルールに基づいて特定された声質を用いてボーカルマスクファイルを生成する。所定のルールは、例えば、両方のボーカル音域を含む音域を声質として特定することである。
In addition, the said 1st Embodiment can be implemented by changing into the following aspects.
The voice quality specifying process (step S2-3) performed by the control unit 21 is any one of a voice quality specifying process (step S3-1) using a similar music piece and a voice quality specifying process using a solo location (step S3-2). Either one is acceptable. When the voice quality is specified by both methods, the control unit 21 generates a vocal mask file using the voice quality specified based on a predetermined rule. The predetermined rule is, for example, to specify a range including both vocal ranges as a voice quality.
 ・分配管理ファイル270に含まれる分配条件は、ボーカル音域を特定することの困難性に基づいて設定されてもよい。例えば、第1のマスク処理によってボーカルマスクファイルが生成された場合には、第2のマスク処理によってボーカルマスクファイルが生成される場合よりも、演奏者分として低い分配率が設定される。さらに、第2のマスク処理によってボーカルマスクファイルが生成された場合には、第3のマスク処理によってボーカルマスクファイルが生成される場合よりも、演奏者分として低い分配率が設定される。これによって、ボーカルマスクファイルを生成することの困難性に応じて、分配率を変更することができる。 The distribution conditions included in the distribution management file 270 may be set based on the difficulty of specifying the vocal sound range. For example, when a vocal mask file is generated by the first mask process, a lower distribution rate is set for the performer than when a vocal mask file is generated by the second mask process. Furthermore, when the vocal mask file is generated by the second mask process, a lower distribution ratio is set for the performer than when the vocal mask file is generated by the third mask process. Thereby, the distribution rate can be changed according to the difficulty of generating the vocal mask file.
 ・楽曲が再生されたか否かを判定する方法は、例えば、楽曲におけるサビの位置が再生されたか否かを判定することであってもよい。この場合には、制御部21が、主旋律を推定する処理(ステップS1-3)において、特定のパターンに基づいてサビの位置を特定する。そして、楽曲管理レコード244に、特定したサビの位置をタイムインデックスに対応付けて記憶しておく。そして、再生管理部214は、再生時間、および、サビの位置に基づいて、楽曲が再生されたか否かの判定を行なう。再生管理部214は、再生時間が再生判定時間を経過していない場合であっても、サビが再生された場合には、再生と判定する。
 ・単一のグループを構成する各ユーザが、自身のユーザ端末40を用いてカラオケサービスを利用する場合には、以下のように再生回数を調整してもよい。
The method for determining whether or not the music has been reproduced may be, for example, determining whether or not the position of the chorus in the music has been reproduced. In this case, the control unit 21 specifies the position of the rust based on the specific pattern in the process of estimating the main melody (step S1-3). Then, the specified chorus position is stored in the music management record 244 in association with the time index. Then, the playback management unit 214 determines whether or not the music has been played based on the playback time and the position of the chorus. Even if the playback time has not passed the playback determination time, the playback management unit 214 determines that playback has been performed when rust is played back.
-When each user who comprises a single group uses a karaoke service using his user terminal 40, you may adjust the frequency | count of reproduction | regeneration as follows.
 図23に示すように、複数のユーザ端末40a,40b、40cが同時期に同じ楽曲を再生した場合には、再生回数が調整される。例えば、管理サーバ20は、グループ情報記憶部を備える。グループ情報記憶部は、グループ管理コードに対して、グループとして利用する各ユーザ端末40のユーザコードに関するデータを記憶する。管理サーバ20は、カラオケサービスを利用する楽曲の指定を取得した場合に、楽曲指定が行なわれたユーザ端末40のユーザコードに関連付けて親機情報を記憶する。そして、サービス管理システムは、図24に示すグループ再生管理処理を実行する。 As shown in FIG. 23, when a plurality of user terminals 40a, 40b, and 40c play the same music at the same time, the number of times of playback is adjusted. For example, the management server 20 includes a group information storage unit. The group information storage unit stores data related to the user code of each user terminal 40 used as a group with respect to the group management code. When the management server 20 acquires the designation of a song that uses the karaoke service, the management server 20 stores the parent device information in association with the user code of the user terminal 40 for which the song is designated. Then, the service management system executes a group reproduction management process shown in FIG.
 (グループ再生管理処理)
 まず、制御部21は、グループを登録する処理を実行する(ステップS11-1)。具体的には、複数のユーザが各自のユーザ端末40を用いてカラオケサービスを利用する場合に、各ユーザ端末40が管理サーバ20にアクセスする。再生管理部214は、グループに対してグループ管理コードを付与する。再生管理部214は、グループ管理コードを用いてログインしたユーザ端末40について、グループ情報記憶部に、グループ管理コードに関連付けてユーザコードを記憶する(グループ登録)。ここでは、ユーザ端末40a~40cが、同じグループ管理コードを用いて、ログインした場合を想定する。
(Group playback management process)
First, the control unit 21 executes a process for registering a group (step S11-1). Specifically, when a plurality of users use the karaoke service using their own user terminals 40, each user terminal 40 accesses the management server 20. The reproduction management unit 214 assigns a group management code to the group. The reproduction management unit 214 stores the user code associated with the group management code in the group information storage unit for the user terminal 40 logged in using the group management code (group registration). Here, it is assumed that the user terminals 40a to 40c log in using the same group management code.
 次に、制御部21は、楽曲指定を受け付ける処理を実行する(ステップS11-2)。具体的には、再生管理部214は、グループ登録された任意のユーザ端末40から、利用したい楽曲の楽曲コードを取得する。例えば、ユーザ端末40bから楽曲指定を受信した場合を想定する。 Next, the control unit 21 executes a process of accepting music designation (step S11-2). Specifically, the reproduction management unit 214 obtains the music code of the music to be used from any user terminal 40 registered as a group. For example, it is assumed that a music designation is received from the user terminal 40b.
 次に、制御部21は、指定者を親機として特定する処理を実行する(ステップS11-3)。具体的には、再生管理部214は、グループ管理コードが付与されているユーザ端末40a~40cにおいて、先に指定された楽曲についてユーザ端末40bを親機として登録する。 Next, the control unit 21 executes a process for specifying the designated person as the parent device (step S11-3). Specifically, the reproduction management unit 214 registers the user terminal 40b as a parent device for the previously specified music in the user terminals 40a to 40c to which the group management code is assigned.
 次に、制御部21は、同期を指示するための処理を実行する(ステップS11-4)。具体的には、再生管理部214は、グループ登録されたユーザ端末40a~40cに対して同期指示を送信する。この場合、再生管理部214は、グループ登録された各ユーザ端末40a~40cにおける楽曲のダウンロード状況を取得する。そして、再生管理部214は、すべてのユーザ端末40a~40cのダウンロードが完了するまで、各ユーザ端末40に待機を指示する。 Next, the control unit 21 executes a process for instructing synchronization (step S11-4). Specifically, the reproduction management unit 214 transmits a synchronization instruction to the user terminals 40a to 40c registered in the group. In this case, the reproduction management unit 214 acquires the music download status at each of the user terminals 40a to 40c registered in the group. Then, the reproduction management unit 214 instructs each user terminal 40 to wait until downloading of all the user terminals 40a to 40c is completed.
 次に、制御部21は、再生処理を実行する(ステップS11-5)。具体的には、再生管理部214は、グループ登録されたすべてのユーザ端末40a~40cにおいて、ダウンロードの完了情報を取得した場合、同時に楽曲の再生開始を指示する。 Next, the control unit 21 executes a reproduction process (step S11-5). Specifically, when the download management information is acquired in all the user terminals 40a to 40c registered in the group, the playback management unit 214 instructs to start playback of the music at the same time.
 次に、制御部21は、再生実績を記憶する処理を実行する(ステップS11-6)。具体的には、再生管理部214は、親機として指定されたユーザ端末40bのユーザコードを用いて、ユーザ実績レコード251を登録する。 Next, the control unit 21 executes a process for storing the playback results (step S11-6). Specifically, the reproduction management unit 214 registers the user performance record 251 using the user code of the user terminal 40b designated as the parent device.
 また、再生管理部214は、グループ登録の場合には、同じ場所での楽曲再生を条件としてもよい。この場合には、各ユーザ端末40a~40cから所在情報を取得する。この所在情報は、例えばユーザ端末40a~40cが備えているGPS機能による位置情報を用いることができる。これによって、同じ場所で同時期にサービスを利用していることを制御部21は判定できる。 In the case of group registration, the playback management unit 214 may use the playback of music at the same place as a condition. In this case, location information is acquired from each of the user terminals 40a to 40c. As the location information, for example, position information by a GPS function provided in the user terminals 40a to 40c can be used. Accordingly, the control unit 21 can determine that the service is being used at the same place at the same time.
 ・歌詞の同期状態は、ユーザからの情報に基づいて確認されてもよい。ここでは、管理サーバ20が、修正情報記憶部を備え、歌詞同期調整の処理を実行する。修正情報記憶部は、ユーザから取得された修正リクエストに関するリクエスト管理レコードを記憶する。リクエスト管理レコードは、楽曲コードに対して、ユーザコード、再生時間、点数、対応結果に関するデータを記憶する。
 ユーザコードデータ領域は、修正リクエストを行なったユーザを特定するための識別子に関するデータを記憶する。
 再生時間データ領域は、楽曲において修正を希望する箇所を特定するためのタイムインデックスに関するデータを記憶する。
 点数データ領域は、修正リクエストに対して付与された点数に関するデータを記憶する。
The lyrics synchronization status may be confirmed based on information from the user. Here, the management server 20 includes a correction information storage unit, and executes lyrics synchronization adjustment processing. The correction information storage unit stores a request management record relating to a correction request acquired from the user. The request management record stores data related to the user code, the reproduction time, the score, and the correspondence result for the music code.
The user code data area stores data relating to an identifier for specifying a user who has made a correction request.
The reproduction time data area stores data relating to a time index for specifying a portion desired to be corrected in the music.
The score data area stores data related to the score given to the correction request.
 対応結果データ領域は、修正リクエストに対する対応結果を特定するためのフラグを記憶する。修正が行われた場合の「修正」フラグ、または、修正が行なわれていない場合の「修正不要」フラグのいずれかが記憶される。 The response result data area stores a flag for specifying the response result for the correction request. Either the “correction” flag when the correction is performed or the “correction unnecessary” flag when the correction is not performed is stored.
 さらに、歌詞管理部212は、修正の要否を確認するための修正確認基準点に関するデータを記憶する。また、歌詞管理部212は、ユーザ属性に基づいて、点数に付与する重み付け情報を記憶する。重み付け情報は、修正リクエストに基づいて修正が行なわれた修正割合に対して付与する重み付け値を記憶する。 Furthermore, the lyrics management unit 212 stores data related to the correction confirmation reference point for confirming whether correction is necessary. In addition, the lyrics management unit 212 stores weighting information to be given to the score based on the user attribute. The weighting information stores a weighting value to be given to the correction ratio corrected based on the correction request.
 次に、図25を用いて歌詞同期調整処理を説明する。
 (歌詞同期調整処理)
 まず、制御部21は、ユーザの修正情報を取得する処理を実行する(ステップS12-1)。具体的には、ユーザ端末40における再生時に、歌詞の同期状態にずれがあると判断された場合、サービス利用者は、ユーザ端末40において、修正依頼を入力する。この場合、ユーザ端末40は、管理サーバ20に修正リクエストを送信する。この修正リクエストは、楽曲コード、修正依頼入力が行なわれた再生時間に関するデータを含める。歌詞管理部212は、ユーザ端末40からの修正リクエストを受信する。
Next, the lyrics synchronization adjustment process will be described with reference to FIG.
(Lyrics synchronization adjustment processing)
First, the control unit 21 executes a process of acquiring user correction information (step S12-1). Specifically, when it is determined that there is a deviation in the synchronization state of lyrics during reproduction on the user terminal 40, the service user inputs a correction request on the user terminal 40. In this case, the user terminal 40 transmits a correction request to the management server 20. This correction request includes data relating to the music code and the playback time at which the correction request was input. The lyrics management unit 212 receives a correction request from the user terminal 40.
 次に、制御部21は、修正情報を記憶する処理を実行する(ステップS12-2)。具体的には、歌詞管理部212は、取得された修正リクエストについてのリクエスト管理レコードを生成し、修正情報記憶部にそれを記憶する。リクエスト管理レコードは、楽曲コード、ユーザコード、再生時間に関するデータを含める。 Next, the control unit 21 executes processing for storing correction information (step S12-2). Specifically, the lyrics management unit 212 generates a request management record for the acquired correction request, and stores it in the correction information storage unit. The request management record includes data related to the music code, user code, and playback time.
 次に、制御部21は、ユーザ属性に応じて重み付けされた点数を付与する処理を実行する(ステップS12-3)。具体的には、歌詞管理部212は、修正情報記憶部から、ユーザコードが記憶されたリクエスト管理レコードにおいて、既処理フラグが記憶されたレコードを抽出する。そして、歌詞管理部212は、リクエスト数に対して、修正フラグが記憶されたレコード数の修正割合を算出する。そして、歌詞管理部212は、修正割合に応じて重み付けした点数を、今回の修正リクエストに対して付与して、リクエスト管理レコードに記憶する。 Next, the control unit 21 executes a process of assigning points weighted according to user attributes (step S12-3). Specifically, the lyrics management unit 212 extracts, from the correction information storage unit, the record in which the processed flag is stored in the request management record in which the user code is stored. Then, the lyrics management unit 212 calculates the correction ratio of the number of records in which the correction flag is stored with respect to the number of requests. And the lyrics management part 212 gives the score weighted according to the correction ratio with respect to this correction request, and memorize | stores it in a request management record.
 次に、制御部21は、修正リクエストと同じ修正があるか否かを判定する処理を実行する(ステップS12-4)。具体的には、歌詞管理部212は、修正情報記憶部から、修正リクエストと同じ楽曲コードが記憶され、かつ、対応結果データ領域にフラグが記憶されていない修正管理レコードを抽出する。また、歌詞管理部212は、抽出された修正管理レコードのなかで、修正リクエストと同じ再生時間帯が記憶されたレコードを抽出する。 Next, the control unit 21 executes a process for determining whether or not there is the same correction as the correction request (step S12-4). Specifically, the lyrics management unit 212 extracts from the correction information storage unit a correction management record in which the same music code as the correction request is stored and no flag is stored in the corresponding result data area. Further, the lyrics management unit 212 extracts a record in which the same playback time zone as the correction request is stored from the extracted correction management records.
 修正リクエストと同じ修正があると判定された場合(ステップS12-4において「YES」の場合)、制御部21は、点数の合計処理を実行する(ステップS12-5)。具体的には、歌詞管理部212は、抽出された修正管理レコードに記憶された点数を合計し、それを合計点として算出する。 When it is determined that there is the same correction as the correction request (in the case of “YES” in step S12-4), the control unit 21 executes the total processing of the points (step S12-5). Specifically, the lyrics management unit 212 sums up the points stored in the extracted correction management records, and calculates it as a total score.
 一方、修正リクエストと同じ修正がないと判定された場合(ステップS12-4において「NO」の場合)、制御部21は、点数の合計処理(ステップS12-5)をスキップする。 On the other hand, when it is determined that there is no correction same as the correction request (in the case of “NO” in step S12-4), the control unit 21 skips the score totaling process (step S12-5).
 次に、制御部21は、合計点が基準点を越えているか否かを判定する処理を実行する(ステップS12-6)。具体的には、歌詞管理部212は、算出された合計点と修正確認基準点とを比較する。
 合計点が基準点を越えていない場合(ステップS12-6において「NO」の場合)、制御部21は、歌詞同期調整処理を終了する。
Next, the control unit 21 executes processing for determining whether or not the total score exceeds the reference point (step S12-6). Specifically, the lyrics management unit 212 compares the calculated total score with the correction confirmation reference point.
If the total score does not exceed the reference point (“NO” in step S12-6), the control unit 21 ends the lyrics synchronization adjustment process.
 一方、合計点が基準点を越えている場合(ステップS12-6において「YES」の場合)、制御部21は、修正を提案する処理を実行する(ステップS12-7)。具体的には、歌詞管理部212は、担当者端末30に対して、確認リクエストを送信する。確認リクエストは、リクエスト管理レコードに記憶された楽曲コード、再生時間に関する情報を含む。 On the other hand, when the total score exceeds the reference point (in the case of “YES” in step S12-6), the control unit 21 executes a process for proposing correction (step S12-7). Specifically, the lyrics management unit 212 transmits a confirmation request to the person-in-charge terminal 30. The confirmation request includes information related to the music code and the reproduction time stored in the request management record.
 次に、制御部21は、修正判断処理を実行する(ステップS12-8)。確認リクエストを取得した担当者端末30は、確認リクエスト内容を表示する。この場合、編集担当者は、楽曲を再生し、歌詞の同期状態を確認する。 Next, the control unit 21 executes a correction determination process (step S12-8). The person-in-charge terminal 30 that has acquired the confirmation request displays the content of the confirmation request. In this case, the editor in charge reproduces the music and confirms the synchronization state of the lyrics.
 そして、歌詞の同期が適切でないと編集担当者が判断した場合に、編集担当者はXMLファイルを修正する。また、歌詞管理部212は、同じ修正リクエストに関するすべての修正管理レコードにおいて、修正済みフラグを記憶する。 Then, when the editor in charge determines that the lyrics are not synchronized properly, the editor in charge modifies the XML file. In addition, the lyrics management unit 212 stores a corrected flag in all correction management records related to the same correction request.
 一方、歌詞の同期が適切であると編集担当者が判断した場合、編集担当者は担当者端末30から修正不要を入力する。また、歌詞管理部212は、同じ修正リクエストに関するすべての修正不要フラグを修正管理レコードに記憶する。
 ・提供されるコンテンツは、ボーカルの音量が個別に調整されたバックグラウンドミュージックや、ボーカルの音量が個別に調整された映像であってもよい。
On the other hand, when the editor in charge determines that the synchronization of lyrics is appropriate, the editor in charge inputs correction unnecessary from the person in charge terminal 30. In addition, the lyrics management unit 212 stores all correction unnecessary flags related to the same correction request in the correction management record.
The provided content may be background music in which the volume of the vocal is individually adjusted or video in which the volume of the vocal is individually adjusted.
 (第2実施形態)
 以下、サービス管理システムの第2実施形態を説明する。第2実施形態のサービス管理システムは、ボーカルの音量を個別に変える度合いを設定することが第1実施形態やその変形例とは異なる。そのため、以下では、第1実施形態やその変形例とは異なる点を特に説明し、相互に重複する構成の説明を割愛する。
(Second Embodiment)
Hereinafter, a second embodiment of the service management system will be described. The service management system according to the second embodiment is different from the first embodiment and the modification thereof in that the degree of individually changing the vocal volume is set. Therefore, below, especially a different point from 1st Embodiment and its modification is demonstrated, and the description of the structure which mutually overlaps is omitted.
 図26が示すように、ユーザ端末40は、制御部41A、通信部42A、入力部43、データ記憶部44、音量検出部45、および、楽曲出力部46を備える。制御部41Aは、CPU、DSP、ROM、RAMなどから構成される。制御部41Aは、入力部43、通信部42A、データ記憶部44、音量検出部45、および、楽曲出力部46などの各部が実行する処理を制御する。 26, the user terminal 40 includes a control unit 41A, a communication unit 42A, an input unit 43, a data storage unit 44, a volume detection unit 45, and a music output unit 46. The control unit 41A includes a CPU, DSP, ROM, RAM, and the like. 41 A of control parts control the process which each part, such as the input part 43, the communication part 42A, the data storage part 44, the sound volume detection part 45, and the music output part 46, performs.
 入力部43は、例えば、操作ボタンなどを供えるリモートコントローラやボリュームスイッチなどである。入力部43は、制御部41Aが処理できる形式で各種の情報を制御部41Aに入力する。入力部43が入力する情報は、指定モード、指定マスク度、および、指定期間を含む。 The input unit 43 is, for example, a remote controller or a volume switch that provides operation buttons. The input unit 43 inputs various types of information to the control unit 41A in a format that can be processed by the control unit 41A. The information input by the input unit 43 includes a designated mode, a designated mask degree, and a designated period.
 マスク度は、楽曲データに基づき楽曲の再生が行われている期間において、ボーカルが個別にマスクされる度合いを示す。言い換えれば、マスク度は、混合音の一例である楽曲の音量に対してボーカルの音量が相対的に下げられる度合いである。 The degree of masking indicates the degree to which vocals are individually masked during the period when the music is being played based on the music data. In other words, the degree of masking is the degree to which the volume of the vocal is relatively lowered with respect to the volume of the music that is an example of the mixed sound.
 なお、楽曲の音量に対するボーカルの音量のなかで、マスクされていないときのボーカルの音量は、原盤ファイルから再生されている音量と等しい原盤音量である。これに対して、マスクされているときのボーカルの音量は、原盤音量よりも小さいサポート音量である。 It should be noted that the vocal volume when not masked in the vocal volume relative to the volume of the music is the master volume equal to the volume being played from the master file. On the other hand, the volume of the vocal when masked is a support volume smaller than the master volume.
 指定マスク度は、カラオケサービスの利用者がユーザ端末40に入力するマスク度である。指定マスク度は、楽曲データを用いる楽曲の再生が行われる前に、その楽曲に含まれるボーカルに対して、1つずつ入力することが可能である。例えば、楽曲がソロ楽曲である場合、ボーカルに対して1つの指定マスク度を入力することが可能である。楽曲がデュエット曲である場合、2種類のボーカルの各々に対して1つずつ指定マスク度を入力することが可能である。 The designated mask degree is a mask degree input by the user of the karaoke service to the user terminal 40. The designated mask degree can be input one by one with respect to the vocals included in the music before the music using the music data is reproduced. For example, when the music is a solo music, it is possible to input one designated mask degree for the vocal. When the music is a duet music, it is possible to input a designated mask degree for each of the two types of vocals.
 指定モードは、カラオケサービスの利用者がユーザ端末40に入力する再生の形式である。指定モードは、楽曲データを用いる楽曲の再生が行われる前に、その楽曲の再生に対して、1つずつ入力することが可能である。指定モードは、例えば、固定モード、第1変動モード、および、第2変動モードのなかのいずれか1つである。
 制御部41Aは、固定モードにおいて、マスク度として固定値を採用し、楽曲データから楽曲を再生する際に、ボーカルを固定値でマスクする。
The designation mode is a reproduction format that is input to the user terminal 40 by the user of the karaoke service. The designated mode can be input one by one with respect to the reproduction of the music before the music using the music data is reproduced. The designation mode is, for example, any one of a fixed mode, a first variation mode, and a second variation mode.
41 A of control parts employ | adopt a fixed value as a mask degree in fixed mode, and mask a vocal with a fixed value, when reproducing | regenerating a music from music data.
 制御部41Aは、第1変動モードにおいて、マスク度として変動値を採用し、楽曲データから楽曲を再生する際に、ボーカルを変動値に基づきマスクする。第1変動モードにおける変動値は、音量検出部45の検出した音量である演奏音の検出音量に従って変わる。 In the first variation mode, the control unit 41A employs a variation value as the masking degree, and masks the vocal based on the variation value when reproducing the song from the song data. The variation value in the first variation mode changes according to the detected sound volume of the performance sound that is the sound volume detected by the sound volume detector 45.
 具体的には、第1変動モードにおける変動値は、例えば、基準音量よりも検出音量が大きいほど、サポート音量が小さいように変わり、また、例えば、基準音量よりも検出音量が小さいほどサポート音量が大きいように変わる。基準音量は、原盤音量を一定の割合でマスクすることによって得られる音量であってもよいし、一定の音量であってもよい。 Specifically, the variation value in the first variation mode changes, for example, as the detected volume is larger than the reference volume, so that the support volume is lower. For example, as the detected volume is smaller than the reference volume, the support volume is It changes as big. The reference volume may be a volume obtained by masking the master volume at a certain rate, or may be a certain volume.
 制御部41Aは、第2変動モードにおいて、マスク度として変動値を採用し、楽曲データから楽曲を再生する際に、ボーカルを変動値でマスクする。第2変動モードにおける変動値は、検出音量とサポート音量との総量を原盤音量に合わせるように変わる。 In the second variation mode, the control unit 41A adopts a variation value as the mask degree, and masks the vocal with the variation value when reproducing the song from the song data. The variation value in the second variation mode changes so that the total amount of the detected volume and the support volume matches the master volume.
 入力部43は、楽曲が再生される期間のなかで指定モードが適用される期間である適用期間を、制御部41Aが処理できる形式で制御部41Aに入力する。適用期間は、例えば、楽曲を構成するAメロ、Bメロ、Cメロ、サビ、チャプターの少なくとも1つである。 The input unit 43 inputs, to the control unit 41A, an application period that is a period in which the designated mode is applied during a period in which music is played, in a format that can be processed by the control unit 41A. The application period is, for example, at least one of A melody, B melody, C melody, rust, and chapter that compose the music.
 通信部42Aは、管理サーバ20に対するカラオケファイルの送信要求や、管理サーバ20に対するカラオケファイルの受信連絡を実行する。制御部41Aは、通信部42Aが実行するデータの送信と、通信部42Aが実行するデータの受信とを制御する。 42 A of communication parts perform the transmission request | requirement of the karaoke file with respect to the management server 20, and the reception notification of the karaoke file with respect to the management server 20. FIG. The control unit 41A controls transmission of data executed by the communication unit 42A and reception of data executed by the communication unit 42A.
 データ記憶部44は、入力部43が制御部41Aに入力した各種のデータ、および、通信部42Aが受信したカラオケファイルなどの各種のデータを記憶するための記憶領域を備える。データ記憶部44は、ボーカルの音量を変えるために制御部41Aが実行する音楽再生プログラム、および、ボーカルの音量を変えるために制御部41Aが用いる各種のデータを記憶する。 The data storage unit 44 includes a storage area for storing various data input by the input unit 43 to the control unit 41A and various data such as a karaoke file received by the communication unit 42A. The data storage unit 44 stores a music playback program executed by the control unit 41A to change the volume of the vocal and various data used by the control unit 41A to change the volume of the vocal.
 データ記憶部44が記憶するデータは、マスク度のデフォルトとして標準マスク度を含む。また、データ記憶部44が記憶するデータは、適用期間のデフォルトとして標準期間を含む。 The data stored in the data storage unit 44 includes a standard mask degree as a default mask degree. The data stored in the data storage unit 44 includes a standard period as a default application period.
 標準マスク度は、楽曲の再生に際して、その楽曲の再生に対して指定マスク度が入力されなかったときに、制御部41Aにおいて指定マスク度に代えて採用される。 The standard mask degree is adopted in place of the designated mask degree in the control unit 41A when the designated mask degree is not inputted for the reproduction of the music during the reproduction of the music.
 標準期間は、楽曲の再生に際して、その楽曲の再生に対して指定期間が入力されなかったときに、制御部41Aにおいて指定期間に代えて採用される。なお、データ記憶部44が記憶するデータは、制御部41Aが指定モードを特定するための識別子に関するデータを含む。
 制御部41Aは、データ記憶部44が実行するデータの記憶と、データ記憶部44に記憶されたデータのデータ記憶部44からの読み出しとを制御する。
The standard period is adopted in place of the designated period in the control unit 41A when the designated period is not input for the reproduction of the music. The data stored in the data storage unit 44 includes data related to an identifier for the control unit 41A to specify the designated mode.
The control unit 41 </ b> A controls data storage executed by the data storage unit 44 and reading of data stored in the data storage unit 44 from the data storage unit 44.
 制御部41Aは、データ記憶部44が記憶する音楽再生プログラムを読み出し、読み出された音楽再生プログラムに従って、楽曲データから楽曲を再生するための音楽再生信号を生成する。 The control unit 41A reads the music reproduction program stored in the data storage unit 44, and generates a music reproduction signal for reproducing the music from the music data according to the read music reproduction program.
 制御部41Aは、生成された楽曲再生信号を楽曲出力部46に入力し、楽曲再生信号に基づく再生を楽曲出力部46に実行させる。制御部41A、および、楽曲出力部46は、楽曲再生プログラムを実行することによって、音量設定部、および、再生処理部として機能する。 The control unit 41A inputs the generated music reproduction signal to the music output unit 46, and causes the music output unit 46 to perform reproduction based on the music reproduction signal. The control unit 41A and the music output unit 46 function as a volume setting unit and a playback processing unit by executing a music playback program.
 音量検出部45は、ユーザ端末40に入力された演奏音の音量を検出音量として検出する。音量検出部45は、例えば、ボーカルを演じる利用者がマイクに入力した歌声の音量を検出音量として検出する。
 制御部41Aは、音量設定部411と、再生信号生成部412とを備える。
The volume detector 45 detects the volume of the performance sound input to the user terminal 40 as a detected volume. The volume detection unit 45 detects, for example, the volume of a singing voice input to a microphone by a user who plays vocals as a detected volume.
The control unit 41A includes a volume setting unit 411 and a reproduction signal generation unit 412.
 音量設定部411は、カラオケファイルに含まれる楽曲データ、および、ボーカルマスクと、データ記憶部44が記憶する指定マスク度、および、標準マスク度を用い、ボーカルの音量を変えるための処理を実行する。 The volume setting unit 411 executes a process for changing the volume of the vocal using the music data and the vocal mask included in the karaoke file, the designated mask degree stored in the data storage unit 44, and the standard mask degree. .
 ボーカルマスクは、ボーカルの音階を含む音域であるボーカルラインを、再生音の音域のなかで特定するためのデータである。ボーカルマスクは、ボーカルの音域をデータとして含み、ボーカルの音域は、楽曲の再生時間であるタイムインデックスに対して規定される。ボーカルラインに含まれる音は、ボーカルマスクに基づいて特定され、特定された音の音量は、マスク度に基づいて変えられ、それによって、サポート音量が変わる。 The vocal mask is data for specifying a vocal line, which is a range including the vocal scale, within the range of the reproduced sound. The vocal mask includes a vocal range as data, and the vocal range is defined with respect to a time index that is a reproduction time of a song. The sound included in the vocal line is specified based on the vocal mask, and the volume of the specified sound is changed based on the mask degree, thereby changing the support volume.
 音量設定部411は、固定モードや第1変動モードにおいて、指定マスク度が入力されている場合、音量の設定として、指定マスク度を設定する。一方、音量設定部411は、指定マスク度が入力されていない場合、音量の設定として、標準マスク度を設定する。音量設定部411が実行する音量設定のなかで、第1変動モード、および、第2変動モードによる設定は、変動するマスク度を用いた補助設定である。 The volume setting unit 411 sets the designated mask degree as the volume setting when the designated mask degree is input in the fixed mode or the first variation mode. On the other hand, the volume setting unit 411 sets the standard mask degree as the volume setting when the designated mask degree is not input. Among the volume settings executed by the volume setting unit 411, the settings in the first variation mode and the second variation mode are auxiliary settings using a varying mask degree.
 音量設定部411は、固定モード、第1変動モード、および、第2変動モードにおいて、指定期間が入力されている場合、音量の設定として、指定期間を設定する。一方、音量設定部411は、指定期間が入力されていない場合、標準期間を設定する。 The volume setting unit 411 sets the specified period as the volume setting when the specified period is input in the fixed mode, the first variation mode, and the second variation mode. On the other hand, the volume setting unit 411 sets the standard period when the specified period is not input.
 例えば、ユーザ端末40の実行する設定の形式が固定モードであり、かつ、指定マスク度が入力されており、また、指定期間が入力されている場合、音量設定部411は、ボーカルの音量を変えるための設定の処理を以下のように実行する。 For example, when the setting format executed by the user terminal 40 is the fixed mode, the specified mask degree is input, and the specified period is input, the volume setting unit 411 changes the volume of the vocal. The processing for setting is executed as follows.
 すなわち、音量設定部411は、楽曲データ、ボーカルマスク、指定マスク度、指定期間を参照する。音量設定部411は、楽曲データから再生されるボーカルの音量が指定マスク度でマスクされたサポート音量となるように、再生信号生成部412が採用するマスク度に指定マスク度を設定する。そして、音量設定部411は、再生信号生成部412が指定マスク度を採用する期間として、指定期間を設定する。 That is, the volume setting unit 411 refers to music data, vocal mask, designated mask degree, and designated period. The volume setting unit 411 sets the designated masking degree to the masking degree adopted by the reproduction signal generating unit 412 so that the volume of the vocal played from the music data becomes the support volume masked by the designated masking degree. Then, the sound volume setting unit 411 sets the designated period as a period during which the reproduction signal generation unit 412 adopts the designated mask degree.
 例えば、ユーザ端末40の実行する設定の形式が第1変動モードであり、かつ、指定マスク度が入力されており、また、指定期間が入力されている場合、音量設定部411は、ボーカルの音量を変えるための設定の処理を以下のように実行する。 For example, when the setting format executed by the user terminal 40 is the first variation mode, the specified mask degree is input, and the specified period is input, the volume setting unit 411 sets the volume of the vocal. The setting process for changing the value is executed as follows.
 すなわち、音量設定部411は、楽曲データ、ボーカルマスク、指定マスク度、指定期間を参照する。音量設定部411は、楽曲データからボーカルの再生が開始されるとき、楽曲データから再生されるボーカルの音量がサポート音量となるように、再生信号生成部412が採用するマスク度に指定マスク度を設定する。 That is, the volume setting unit 411 refers to music data, vocal mask, designated mask degree, and designated period. The volume setting unit 411 sets the specified mask degree to the mask degree adopted by the reproduction signal generation unit 412 so that the vocal volume reproduced from the music data becomes the support volume when vocal reproduction is started from the music data. Set.
 そして、ボーカルの再生が開始されて以降、音量検出部45の検出した音量が下がるときにサポート音量を上げるよう、音量設定部411は、再生信号生成部412が採用するマスク度を下げる。反対に、音量設定部411は、音量検出部45の検出した音量が上がるときにサポート音量を下げるよう、再生信号生成部412が採用するマスク度を上げる。なお、音量設定部411は、再生信号生成部412がこれらのマスク度を採用する期間として、指定期間を設定する。 Then, after the reproduction of the vocal is started, the volume setting unit 411 decreases the masking degree adopted by the reproduction signal generation unit 412 so that the support volume is increased when the volume detected by the volume detection unit 45 is decreased. On the contrary, the volume setting unit 411 increases the masking degree adopted by the reproduction signal generation unit 412 so that the support volume is lowered when the volume detected by the volume detection unit 45 is increased. Note that the volume setting unit 411 sets a designated period as a period during which the reproduction signal generation unit 412 adopts these mask degrees.
 例えば、ユーザ端末40の実行する設定の形式が第2変動モードであり、指定期間が入力されている場合、音量設定部411は、ボーカルの音量を変えるための設定の処理を以下のように実行する。 For example, when the setting format executed by the user terminal 40 is the second variation mode and the specified period is input, the volume setting unit 411 executes the setting process for changing the volume of the vocal as follows. To do.
 すなわち、音量設定部411は、楽曲データ、ボーカルマスク、指定期間を参照し、再生信号生成部412が採用するマスク度の調整を逐次実行する。この際、音量検出部45が検出した音量とサポート音量との総量が原盤音量となるように、音量設定部411はマスク度を調整する。また、音量設定部411は、再生信号生成部412がこうしたマスク度を採用する期間として、指定期間を設定する。 That is, the volume setting unit 411 sequentially performs adjustment of the mask degree adopted by the reproduction signal generation unit 412 with reference to the music data, vocal mask, and specified period. At this time, the volume setting unit 411 adjusts the mask degree so that the total volume of the volume detected by the volume detection unit 45 and the support volume becomes the master volume. Further, the volume setting unit 411 sets a designated period as a period in which the reproduction signal generation unit 412 adopts such a mask degree.
 再生信号生成部412は、管理サーバ20からダウンロードされた楽曲データから、楽曲を再生するための信号である楽曲再生信号を生成する。再生信号生成部412は、管理サーバ20からダウンロードされた映像データから、映像を再生するための信号である映像再生信号を生成する。 The reproduction signal generation unit 412 generates a music reproduction signal that is a signal for reproducing music from the music data downloaded from the management server 20. The reproduction signal generation unit 412 generates a video reproduction signal that is a signal for reproducing a video from the video data downloaded from the management server 20.
 再生信号生成部412は、楽曲データとボーカルマスクとを用い、楽曲データから再生する楽曲のなかからボーカルを特定する。すなわち、再生信号生成部412は、楽曲を再生するための信号を楽曲データから生成すると共に、生成された信号のなかでボーカルに相当する信号を特定する。また、再生信号生成部412は、音量設定部411が設定したマスク度を採用する。そして、再生信号生成部412は、ボーカルとして特定される信号がマスク度でマスクされた信号となるように、楽曲を再生するための信号を加工し、それを楽曲再生信号とする。なお、再生信号生成部412は、マスク度が採用される期間として、音量設定部411が設定した期間を採用し、その設定された期間でのみ上記加工の処理を実行する。 The reproduction signal generation unit 412 uses the music data and the vocal mask, and identifies the vocal from the music reproduced from the music data. In other words, the reproduction signal generation unit 412 generates a signal for reproducing the music from the music data, and specifies a signal corresponding to vocal among the generated signals. Further, the reproduction signal generation unit 412 employs the mask degree set by the volume setting unit 411. Then, the reproduction signal generation unit 412 processes a signal for reproducing the music so that the signal specified as vocal becomes a signal masked with the mask degree, and uses it as a music reproduction signal. Note that the reproduction signal generation unit 412 adopts the period set by the volume setting unit 411 as the period in which the masking degree is adopted, and executes the above processing only during the set period.
 図27が示すように、まず、音量設定部411は、管理サーバ20から提供されたカラオケファイルに含まれる楽曲データをデータ記憶部44から読み出す(ステップS13-1)。 As shown in FIG. 27, first, the volume setting unit 411 reads the music data included in the karaoke file provided from the management server 20 from the data storage unit 44 (step S13-1).
 次に、音量設定部411は、管理サーバ20から提供されたカラオケファイルに含まれるボーカルマスクをデータ記憶部44から読み出す(ステップS13-2)。 Next, the volume setting unit 411 reads the vocal mask included in the karaoke file provided from the management server 20 from the data storage unit 44 (step S13-2).
 次に、音量設定部411は、今回の楽曲再生に際して入力部43から入力された指定モードをデータ記憶部44から読み出し、読み出された指定モードを再生信号生成部412に設定する(ステップS13-3)。 Next, the volume setting unit 411 reads the designation mode input from the input unit 43 during the current music reproduction from the data storage unit 44, and sets the read designation mode in the reproduction signal generation unit 412 (step S13-). 3).
 音量設定部411は、指定期間が入力されている場合、今回の楽曲再生に採用されるマスク度の採用期間として、指定期間を再生信号生成部412に設定する。音量設定部411は、指定期間が入力されていない場合、今回の楽曲再生に採用されるマスク度の採用期間として、標準期間を再生信号生成部412に設定する。 When the designated period is input, the volume setting unit 411 sets the designated period in the reproduction signal generating unit 412 as the adoption period of the mask degree adopted for the current music reproduction. When the designated period is not input, the volume setting unit 411 sets the standard period in the reproduction signal generation unit 412 as the adoption period of the mask degree employed for the current music reproduction.
 音量設定部411は、指定マスク度が入力されている場合、今回の楽曲再生に採用されるマスク度を指定マスク度として再生信号生成部412に設定する。音量設定部411は、指定マスク度が入力されていない場合、今回の楽曲再生に採用されるマスク度を標準マスク度として再生信号生成部412に設定する。 When the specified mask level is input, the volume setting unit 411 sets the mask level used for the current music playback to the playback signal generation unit 412 as the specified mask level. When the designated mask level is not input, the volume setting unit 411 sets the mask level employed for the current music playback as the standard mask level in the playback signal generation unit 412.
 次に、制御部41Aは、今回の楽曲再生に採用される指定モードが固定モードか否かを判定する(ステップS13-4)。
 今回の楽曲再生に採用される指定モードが第1変動モード、もしくは、第2変動モードであると判定された場合(ステップS13-4において「NO」の場合)、制御部41Aは、指定モードである第1変動モード、もしくは、第2変動モードによる楽曲の再生を開始する(ステップS13-5)。
Next, the control unit 41A determines whether or not the designated mode employed for the current music reproduction is the fixed mode (step S13-4).
When it is determined that the designated mode employed for the current music reproduction is the first variation mode or the second variation mode (in the case of “NO” in step S13-4), the control unit 41A performs the designation mode. The reproduction of music in a certain first variation mode or second variation mode is started (step S13-5).
 一方、今回の楽曲再生に採用される指定モードが固定モードであると判定された場合(ステップS13-4において「YES」の場合)、制御部41Aは、指定モードである固定モードによる楽曲の再生を開始する(ステップS13-15)。 On the other hand, when it is determined that the designated mode employed in the current music reproduction is the fixed mode (in the case of “YES” in step S13-4), the control unit 41A reproduces the music in the fixed mode which is the designated mode. Is started (step S13-15).
 次に、制御部41Aは、第1変動モードでの楽曲の再生、あるいは、第2変動モードでの楽曲の再生において、再生時間が指定期間に到達するごとに、指定期間内において、音量検出部45が検出する音量に基づいて、以下の処理を繰り返す。 Next, the control unit 41A performs a sound volume detection unit within the designated period every time the reproduction time reaches the designated period in the reproduction of the music in the first fluctuation mode or the reproduction of the music in the second fluctuation mode. The following processing is repeated based on the volume detected by 45.
 すなわち、制御部41Aの音量設定部411は、第1変動モードでの楽曲の再生において、音量検出部45が音量を検出するごとに、音量検出部45の検出した音量が基準音量よりも下がるときにサポート音量を上げるよう、再生信号生成部412が採用するマスク度を下げる。反対に、音量設定部411は、音量検出部45の検出した音量が基準音量よりも上がるときにサポート音量を下げるよう、再生信号生成部412が採用するマスク度を上げる。 In other words, the volume setting unit 411 of the control unit 41A, when playing the music in the first variation mode, every time the volume detection unit 45 detects the volume, the volume detected by the volume detection unit 45 falls below the reference volume. The masking degree adopted by the reproduction signal generation unit 412 is lowered so as to increase the support volume. On the contrary, the volume setting unit 411 increases the masking degree adopted by the reproduction signal generation unit 412 so that the support volume is lowered when the volume detected by the volume detection unit 45 is higher than the reference volume.
 また、音量設定部411は、第2変動モードでの楽曲の再生において、音量検出部45が音量を検出するごとに、音量検出部45が検出した音量とサポート音量との総量が原盤音量となるように、マスク度を調整する(ステップS13-6)。 In addition, the volume setting unit 411, when playing the music in the second variation mode, every time the volume detection unit 45 detects the volume, the total volume of the volume detected by the volume detection unit 45 and the support volume becomes the master volume. Thus, the mask degree is adjusted (step S13-6).
 そして、再生信号生成部412は、今回の検出結果が反映されたマスク度を用い、ボーカルの新たな音量での再生音を生成するための楽曲再生信号を生成する(ステップS13-7)。 Then, the reproduction signal generation unit 412 generates a music reproduction signal for generating a reproduction sound at a new vocal volume by using the mask degree reflecting the detection result of this time (step S13-7).
 なお、制御部41Aは、固定モードでの楽曲の再生において、再生時間が指定期間に到達するごとに、ボーカルの音量が今回のマスク度でマスクされるように標準マスク度を設定する(ステップS13-17)。そして、再生信号生成部412は、今回の設定されたマスク度を用い、ボーカルの新たな音量での再生音を生成するための楽曲再生信号を生成する(ステップS13-18)。
 図28から図30を参照して固定モード、第1変動モード、および、第2変動モードによる楽曲の再生結果の一例を説明する。
Note that the control unit 41A sets the standard mask degree so that the volume of the vocal is masked with the current mask degree every time the reproduction time reaches the specified period in the reproduction of the music in the fixed mode (step S13). -17). Then, the reproduction signal generation unit 412 generates a music reproduction signal for generating a reproduction sound at a new vocal volume using the mask degree set this time (step S13-18).
With reference to FIGS. 28-30, an example of the reproduction | regeneration result of the music by fixed mode, 1st fluctuation mode, and 2nd fluctuation mode is demonstrated.
 なお、図28から図30の各々において、実線は、音量検出部45が検出した音量である検出音量の再生時間に対する推移を示す。破線は、ユーザ端末40が再生した楽曲のなかのボーカルであるサポート音量の再生時間に対する推移を示す。二点鎖線は、マスクされていないボーカルの音量である原盤音量の再生時間に対する推移を示す。 In each of FIGS. 28 to 30, the solid line indicates the transition of the detected volume, which is the volume detected by the volume detector 45, with respect to the playback time. A broken line indicates transition of the support volume, which is a vocal in the music reproduced by the user terminal 40, with respect to the reproduction time. The alternate long and two short dashes line shows the transition of the master volume, which is the volume of the unmasked vocal, with respect to the playback time.
 図28は、固定モードにおける各音の音量の推移を示す。図29は、第1変動モードにおける各音の音量の推移を示す。図30は、第2変動モードにおける各音の音量の推移を示す。 FIG. 28 shows the transition of the volume of each sound in the fixed mode. FIG. 29 shows the transition of the volume of each sound in the first variation mode. FIG. 30 shows the transition of the volume of each sound in the second variation mode.
 図28が示すように、再生時間が経過することに伴い、検出音量は原盤音量の変化に追従するように変化する。一方で、再生時間が期間Tであるとき、例えば、ボーカルに求められる声域から利用者の声域が外れたり、利用者の声が枯れたりすることに伴って、利用者がボーカルを奏でることができず、それによって、検出音量がほとんど検出されないレベルにまで下がる。 As shown in FIG. 28, as the playback time elapses, the detected volume changes so as to follow the change in the master volume. On the other hand, when the playback time is the period T, for example, the user can play vocals as the user's voice range deviates from the vocal range required for vocals or the user's voice dies. Accordingly, the detected sound volume is lowered to a level where it is hardly detected.
 ここで、一定のマスク度でマスクされたサポート音量は、原盤音量よりも小さい音量で、それの変化に追従するように、再生時間の全体にわたり変化する。結果として、利用者がボーカルを演じるべき期間Tにおいては、利用者がボーカルを奏でることができないとしても、ユーザ端末40の再生するサポート音と、それ以外の音とによって楽曲は構成される。そのため、楽曲を奏でることがユーザ端末40によって補助される。 Here, the support volume masked at a certain masking degree is smaller than the master volume, and changes throughout the playback time so as to follow the change. As a result, in the period T in which the user should play vocals, even if the user cannot play vocals, the music is composed of the support sound played by the user terminal 40 and other sounds. Therefore, playing the music is assisted by the user terminal 40.
 図29が示すように、ここでもまた、再生時間が経過することに伴い、検出音量は原盤音量の変化に追従するように変化する。一方で、再生時間が期間Tであるとき、検出音量がほとんど検出されないレベルにまで下がる。 29, as shown in FIG. 29, the detected sound volume also changes so as to follow the change in the master disk sound volume as the reproduction time elapses. On the other hand, when the reproduction time is the period T, the detected sound volume is reduced to a level where it is hardly detected.
 ここで、変動するマスク度によるサポート音量は、固定されたマスク度によるサポート音量を基準として、検出音量と原盤音量との差dV1の変化に追従して変わる。指定マスク度が入力されている場合、固定されたマスク度は指定マスク度であり、指定マスク度が入力されていない場合、固定されたマスク度は標準マスク度である。 Here, the support sound volume due to the varying masking degree changes following the change of the difference dV1 between the detected sound volume and the master disk sound volume with reference to the support sound volume due to the fixed masking degree. When the specified mask degree is input, the fixed mask degree is the specified mask degree. When the specified mask degree is not input, the fixed mask degree is the standard mask degree.
 そして、変動するマスク度で生成されたサポート音量は、歌声の音量が原盤音量よりも小さいほど、大きいレベルである。また、変動するマスク度で生成されたサポート音量は、歌声の音量が原盤音量よりも大きいほど、小さいレベルである。 Further, the support volume generated with the varying mask degree is higher as the volume of the singing voice is lower than the master volume. Further, the support volume generated with the varying mask degree is lower as the volume of the singing voice is higher than the master volume.
 すなわち、制御部41Aは、固定されたマスク度で原盤音量をマスクした音量を基準となる音量とし、また、検出音量と原盤音量との差dV1に1以下の所定値αを乗算した音量を変動分とし、これらの合計となるレベルをサポート音量とする。 That is, the control unit 41A changes the volume obtained by multiplying the difference dV1 between the detected volume and the master volume by a predetermined value α of 1 or less as a reference volume by masking the master volume with a fixed degree of masking. Minutes, and the total level of these is the support volume.
 こうして得られるサポート音量は、固定されたマスク度で原盤音量をマスクした音量よりも大きいときもあれば、小さいときもある。なお、所定値αは、指定モード、指定マスク度、および、指定期間などと共に、入力部43が入力する構成であってもよいし、標準マスク度などと同じく、データ記憶部44が予め記憶する構成であってもよい。 The support volume obtained in this way may be higher or lower than the volume obtained by masking the master volume with a fixed degree of masking. The predetermined value α may be configured to be input by the input unit 43 together with the specified mode, the specified mask degree, the specified period, etc. It may be a configuration.
 結果として、上述したように、利用者がボーカルを奏でることができない場合であっても、ユーザ端末40の生成するサポート音と、それ以外の音とによって楽曲を構成すること、ひいては、楽曲を奏でることを補助することが可能となる。 As a result, as described above, even if the user cannot play vocals, the music is composed of the support sound generated by the user terminal 40 and other sounds, and consequently the music is played. It becomes possible to assist.
 しかも、利用者がボーカルを十分に奏でることができる期間では、固定されたマスク度で生成されるサポート音量よりも十分に小さい音量でサポート音が生成される。そのため、ユーザ端末40によるサポート音の再生が、利用者によるボーカルの演奏を妨げることが抑えられる。 In addition, in a period in which the user can sufficiently play vocals, the support sound is generated with a volume sufficiently lower than the support volume generated with a fixed mask degree. Therefore, it is possible to prevent the support sound from being played by the user terminal 40 from interfering with the vocal performance by the user.
 図30が示すように、ここでもまた、再生時間が経過することに伴い、検出音量は原盤音量の変化に追従するように変化する。一方で、再生時間が期間Tであるとき、検出音量がほとんど検出されないレベルにまで下がる。 As shown in FIG. 30, the detected sound volume also changes so as to follow the change in the master disk sound volume as the reproduction time elapses. On the other hand, when the reproduction time is the period T, the detected sound volume is reduced to a level where it is hardly detected.
 ここで、変動するマスク度によるサポート音量は、検出音量が原盤音量よりも小さいときに再生され、その音量のレベルは、検出音量と原盤音量との差dV2に相当する値に設定されて、検出音量と原盤音量との差dV2の変化に追従して変わる。 Here, the support volume based on the varying masking degree is reproduced when the detected volume is smaller than the master volume, and the level of the volume is set to a value corresponding to the difference dV2 between the detected volume and the master volume. It changes following the change of the difference dV2 between the volume and the master volume.
 そして、変動するマスク度によるサポート音量は、原盤音量と同じレベルにまで、期間Tでは増える。一方、利用者がボーカルを十分に奏でることができる期間では、こうしたサポート音量が十分に抑えられる。 And the support volume due to the changing mask level increases in the period T to the same level as the master volume. On the other hand, such a support volume is sufficiently suppressed during a period in which the user can sufficiently play vocals.
 結果として、上述したように、利用者がボーカルを奏でることができない場合であっても、ユーザ端末40の生成するサポート音と、それ以外の音とによって楽曲を構成すること、ひいては、楽曲を奏でることを補助することができる。しかも、ユーザ端末40によるボーカルの再生が、利用者によるボーカルの演奏を妨げることがさらに抑えられる。 As a result, as described above, even when the user cannot play vocals, the music is composed of the support sound generated by the user terminal 40 and other sounds, and consequently the music is played. Can help. In addition, it is further suppressed that the vocal reproduction by the user terminal 40 hinders the vocal performance by the user.
 なお、第1変動モードや第2変動モードであれば、例えば、女性ボーカルと男性ボーカルとが含まれるデュエット楽曲において、これらの双方のボーカルマスクが用いられ、いずれか一方の検出音量がほとんど検出されないレベルであっても、そのパートをサポート音量で補助することもできる。 In the first variation mode and the second variation mode, for example, in a duet music including female vocals and male vocals, both vocal masks are used, and the detected sound volume of either one is hardly detected. Even if it is a level, the part can be assisted by the support volume.
 そのため、男性ボーカルのみを再生するために、女性ボーカルのみをマスクするためのボーカルマスクや、女性ボーカルのみを再生するために、男性ボーカルのみをマスクするためのボーカルマスクを、別途生成することを要しない。 Therefore, it is necessary to separately generate a vocal mask for masking only female vocals in order to reproduce only male vocals, and a vocal mask for masking only male vocals in order to reproduce only female vocals. do not do.
 以上、第2実施形態によれば以下に示す効果が得られる。
 (1)利用者がボーカルを演じられない場合であっても、伴奏音のみが空間に流れることがサポート音の生成によって抑えられる。しかも、音楽原盤に含まれるボーカルなどの対象音がサポート音として採用されるため、再生音を利用者と共に聴く共用者が、サポート音に違和感を覚えることが抑えられる。
As described above, according to the second embodiment, the following effects can be obtained.
(1) Even when the user cannot play vocals, only the accompaniment sound flows in the space can be suppressed by generating the support sound. In addition, since the target sound such as vocals included in the master music is adopted as the support sound, it is possible to prevent a common person who listens to the playback sound together with the user from feeling uncomfortable with the support sound.
 (2)固定されたマスク度で生成されるサポート音量と検出音量との乖離を小さくするように変動マスク度が生成されるため、楽曲データから再生された音と利用者の歌声とが奏でる音楽は、楽曲として調和が図られたものとなる。 (2) Since the variation mask degree is generated so as to reduce the discrepancy between the support volume and the detected volume generated with a fixed mask degree, the music played by the sound reproduced from the song data and the user's singing voice Will be harmonized as music.
 (3)検出音量とサポート音量との総量が原盤音量となるようにサポート音量が変動するため、楽曲データから再生された音と利用者の歌声とが奏でる音楽は、楽曲としてさらに調和が図られたものとなる。 (3) Since the support volume fluctuates so that the total volume of the detected volume and the support volume becomes the master volume, the music played by the sound reproduced from the song data and the user's singing voice is further harmonized as a song. It will be.
 (第3実施形態)
 図31、および、図32を参照して、サービス管理システムの第3実施形態を説明する。なお、第3実施形態は、前回までの音量の設定を履歴として取り扱う点において、第2実施形態における音楽再生装置とは異なる。以下では、第2実施形態で説明した各構成とは異なる点を主に説明し、第2実施形態で説明した構成と同じ機能を有する構成に関しては、同じ符号を付してそれの説明を省略する。
(Third embodiment)
With reference to FIGS. 31 and 32, a third embodiment of the service management system will be described. The third embodiment differs from the music playback device in the second embodiment in that the volume setting up to the previous time is handled as a history. In the following, points different from the configurations described in the second embodiment will be mainly described, and components having the same functions as those described in the second embodiment will be denoted by the same reference numerals and description thereof will be omitted. To do.
 通信部42Aは、楽曲データを用いた今回の再生において、音量設定部411が実行した音量の設定を次回以降の設定候補として管理サーバ20に送信する。制御部41Aは、楽曲データを用いた楽曲の再生が実行されるごとに、通信部42Aによる設定候補の送信を実行する。 42 A of communication parts transmit the setting of the volume which the volume setting part 411 performed in this reproduction | regeneration using music data to the management server 20 as a setting candidate after the next time. The control unit 41A performs transmission of setting candidates by the communication unit 42A each time reproduction of a song using the song data is executed.
 管理サーバ20の管理部210は、履歴管理部として機能し、ユーザ端末40から受信した設定候補を再生実績記憶部25に格納する処理を実行する。再生実績記憶部25は、再生日時、利用者コード、楽曲コード、および、設定候補を再生実績データ250に含める。 The management unit 210 of the management server 20 functions as a history management unit, and executes a process of storing the setting candidates received from the user terminal 40 in the reproduction result storage unit 25. The playback record storage unit 25 includes the playback date and time, the user code, the music code, and the setting candidate in the playback record data 250.
 再生実績データ250に含まれる設定候補は、今回の再生における指定モードを特定するための識別子に関するデータを含む。また、再生実績データ250に含まれる設定候補は、指定マスク度や標準マスク度のように、サポート音量の再生に用いられたマスク度を含む。また、再生実績データ250に含まれる設定候補は、指定期間や標準期間のようにサポート音量の再生に用いられた適用期間を含む。 The setting candidates included in the playback performance data 250 include data relating to an identifier for specifying the designated mode in the current playback. Further, the setting candidates included in the reproduction performance data 250 include the mask degree used for reproducing the support sound volume, such as the designated mask degree and the standard mask degree. In addition, the setting candidates included in the playback record data 250 include an application period used for playback of the support volume, such as a specified period and a standard period.
 管理サーバ20の制御部21は、ユーザ端末40に対してカラオケファイルを提供する際に、再生実績記憶部25が記憶する再生実績データ250を参照する。制御部21は、今回の再生に関わる利用者コードと、楽曲コードとに対応する設定候補の有無を判定する。 When the control unit 21 of the management server 20 provides a karaoke file to the user terminal 40, the control unit 21 refers to the reproduction result data 250 stored in the reproduction result storage unit 25. The control unit 21 determines whether or not there is a setting candidate corresponding to the user code related to the current reproduction and the music code.
 再生実績記憶部25に設定候補が記憶されていると判定された場合、管理サーバ20の制御部21は、今回の再生に関わる利用者コードと楽曲コードとに対応する全ての設定候補を、カラオケファイルに含めてユーザ端末40に送信する。一方、再生実績記憶部25に設定候補が記憶されていないと判定された場合、管理サーバ20の制御部21は、設定候補がない旨をカラオケファイルに含めてユーザ端末40に送信する。 When it is determined that the setting candidates are stored in the reproduction record storage unit 25, the control unit 21 of the management server 20 sets all the setting candidates corresponding to the user code and the music code related to the current reproduction as karaoke. It is included in the file and transmitted to the user terminal 40. On the other hand, when it is determined that the setting candidate is not stored in the reproduction record storage unit 25, the control unit 21 of the management server 20 transmits to the user terminal 40 that the karaoke file includes no setting candidate.
 ユーザ端末40は、管理サーバ20から受信した設定候補を解析するための履歴解析部413を備える。履歴解析部413は、管理サーバ20から受信した各設定候補を利用者による設定の履歴として取り扱い、所定のアルゴリズムに従って、利用者による設定の履歴から今回の再生における音量の設定を生成する。 The user terminal 40 includes a history analysis unit 413 for analyzing the setting candidates received from the management server 20. The history analysis unit 413 handles each setting candidate received from the management server 20 as a setting history by the user, and generates a volume setting for the current reproduction from the setting history by the user according to a predetermined algorithm.
 履歴解析部413は、例えば、採用された回数が最も多い設定を複数の設定候補の中から抽出し、抽出された設定を今回の設定とする。履歴解析部413は、例えば、採用された回数が最も多い指定モードを設定候補の中から抽出し、抽出された指定モードを今回の指定モードとして設定する。 The history analysis unit 413, for example, extracts the setting that has been adopted most frequently from a plurality of setting candidates, and sets the extracted setting as the current setting. For example, the history analysis unit 413 extracts the designated mode that has been adopted most frequently from the setting candidates, and sets the extracted designated mode as the current designated mode.
 また、履歴解析部413は、例えば、再生日時が最も新しい設定候補から順に、連続している指定モードの有無を判断し、連続している最新の指定モードを今回の指定モードとして設定する。また、履歴解析部413は、例えば、設定候補に含まれるマスク度を用い、それらの平均値や最頻値を代表値として算出し、算出された代表値を今回の指定マスク度として設定する。 Further, for example, the history analysis unit 413 determines whether or not there is a continuous designation mode in order from the setting candidate with the latest reproduction date and time, and sets the latest continuous designation mode as the current designation mode. Further, the history analysis unit 413 uses, for example, the mask degrees included in the setting candidates, calculates the average value or mode value thereof as a representative value, and sets the calculated representative value as the current designated mask degree.
 図32が示すように、まず、制御部41Aは、提供されたカラオケファイルから楽曲データ、および、ボーカルマスクを読み込む(ステップS13-1)。この際、制御部41Aは、カラオケファイルに設定候補が含まれているか否かを判定する(ステップS13-2)。 As shown in FIG. 32, first, the control unit 41A reads music data and a vocal mask from the provided karaoke file (step S13-1). At this time, the control unit 41A determines whether or not setting candidates are included in the karaoke file (step S13-2).
 次に、カラオケファイルに設定候補が含まれていない場合(ステップS13-2において「NO」の場合)、音量設定部411は、再生信号生成部412が採用する指定モードに固定モードを設定する。また、音量設定部411は、再生信号生成部412が採用するマスク度に標準マスク度を設定し、再生信号生成部412が採用する適用期間に標準期間を設定する(ステップS13-3)。 Next, when the setting candidate is not included in the karaoke file (in the case of “NO” in step S13-2), the volume setting unit 411 sets the fixed mode as the designated mode adopted by the reproduction signal generating unit 412. Further, the volume setting unit 411 sets the standard mask degree as the mask degree adopted by the reproduction signal generation unit 412 and sets the standard period as the application period adopted by the reproduction signal generation unit 412 (step S13-3).
 一方、カラオケファイルに設定候補が含まれている場合(ステップS13-2において「YES」の場合)、ユーザ端末40の履歴解析部413は、今回の楽曲再生に採用する指定モード、マスク度、および、適用期間を、読み出された設定候補から生成する(ステップS14-3)。 On the other hand, if the karaoke file contains setting candidates (in the case of “YES” in step S13-2), the history analysis unit 413 of the user terminal 40 uses the designated mode, mask degree, and The application period is generated from the read setting candidates (step S14-3).
 また、音量設定部411は、履歴解析部413による解析の結果に従って、再生信号生成部412が採用する指定モード、マスク度、および、指定期間を仮設定する。また、音量設定部411は、仮設定された内容を表示部などに出力させることによって、今回の再生における設定の確認を利用者に促す(ステップS14-4)。 Also, the volume setting unit 411 temporarily sets the designation mode, the masking degree, and the designation period adopted by the reproduction signal generation unit 412 according to the analysis result by the history analysis unit 413. Further, the volume setting unit 411 prompts the user to confirm the setting in the current reproduction by outputting the temporarily set content to the display unit or the like (step S14-4).
 この際、音量設定部411は、ユーザ端末40の入力部43に対する外部からの操作によって、指定モード、マスク度、および、指定期間の変更を許可する。そして、音量設定部411は、再生信号生成部412が採用する指定モード、マスク度、および、適用期間に、確認された内容を設定する(ステップS13-3)。 At this time, the volume setting unit 411 permits the change of the designated mode, the degree of masking, and the designated period by an external operation on the input unit 43 of the user terminal 40. Then, the volume setting unit 411 sets the confirmed contents in the designation mode, the masking degree, and the application period adopted by the reproduction signal generation unit 412 (step S13-3).
 以降、制御部41Aは、第2実施形態と同様に、音量設定部411による音量の設定に従って、カラオケファイルに含まれる楽曲データから楽曲を再生する。そして、制御部41Aは、今回の再生における音量の設定を次回の設定候補として管理サーバ20に送信する(ステップS14-5)。 Thereafter, similarly to the second embodiment, the control unit 41A reproduces the music from the music data included in the karaoke file according to the volume setting by the volume setting unit 411. Then, the control unit 41A transmits the volume setting for the current reproduction to the management server 20 as the next setting candidate (step S14-5).
 以上、第3実施形態によれば、以下に列挙する効果が得られる。
 (1)サポート音の音量として今回の再生に求められる大きさやその適用期間は、再生する楽曲が同じであれば、前回の再生までに行われた設定のいずれかと同じ大きさや同じ適用期間であることが少なくない。この点、前回の再生までに行われた音量の設定を用いて、音量設定部411が今回の音量の設定を行うため、新たな設定を入力する利用者の負荷が軽減される。
As described above, according to the third embodiment, the effects listed below can be obtained.
(1) The volume required for the current playback as the volume of the support sound and the application period thereof are the same volume and the same application period as any of the settings made up to the previous playback if the music to be played is the same. There are many things. In this regard, since the volume setting unit 411 sets the current volume using the volume setting performed up to the previous playback, the load on the user who inputs a new setting is reduced.
 なお、上記第2実施形態、および、第3実施形態は、以下の態様に変更してもよい。
 ・ユーザ端末40は、ネットワークを介して管理サーバ20に接続された端末に限らず、例えば、ネットワークに接続されていない装置であってもよい。この際、ユーザ端末40は、管理サーバ20が備えるボーカルマスクの生成機能を備えてもよい。すなわち、ユーザ端末は、管理サーバ20が備えるボーカル処理部211に相当する機能部を備え、原盤ファイルに含まれる楽曲データからボーカルの音域を特定し、ボーカルの音量を変えるためのボーカルマスクの生成処理を実行してもよい。
In addition, you may change the said 2nd Embodiment and 3rd Embodiment into the following aspects.
The user terminal 40 is not limited to a terminal connected to the management server 20 via a network, and may be a device that is not connected to the network, for example. At this time, the user terminal 40 may include a vocal mask generation function included in the management server 20. That is, the user terminal includes a functional unit corresponding to the vocal processing unit 211 included in the management server 20, and specifies a vocal range from the music data included in the master file and generates a vocal mask for changing the vocal volume. May be executed.
 ・ユーザ端末40が行う楽曲の再生は、楽曲記憶部24に記憶された楽曲データや、ボーカルマスクを用いる処理に限らない。ユーザ端末40は、例えば、混合音を再生するための楽曲用のストリーミングファイルと、混合音のなかの対象音をマスクするためのマスク用のストリーミングファイルとを管理サーバ20から取得し、これら取得したファイルを用いて楽曲をストリーミング再生してもよい。なお、第1実施形態におけるユーザ端末40も、こうしたストリーミング再生を行うための同様の機能を備えてもよい。 The reproduction of the music performed by the user terminal 40 is not limited to the process using the music data stored in the music storage unit 24 or the vocal mask. The user terminal 40 acquires, for example, a streaming file for music for reproducing the mixed sound and a streaming file for masking for masking the target sound in the mixed sound from the management server 20, and acquires these. The music may be streamed using the file. Note that the user terminal 40 in the first embodiment may also have a similar function for performing such streaming playback.
 ・調整対象音は、例えば、オーケストラを構成する楽器であって各々が相互に異なる二つ以上の楽器によって奏でられる音であってもよい。このように、各々が共通するタイムインデックスに対応付けられた2つ以上の音域に、対象音が別々に含まれてもよい。この際、ユーザ端末40は、各々が相互に異なる音域をマスクするための複数のボーカルマスクと、各ボーカルマスクに1つずつのマスク度を用いてもよい。あるいは、ユーザ端末40は、相互に異なる音域をマスクするように定義された1つのボーカルマスクと、各音域に1つずつのマスク度を用いてもよい。このように、2つ以上の音域に別々の調整対象音が含まれる構成であれば、各々が相互に異なる音階を有して外部で奏でられる複数の音と、それ以外の再生音とによって生成される混合音に関して、それを奏でることの補助を行うことが可能である。 The adjustment target sound may be, for example, a sound that is played by two or more musical instruments that constitute an orchestra and are different from each other. In this way, the target sound may be included separately in two or more sound ranges associated with the common time index. At this time, the user terminal 40 may use a plurality of vocal masks for masking different sound ranges, and a masking degree for each vocal mask. Alternatively, the user terminal 40 may use one vocal mask defined so as to mask different sound ranges and one mask degree for each sound range. In this way, if two or more sound ranges include different adjustment target sounds, they are generated from a plurality of sounds that are played outside each having a different scale and other reproduced sounds. It is possible to assist in playing the mixed sound.
 ・各々が相互に異なる音域をマスクする複数のボーカルマスクが、1つの楽曲の再生に際して採用されてもよい。この際、各ボーカルマスクと、それが採用される適用期間とは、入力部43に入力される操作などに基づいて、ユーザ端末40においてボーカルマスクに1つずつ対応付けられる。 A plurality of vocal masks, each masking a different range, may be adopted when playing one piece of music. At this time, each vocal mask and the application period in which it is adopted are associated one by one with the vocal mask in the user terminal 40 based on an operation input to the input unit 43 or the like.
 こうしたボーカルマスクの採用の形態であれば、例えば、ボーカルマスクによるマスクの精度が高いことが求められる適用期間には、第3ボーカルマスクのような動的なボーカルマスクが用いられる。また、ボーカルマスクによるマスクの精度が求められない適用期間には、第1ボーカルマスクのような静的ボーカルマスクが用いられる。それによって、第3ボーカルマスクのような動的なボーカルマスクのみが用いられる構成と比べて、ボーカルマスクの記憶に求められる記憶容量を抑えることが可能であり、また、ボーカルマスクの生成に要する手間を軽減することも可能である。 In the case of adopting such a vocal mask, for example, a dynamic vocal mask such as the third vocal mask is used in an application period in which high accuracy of the mask by the vocal mask is required. Further, a static vocal mask such as the first vocal mask is used in an application period in which the accuracy of the mask by the vocal mask is not required. As a result, it is possible to reduce the storage capacity required for storing the vocal mask as compared with a configuration in which only a dynamic vocal mask such as the third vocal mask is used, and the effort required to generate the vocal mask. Can be reduced.
 ・データ記憶部44が記憶した楽曲データから所定のアルゴリズムに従ってボーカルマスクを生成すると共に、生成されたボーカルマスクをデータ記憶部44に記憶させるデータ処理部を、ユーザ端末40がさらに備えてもよい。この際、データ処理部は、例えば、CPU、ROM、RAMを備え、ボーカルマスクを生成するためのアルゴリズムを記憶すると共に、入力部43から入力された楽曲データから所定のアルゴリズムに従ってボーカルマスクを生成する。再生処理部は、データ処理部の処理の結果を受けて、データ記憶部44から楽曲データとボーカルマスクとを読み出し、ボーカルマスクとマスク度とを用いて楽曲データから楽曲を再生する。 The user terminal 40 may further include a data processing unit that generates a vocal mask from the music data stored in the data storage unit 44 according to a predetermined algorithm and causes the data storage unit 44 to store the generated vocal mask. At this time, the data processing unit includes, for example, a CPU, a ROM, and a RAM, stores an algorithm for generating a vocal mask, and generates a vocal mask from music data input from the input unit 43 according to a predetermined algorithm. . The reproduction processing unit receives the processing result of the data processing unit, reads out the music data and the vocal mask from the data storage unit 44, and reproduces the music from the music data using the vocal mask and the mask degree.
 ・適用期間は、Aメロ、Bメロ、サビ、チャプターに限らず、例えば、ボーカルの音域のなかで所定の音階よりも高い音域である期間や、ボーカルの音域のなかで所定の音階よりも低い音域である期間として予め定められてもよい。こうした期間の設定は、まず、適用期間を定めるためのボーカルの音域を定め、ボーカルの音階がその音域に含まれるタイムインデックスを適用期間とする。 -The application period is not limited to A melody, B melody, rust, and chapter, for example, a period that is higher than a predetermined scale in a vocal range or a lower range than a predetermined scale in a vocal range It may be determined in advance as a period that is a sound range. In setting such a period, first, a vocal range for determining the application period is determined, and a time index in which the vocal scale is included in the range is set as the application period.
 人声の音域は、例えば、高音域や低音域に分けられ、高音域に声域を有しない利用者は、高音域にサポート音を必要とされ、低音域に声域を有しない利用者は、低音域にサポート音を必要とされる。この点、適用期間がボーカルの音域によって定められる構成であれば、サポート音が本来必要とされる適用期間のみでサポート音の生成が実行されるため、サポート音をマスク度でマスクする音量の制御において、それの負荷を軽減することが可能ともなる。なお、こうした適用期間において第1変動モードや第2変動モードによる補助設定が行われる構成であれば、サポート音が本来必要とされる適用期間でも、さらにサポート音量として好ましい音量で補助を行うことが可能でもある。 The human voice range is divided into, for example, a high range and a low range.A user who does not have a voice range in the high range requires a support sound in the high range, and a user who does not have a voice range in the low range has a low range. Support sound is required in the range. In this regard, if the application period is determined by the vocal range, the support sound is generated only during the application period in which the support sound is originally required. Therefore, it is possible to reduce the load. In addition, if the auxiliary setting is performed in the first variation mode and the second variation mode in such an application period, assistance can be performed at a volume that is preferable as a support volume even in an application period in which the support sound is originally required. It is also possible.
 ・図33が示すように、ユーザ端末40は、演奏者がマイクに入力する音から、その演奏者が奏でる音の音域を特定する音域特定部47をさらに備えてもよい。音域特定部47が行う音域を特定するための処理は、楽曲を再生する前に行われてもよいし、楽曲の再生中に逐次行われてもよい。 As shown in FIG. 33, the user terminal 40 may further include a sound range specifying unit 47 that specifies the sound range of the sound played by the performer from the sound input by the performer to the microphone. The processing for specifying the sound range performed by the sound range specifying unit 47 may be performed before the music is reproduced, or may be sequentially performed during the reproduction of the music.
 例えば、音域特定部47は、演奏者が奏でる音の音域を予め特定し、音量設定部411は、指定モードに基づく補助を、予め特定された音域以外の音域で行うためのモード特定部414を備えてもよい。 For example, the range specifying unit 47 specifies the range of the sound played by the performer in advance, and the volume setting unit 411 includes the mode specifying unit 414 for performing assistance based on the designated mode in a range other than the range specified in advance. You may prepare.
 モード特定部414は、例えば、調整対象音が対応付けられるタイムインデックスのなかで、音域特定部47が特定した音域以外の音域が対応付けられるタイムインデックスを、楽曲データに基づき特定する。そして、モード特定部414は、指定マスク度や標準マスク度を採用する期間を、その特定されたタイムインデックスに設定する。こうした再生によれば、調整対象音がサポート音量で再生される期間を、演奏者の演奏できない音域に特化させたものとすることが可能でもある。 The mode specifying unit 414 specifies, for example, a time index associated with a sound range other than the sound range specified by the sound range specifying unit 47 in the time index associated with the adjustment target sound based on the music data. And the mode specific | specification part 414 sets the period which employ | adopts designated mask degree and standard mask degree to the specified time index. According to such reproduction, it is possible to specialize the period during which the adjustment target sound is reproduced at the support sound volume in a range where the performer cannot perform.
 例えば、音域特定部47は、演奏者が奏でる音の音域を演奏中に逐次特定し、音量設定部411は、指定モードに基づく補助を、逐次特定された音域に基づいて行うためのモード特定部414を備えてもよい。このモード特定部414は、例えば、音域特定部47が逐次特定する演奏者の音域と、楽曲ファイルに基づく対象音の音域との差が、所定範囲内であるか否かの判断を、楽曲の再生中に繰り返す。そして、音域特定部47が特定する演奏者の音域と、楽曲ファイルに基づく対象音の音域との差が、所定範囲外であるとき、モード特定部414は、そのときの指定モードに基づくマスク度を設定する。 For example, the sound range specifying unit 47 sequentially specifies the sound range of the sound played by the performer during performance, and the volume setting unit 411 is a mode specifying unit for performing assistance based on the designated mode based on the sequentially specified sound range. 414 may be provided. For example, the mode specifying unit 414 determines whether or not the difference between the player's range specified by the range specifying unit 47 and the range of the target sound based on the song file is within a predetermined range. Repeat during playback. When the difference between the player's range specified by the range specification unit 47 and the range of the target sound based on the music file is outside the predetermined range, the mode specification unit 414 determines the mask degree based on the designated mode at that time. Set.
 なお、音域特定部47が特定する演奏者の音域と、楽曲ファイルに基づく調整対象音の音域との差の許容される範囲である上記所定範囲は、指定マスク度などと同じく、入力部43から入力される構成であってもよいし、データ記憶部44が予め記憶する構成であってもよい。こうした再生によれば、演奏者が奏でる音の音域と、調整対象音の音域との乖離が大きいときに、サポート音で調整対象音が再生される。そのため、演奏者の奏でる音の音域が、その楽曲に本来求められる音域から外れていることを、演奏者に把握させることが可能となる。また、その楽曲に本来求められる音域で対象音を再生するため、これらの混合音を鑑賞する鑑賞者に対して、上記乖離による違和感を抑えることが可能ともなる。 Note that the predetermined range, which is an allowable range of the difference between the performer's range specified by the range specification unit 47 and the range of the adjustment target sound based on the music file, is the same as the specified mask degree and the like from the input unit 43. The structure which is input may be sufficient and the structure which the data storage part 44 memorize | stores beforehand may be sufficient. According to such reproduction, the adjustment target sound is reproduced with the support sound when the difference between the range of the sound played by the performer and the range of the adjustment target sound is large. Therefore, it becomes possible for the performer to grasp that the sound range of the sound played by the performer is out of the sound range originally required for the music. In addition, since the target sound is reproduced in a sound range originally required for the music, it is possible to suppress a sense of incongruity due to the divergence for a viewer who appreciates the mixed sound.
 例えば、音域特定部47は、演奏者が奏でる音の音域を予め特定し、制御部41Aは、通信部42Aの通信を介して、特定された演奏者の音域を制御部21に送信してもよい。この際、制御部21は、各楽曲管理データ240のなかで音域に関するデータを参照し、音域に関するデータが示す音域が演奏者の音域の含まれる楽曲管理データ240を特定する。また、制御部21は、その特定された楽曲管理データ240を、推奨曲の楽曲管理データ240として通信部42Aに送信する。そして、制御部41Aは、通信部42Aが受信した推奨曲の楽曲管理データ240に基づき、推奨曲のリストを表示部などに出力させてもよい。こうした楽曲の推奨によれば、演奏者が奏でられる音域に適した楽曲を、演奏者に推奨することが可能ともなる。 For example, the range specifying unit 47 specifies the range of the sound played by the performer in advance, and the control unit 41A transmits the specified performer's range to the control unit 21 via the communication of the communication unit 42A. Good. At this time, the control unit 21 refers to the data related to the sound range in each piece of music management data 240, and specifies the music management data 240 in which the sound range indicated by the data related to the sound range includes the player's sound range. In addition, the control unit 21 transmits the specified music management data 240 to the communication unit 42A as music management data 240 of the recommended music. Then, the control unit 41A may cause the display unit to output a list of recommended songs based on the song management data 240 of the recommended songs received by the communication unit 42A. According to the recommendation of such music, it is possible to recommend to the performer a music suitable for the sound range played by the performer.
 なお、制御部41Aは、演奏者が奏でる音の音域に加え、さらに、通信部42Aの通信を介して、演奏者が求める情景などに関する情報を、制御部21に送信してもよい。そして、制御部21は、各楽曲管理データ240のなかで情景に関するデータを参照し、演奏者が求める情景に含まれる情景が記憶された楽曲管理データ240を特定する。また、制御部21は、その特定された楽曲管理データ240を、推奨曲の楽曲管理データ240として通信部42Aに送信する。こうした楽曲の推奨によれば、演奏者が奏でられる音域に適した楽曲であり、かつ、演奏者が求める情景に適した楽曲を、演奏者に推奨することが可能ともなる。 Note that the control unit 41A may transmit to the control unit 21 information related to a scene requested by the performer via communication of the communication unit 42A in addition to the range of the sound played by the performer. And the control part 21 specifies the music management data 240 in which the scene contained in the scene which a player requires is referred with reference to the data regarding a scene in each music management data 240. FIG. In addition, the control unit 21 transmits the specified music management data 240 to the communication unit 42A as music management data 240 of the recommended music. According to the recommendation of such music, it is possible to recommend to the performer a music that is suitable for the sound range played by the performer and suitable for the scene that the performer desires.
 ・再生判定処理が実行する再生判定方法は、再生判定時間による方法に限定されるものではない。例えば、楽曲におけるサビの位置に基づいて、再生判定を行なうようにしてもよい。この場合には、管理サーバ20の制御部21が、主旋律の推定処理(ステップS1-3)において、特定のパターンに基づいてサビ位置を特定する。そして、楽曲管理データ240に、特定したサビ位置を記憶しておく。そして、再生管理部214は、タイムインデックスとサビ位置とに基づいて、再生判定を行なう。ここでは、再生判定時間を経過していない場合であっても、サビが再生された場合には、再生と判定する。 · The playback determination method executed by the playback determination process is not limited to the method based on the playback determination time. For example, the reproduction determination may be performed based on the position of rust in the music. In this case, the control unit 21 of the management server 20 specifies the chorus position based on the specific pattern in the main melody estimation process (step S1-3). The specified chorus position is stored in the music management data 240. Then, the playback management unit 214 performs playback determination based on the time index and the chorus position. Here, even if the playback determination time has not elapsed, if the rust is played back, it is determined to be played back.
 ・設定されたマスク度に応じて、課金方法を変更するようにしてもよい。この場合には、マスクファイルのマスク度に応じて出力されるボーカルの音量に基づいて、分配金を算出するためルールを分配管理ファイル270に記憶させておく。 · The billing method may be changed according to the set mask degree. In this case, a rule is stored in the distribution management file 270 to calculate the distribution money based on the volume of vocals output according to the masking degree of the mask file.
 10…コンテンツサーバ、20…管理サーバ、21,41,41A…制御部、22…原盤記憶部、23…歌詞情報記憶部、24…楽曲記憶部、25…再生実績記憶部、30…担当者端末、40,40a,40b,40c…ユーザ端末、41…制御部、42…通信部、43…入力部、44…データ記憶部、45…音量検出部、46…楽曲出力部、211…管理部、212…ボーカル処理部、213…歌詞管理部、214…エディタ、215…再生管理部、220…原盤ファイル、230…歌詞ファイル、240…楽曲管理データ、241…HLSファイル、242…XMLファイル、243…ボーカルマスク、244…楽曲管理レコード、250…再生実績データ、251…利用者実績レコード、252…楽曲実績レコード、26…原資記憶部、411…音量設定部、412…再生信号生成部、42A…通信部、500,510,520,600,601,602,603…音階推移グラフ、501…第1ボーカルマスク、511…第2ボーカルマスク、521…第3ボーカルマスク、700…XMLファイル。 DESCRIPTION OF SYMBOLS 10 ... Content server, 20 ... Management server, 21, 41, 41A ... Control part, 22 ... Master disk storage part, 23 ... Lyric information storage part, 24 ... Music storage part, 25 ... Reproduction record storage part, 30 ... Person in charge terminal 40, 40a, 40b, 40c ... user terminal, 41 ... control unit, 42 ... communication unit, 43 ... input unit, 44 ... data storage unit, 45 ... volume detection unit, 46 ... music output unit, 211 ... management unit, 212 ... Vocal processing unit, 213 ... Lyric management unit, 214 ... Editor, 215 ... Playback management unit, 220 ... Master file, 230 ... Lyric file, 240 ... Music management data, 241 ... HLS file, 242 ... XML file, 243 ... Vocal mask, 244 ... music management record, 250 ... reproduction performance data, 251 ... user performance record, 252 ... music performance record, 26 ... stock memory 411: Volume setting unit, 412: Reproduction signal generation unit, 42A: Communication unit, 500, 510, 520, 600, 601, 602, 603 ... Scale transition graph, 501: First vocal mask, 511: Second vocal mask 521 ... Third vocal mask, 700 ... XML file.

Claims (9)

  1.  音量調整の対象となる調整対象音とそれ以外の音とを含む音が混合音であり、前記混合音を再生するためのデータを含む原盤ファイルを記憶した原盤記憶部と、
     音調整コンテンツをユーザ端末に提供するサービスの原資を記憶した原資記憶部と、
     前記ユーザ端末による音調整コンテンツの再生回数を記憶する再生実績記憶部と、
     前記ユーザ端末に接続される制御部とを備え、
     前記制御部が、
     前記原盤記憶部に記憶された前記原盤ファイルにおいて前記調整対象音の音域を特定し、
     前記特定された音域に基づき、前記混合音のなかで前記調整対象音の音量を個別に調整するためのマスクファイルを生成し、
     前記マスクファイルを含む前記音調整コンテンツを前記ユーザ端末に提供し、前記再生回数を前記再生実績記憶部に記憶し、
     前記原資記憶部に記憶した原資と、前記再生実績記憶部に記憶された前記再生回数とに基づいて、前記調整対象音の分配対象者に分配する分配金額を算出する
     サービス管理システム。
    A master disk storage unit that stores a master disk file that includes data for reproducing the mixed sound, and the sound including the adjustment target sound that is the target of volume adjustment and the other sound is a mixed sound;
    A resource storage unit that stores resources for providing sound adjustment content to the user terminal;
    A playback record storage unit for storing the number of playback times of the sound adjustment content by the user terminal;
    A control unit connected to the user terminal,
    The control unit is
    In the master file stored in the master storage unit, the range of the adjustment target sound is specified,
    Based on the specified range, generate a mask file for individually adjusting the volume of the adjustment target sound in the mixed sound,
    Providing the sound adjustment content including the mask file to the user terminal, storing the number of times of reproduction in the reproduction result storage unit,
    A service management system that calculates a distribution amount to be distributed to the distribution target person of the adjustment target sound based on the resource stored in the resource storage unit and the number of reproductions stored in the reproduction result storage unit.
  2.  前記制御部は、
     複数の前記音調整コンテンツを提供し、
     前記制御部が提供した前記音調整コンテンツごとの前記再生回数の合計である総再生回数を算出し、
     前記総再生回数に対する前記各再生回数の割合に基づいて、前記各音調整コンテンツの前記分配金額を算出する
     請求項1に記載のサービス管理システム。
    The controller is
    Providing a plurality of the sound adjustment contents;
    Calculating a total number of times of reproduction that is the total number of times of reproduction for each of the sound adjustment contents provided by the control unit;
    The service management system according to claim 1, wherein the distribution amount of each sound adjustment content is calculated based on a ratio of each reproduction count to the total reproduction count.
  3.  前記制御部が、
     複数の前記ユーザ端末に前記音調整コンテンツを提供し、
     前記各ユーザ端末に提供される前記音調整コンテンツが相互に等しく、かつ、前記各ユーザ端末が前記音調整コンテンツの再生を相互に同期させた場合には、前記音調整コンテンツの前記再生回数を調整する
     請求項1または2に記載のサービス管理システム。
    The control unit is
    Providing the sound adjustment content to a plurality of the user terminals;
    When the sound adjustment contents provided to the user terminals are equal to each other and the user terminals synchronize the reproduction of the sound adjustment contents, the number of reproductions of the sound adjustment contents is adjusted. The service management system according to claim 1 or 2.
  4.  前記制御部が、
     各原盤音源の一部から構成されたメドレーが前記混合音であり、前記メドレーを構成する原盤音源を特定し、
     前記原盤音源の前記音調整コンテンツに前記再生回数を分配する
     請求項1から3のいずれか一項に記載のサービス管理システム。
    The control unit is
    A medley composed of a part of each master sound source is the mixed sound, and specifies a master sound source constituting the medley,
    The service management system according to any one of claims 1 to 3, wherein the reproduction count is distributed to the sound adjustment content of the master sound source.
  5.  前記ユーザ端末をさらに備え、
     前記ユーザ端末は、
     音量の設定を行う音量設定部と、
     前記調整対象音の音量が前記音量設定部の設定に追従して変わるように、前記音量設定部による設定に基づいて、前記混合音を再生するための楽曲データと前記マスクファイルとを用いて前記混合音を再生する再生処理部とを備える
     請求項1から4のいずれか一項に記載のサービス管理システム。
    Further comprising the user terminal;
    The user terminal is
    A volume setting section for setting the volume;
    Based on the setting by the volume setting unit, the music data for playing the mixed sound and the mask file are used so that the volume of the adjustment target sound changes following the setting of the volume setting unit. The service management system according to claim 1, further comprising: a reproduction processing unit that reproduces the mixed sound.
  6.  前記音量設定部は、前記音量設定部の設定を前記音調整コンテンツの再生ごとに設定候補として記憶すると共に、前記記憶された設定候補を用いて今回の再生における設定を行う
     請求項5に記載のサービス管理システム。
    The said volume setting part memorize | stores the setting of the said volume setting part as a setting candidate for every reproduction | regeneration of the said sound adjustment content, and performs the setting in this reproduction | regeneration using the said stored setting candidate. Service management system.
  7.  前記音量の設定に関する履歴を利用者ごとに管理して前記履歴を前記ユーザ端末へ送信する履歴管理部をさらに備え、
     前記ユーザ端末は、
     今回のユーザに対して管理された前回の利用までの前記履歴を前記履歴管理部から受信すると共に、今回行われた設定に関する情報を前記履歴管理部へ送信する通信部と、
     前記通信部が受信した前記履歴を用いて今回の設定を行う前記音量設定部と、
     前記調整対象音の音量が前記今回の設定に追従して変わるように、今回の設定に基づいて前記混合音を再生する再生処理部とを備える
     請求項6に記載のサービス管理システム。
    A history management unit that manages the history of the volume setting for each user and transmits the history to the user terminal;
    The user terminal is
    A communication unit that receives the history up to the previous use managed for the user from the history management unit, and transmits information related to the setting performed this time to the history management unit;
    The volume setting unit that performs the current setting using the history received by the communication unit;
    The service management system according to claim 6, further comprising: a reproduction processing unit that reproduces the mixed sound based on the current setting so that the volume of the adjustment target sound changes following the current setting.
  8.  音量調整の対象となる調整対象音とそれ以外の音とを含む音が混合音であり、前記混合音を再生するためのデータを含む原盤ファイルを記憶した原盤記憶部と、
     音調整コンテンツをユーザ端末に提供するサービスの原資を記憶した原資記憶部と、
     前記ユーザ端末による音調整コンテンツの再生回数を記憶する再生実績記憶部と、
     前記ユーザ端末に接続される制御部とを備えたサービス管理システムを用いて、サービス管理を行なうプログラムが記憶されたコンピュータ可読記憶媒体であって、
     前記プログラムの実行時、前記制御部は、
     前記原盤記憶部に記憶された前記原盤ファイルにおいて前記調整対象音の音域を特定し、
     前記特定された音域に基づき、前記混合音のなかで前記調整対象音の音量を個別に調整するためのマスクファイルを生成し、
     前記マスクファイルを含む前記音調整コンテンツを前記ユーザ端末に提供し、前記再生回数を前記再生実績記憶部に記憶し、
     前記原資記憶部に記憶された原資と、前記再生実績記憶部に記憶された前記再生回数とに基づいて、前記調整対象音の分配対象者に分配する分配金額を算出する
    ように機能するコンピュータ可読記憶媒体。
    A master disk storage unit that stores a master disk file that includes data for reproducing the mixed sound, and the sound including the adjustment target sound that is the target of volume adjustment and the other sound is a mixed sound;
    A resource storage unit that stores resources for providing sound adjustment content to the user terminal;
    A playback record storage unit for storing the number of playback times of the sound adjustment content by the user terminal;
    A computer-readable storage medium storing a program for performing service management using a service management system including a control unit connected to the user terminal,
    During execution of the program, the control unit
    In the master file stored in the master storage unit, the range of the adjustment target sound is specified,
    Based on the specified range, generate a mask file for individually adjusting the volume of the adjustment target sound in the mixed sound,
    Providing the sound adjustment content including the mask file to the user terminal, storing the number of times of reproduction in the reproduction result storage unit,
    A computer readable computer that functions to calculate a distribution amount to be distributed to the distribution target person of the adjustment target sound based on the resource stored in the resource storage unit and the number of reproductions stored in the reproduction result storage unit. Storage medium.
  9.  音量調整の対象となる調整対象音とそれ以外の音とを含む音が混合音であり、前記混合音を再生するためのデータを含む原盤ファイルを記憶した原盤記憶部と、
     音調整コンテンツをユーザ端末に提供するサービスの原資を記憶した原資記憶部と、
     前記ユーザ端末による音調整コンテンツの再生回数を記憶する再生実績記憶部と、
     前記ユーザ端末に接続される制御部とを備えたサービス管理システムを用いて、サービス管理を行なう方法であって、
     前記制御部が、
     前記原盤記憶部に記憶された前記原盤ファイルにおいて前記調整対象音の音域を特定し、
     前記特定された音域に基づき、前記混合音のなかで前記調整対象音の音量を個別に調整するためのマスクファイルを生成し、
     前記マスクファイルを含む前記音調整コンテンツを前記ユーザ端末に提供し、前記再生回数を前記再生実績記憶部に記憶し、
     前記原資記憶部に記憶された原資と、前記再生実績記憶部に記憶された前記再生回数とに基づいて、前記調整対象音の分配対象者に分配する分配金額を算出する
     サービス管理方法。
    A master disk storage unit that stores a master disk file that includes data for reproducing the mixed sound, and the sound including the adjustment target sound that is the target of volume adjustment and the other sound is a mixed sound;
    A resource storage unit that stores resources for providing sound adjustment content to the user terminal;
    A playback record storage unit for storing the number of playback times of the sound adjustment content by the user terminal;
    A service management system using a service management system including a control unit connected to the user terminal,
    The control unit is
    In the master file stored in the master storage unit, the range of the adjustment target sound is specified,
    Based on the specified range, generate a mask file for individually adjusting the volume of the adjustment target sound in the mixed sound,
    Providing the sound adjustment content including the mask file to the user terminal, storing the number of times of reproduction in the reproduction result storage unit,
    A service management method for calculating a distribution amount to be distributed to the distribution target person of the adjustment target sound based on the resource stored in the resource storage unit and the number of reproductions stored in the reproduction result storage unit.
PCT/JP2016/067122 2015-12-08 2016-06-08 Service management system, computer readable storage medium, and service management method WO2017098743A1 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP2015-239691 2015-12-08
JP2015239691A JP2016110154A (en) 2014-12-09 2015-12-08 Service management system, service management method, and service management program
JP2016-035617 2016-02-26
JP2016035617 2016-02-26

Publications (1)

Publication Number Publication Date
WO2017098743A1 true WO2017098743A1 (en) 2017-06-15

Family

ID=59012969

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2016/067122 WO2017098743A1 (en) 2015-12-08 2016-06-08 Service management system, computer readable storage medium, and service management method

Country Status (2)

Country Link
JP (1) JP2017156749A (en)
WO (1) WO2017098743A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002108357A (en) * 2000-09-29 2002-04-10 Casio Comput Co Ltd Downloading system, information processor, and recording medium
JP2005077485A (en) * 2003-08-28 2005-03-24 National Institute Of Advanced Industrial & Technology Duet and chorus karaoke control system at multiple places
JP2009266310A (en) * 2008-04-25 2009-11-12 Kenwood Corp Recording device and recording method

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002108357A (en) * 2000-09-29 2002-04-10 Casio Comput Co Ltd Downloading system, information processor, and recording medium
JP2005077485A (en) * 2003-08-28 2005-03-24 National Institute Of Advanced Industrial & Technology Duet and chorus karaoke control system at multiple places
JP2009266310A (en) * 2008-04-25 2009-11-12 Kenwood Corp Recording device and recording method

Also Published As

Publication number Publication date
JP2017156749A (en) 2017-09-07

Similar Documents

Publication Publication Date Title
US8513512B2 (en) Personalized music remixing
US8855334B1 (en) Mixed content for a communications device
JP5331494B2 (en) Karaoke service system, terminal device
US9075760B2 (en) Narration settings distribution for content customization
US7663048B2 (en) Content distributing server, content distributing method, and content distributing program
WO2021216115A1 (en) System, method, and non-transitory computer-readable storage medium for collaborating on a musical composition over a communication network
CN101111884A (en) Methods and apparatus for use in sound modification
KR20190054282A (en) Method and system for cooperating production of online sound source
JP5772054B2 (en) Singing evaluation device
WO2017098743A1 (en) Service management system, computer readable storage medium, and service management method
WO2003073414A1 (en) Music providing method and system, and music creation system
KR102568089B1 (en) System for providing collaboration based music contents producing service
JP2016110154A (en) Service management system, service management method, and service management program
KR20080082019A (en) Method and system for original sound noraebang service
WO2012004650A1 (en) Systems and methods for dynamic, distributed creation of a musical composition to accompany a visual composition
CN2909452Y (en) Electronic musical instrument for playback received musice
WO2024075638A1 (en) Acoustic model training method
WO2024075633A1 (en) Training system and method for acoustic model
JP2002297136A (en) Musical piece generating device, music distribution system, and program
US20130166478A1 (en) Content Management Apparatus
WO2024075634A1 (en) Display method related to characteristic distribution of sound waveform
Bell Creating Commercial Music: Advertising* Library Music* TV Themes* and More
WO2023276279A1 (en) Image processing device, image processing method, and program
JP2004295379A (en) Data providing system, data providing method, and data providing program
KR102592818B1 (en) System for creating digital contents by tuning selectively expansion and combination of sound sources

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16872643

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 16872643

Country of ref document: EP

Kind code of ref document: A1