US20120269357A1 - Audio signal processing system for live music performance - Google Patents
Audio signal processing system for live music performance Download PDFInfo
- Publication number
- US20120269357A1 US20120269357A1 US13/471,225 US201213471225A US2012269357A1 US 20120269357 A1 US20120269357 A1 US 20120269357A1 US 201213471225 A US201213471225 A US 201213471225A US 2012269357 A1 US2012269357 A1 US 2012269357A1
- Authority
- US
- United States
- Prior art keywords
- signal
- signals
- audio
- signal processing
- manipulated
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0091—Means for obtaining special acoustic effects
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0033—Recording/reproducing or transmission of music for electrophonic musical instruments
- G10H1/0041—Recording/reproducing or transmission of music for electrophonic musical instruments in coded form
- G10H1/0058—Transmission between separate instruments or between individual components of a musical system
- G10H1/0066—Transmission between separate instruments or between individual components of a musical system using a MIDI interface
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2220/00—Input/output interfacing specifically adapted for electrophonic musical tools or instruments
- G10H2220/091—Graphical user interface [GUI] specifically adapted for electrophonic musical instruments, e.g. interactive musical displays, musical instrument icons or menus; Details of user interactions therewith
- G10H2220/101—Graphical user interface [GUI] specifically adapted for electrophonic musical instruments, e.g. interactive musical displays, musical instrument icons or menus; Details of user interactions therewith for graphical creation, edition or control of musical data or parameters
Definitions
- the invention is an electronic system for processing audio signals such as those produced by an instrument or a vocalist through a microphone during live music performances in an effectively infinite variety of ways.
- Instrument libraries include the notes and other sound features of virtually all commonly used instruments. Using any MIDI-compatible source such as a keyboard, drum pad or stringed instrument, a musician can “play” and record music with any of these instruments. Systems of these types are, however, very complex and require extensive training to be used effectively. They are also relatively expensive. For these reasons they are not suitable for use during live musical and/or vocal performances.
- Audio sound manipulation systems used for live performances are also available, although these systems generally offer relatively limited functionality.
- Guitarists for example, commonly use effects pedals or stomp boxes to manipulate the sound of their guitars during live performances.
- Stomp boxes are special-purpose audio processors connected between the guitar and amplifier that manipulate the clean guitar signal in predetermined manners. Distortion, fuzz, reverberation, and wah-wah are examples of the effects that can be added to the signal produced by the guitar itself before it is amplified and played to the listeners through speakers during a performance.
- a number of different stomp boxes can be chained together to provide the guitarist the ability to effect the sound in many different ways.
- An effects processor that has the capability of providing greater varieties of plug-ins for live performances is the Plugzilla audio processor available from Manifold Labs. Audio sources interface to the Plugzilla processor through a conventional mixer. The functionality of this processor is, however, relatively limited, and it can be difficult to operate.
- Such a system should be capable of providing a large variety of sound manipulation functions.
- the system should be relatively easy to use and operate. To be commercially viable, it should also be relatively inexpensive.
- the invention is an improved signal processing system for generating and/or manipulating sound in real time.
- the system includes one or more audio inputs for receiving audio signals, one or more trigger inputs for receiving virtual instrument trigger signals, and memory for storing sound effects plug-ins and libraries of virtual instruments.
- a graphical user interface enables a musician to select one or more of the sound effects plug-ins and/or virtual instruments from the memory.
- a digital processor coupled to the audio inputs, trigger inputs, memory and user interface processes the signals in real time. Music signals produced by the processor are outputted in real time through one or more audio outputs.
- Functions that can be provided by the processor include: (1) manipulating the received audio signals as a function of the selected sound effects plug-ins to produce manipulated audio signals, and/or (2) generating virtual instrument sound signals as a function of the received trigger signals and the selected virtual instruments and/or (3) manipulating the virtual instrument sound signals as a function of the selected sound effect plug-ins to produce manipulated virtual instrument signals, and/or (4) combining the received audio signals and/or the manipulated audio signals and/or the virtual instrument sound signals and/or the manipulated virtual instrument signals to produce combined signals, and/or (5) manipulating any or all of the combined signals to produce manipulated combined signals, and/or (6) repeating operations (4) and/or (5) with any or all of the combined signals and/or with any or all of the manipulated combined signals to produce iteratively processed signals, and (7) producing in real time as one or more output music signals the received audio signals and/or the manipulated audio signals and/or the virtual instrument signals and/or the manipulated virtual instrument signals and/or the combined signals and/or the
- FIG. 1 is a block diagram of a live music performance system including an audio signal processing system in accordance with the present invention.
- FIG. 2 is a detailed block diagram of one embodiment of the signal processing system shown in FIG. 1 .
- FIG. 3 is a flow diagram illustrating the music processing schemes that can be implemented with the signal processing system shown in FIG. 2 .
- FIG. 1 is a live music performance system 8 including an audio signal processing system 10 in accordance with the present invention.
- system 8 includes one or more audio sources 12 and one or more musical instrument digital interface (MIDI) trigger sources 16 connected to signal processing system 10 .
- Audio sources 12 are also connected to the signal processing system 10 through a conventional audio mixer 14 in the illustrated embodiment. Other embodiments of the invention (not shown) do not include mixer 14 .
- Audio sources 12 can be any source of electrical signals representative of audible sound such as guitars, keyboards, or other electric instruments and microphones (for providing vocal sound signals). Alternatively, audio sources 12 can be recorded or stored files of electrical signals that are operated to play back the electrical signals in real time.
- MIDI trigger sources 16 can be any sources of MIDI protocol electrical trigger signals such as keyboards, drum pads and guitars. Alternatively, trigger sources 16 can be stored files of such trigger signals that are executed to generate the trigger signals.
- signal processing system 10 includes a wide variety of software sound effects and other plug-ins, software instrument libraries and software vocal libraries. A musician or other operator can use the signal processing system 10 to select and generate sound or “play” any of the instruments or vocals from the libraries in response to the MIDI trigger sources 16 . The musician can also select any of the plug-ins and cause the sound of the instruments and/or vocals to be manipulated by the plug-ins.
- the musician can select plug-ins that are used to manipulate the sound of the audio sources 12 . After they are generated and/or manipulated by the signal processing system 10 , the audio signals are outputted to a conventional audio amplifier 18 which drives one or more speakers 20 . A listener (not shown) can then hear in real time or substantially real time the live music performance as it is created by the musician.
- FIG. 2 is a detailed block diagram of one embodiment of the signal processing system 10 .
- signal processing system 10 includes a central processing unit (CPU) 21 coupled to a graphic user interface 22 having a display screen 24 and user actuated controls 26 .
- Analog audio signals from the audio sources ( FIG. 1 ) are inputted to the signal processing system 10 through audio inputs 28 and converted into digital form by A/D (analog-to-digital) converters 30 .
- An audio interface 32 couples the digital audio signals from A/D converter 30 to CPU 21 .
- CPU 21 includes memory (e.g., random access memory) for storing data and signals such as the analog audio signals during the processing operations.
- Processed digital audio output signals produced by CPU 21 are converted to analog form by digital-to-analog (D/A) converter 34 and outputted from the signal processing system 10 through audio outputs 36 .
- audio interface 32 couples the CPU 21 to the A/D converter 34 .
- CPU 21 is controlled by an operating system 38 .
- Random access memory (RAM) 40 is coupled to the CPU 21 through an audio host 42 .
- memory 40 includes sound effect plug-ins 44 and libraries of virtual instruments 46 .
- Trigger signals from a MIDI source FIG. 1
- Audio inputs 28 and audio outputs 36 can be conventional analog devices such as commonly-used 1 ⁇ 4′′ balanced or unbalanced jacks.
- One embodiment of the invention includes an 8-channel audio input 28 and an 8-channel audio output 36 , although other embodiments have greater or fewer channels.
- A/D converters 30 and D/A converters 34 can be conventional devices operating at conventional sampling frequencies. By way of example, converters 30 and 34 can be 16- or 24-bit devices operating at sample frequencies of 41K Hz or higher.
- Other embodiments of the signal processing system 10 do not include A/D converters 30 and/or D/A converters 36 , and instead are configured to receive and output digital audio signals.
- the audio inputs 28 and audio outputs 36 can be conventional ADAT or S/PDIF jacks.
- Audio interface 32 converts the format of the digital signals provided by A/D converter 30 (or received from digital audio inputs 28 in the embodiments with no built-in A/D converter) to a format suitable for inputting into CPU 21 . Similarly, the audio interface 32 converts the format of the digital audio signals outputted from CPU 21 to a format suitable for inputting into D/A converter 34 (or to digital audio inputs 28 in the embodiments with no built-in D/A converter).
- CPU 21 includes one or more high speed microprocessors and associated random access memory.
- the operating system 38 run by CPU 21 can be a commercially available operating system such as OSX, Windows XP, Vista or Linux. Alternatively, the operating system 38 can be a proprietary system.
- Memory 40 is high-capacity, high-speed random access memory (RAM).
- RAM random access memory
- One embodiment of the invention includes 5 Gb of memory, although other embodiments include greater or lesser amounts. In general, the greater the amount of memory, the greater the number and the higher the quality of the sound effect plug-ins 44 and the virtual instruments 46 that can be stored in the memory 40 .
- Memory 40 can be included within the same housing or enclosure as other components of signal processing system 10 , or in a separate enclosure that is connected to the other components of the signal processing system by a conventional interface.
- Preferably stored within memory 40 is a large number and wide variety of software plug-ins 44 that can be used by CPU 21 to manipulate the audio signals.
- sound effects plug-ins and sampling sequences can be stored in memory 40 .
- These plug-ins 44 can be commercially available software and/or proprietary software.
- preferred embodiments of the invention include a large number and a wide variety of software virtual instruments 46 that can be used by CPU 21 to generate audio signals in response to MIDI trigger sources. Examples of virtual instruments of these types include vocal and synthetic sounds as well as those producing conventional instrument sounds.
- the virtual instruments 46 within memory 40 can be commercially available software and/or proprietary software.
- preferred embodiments of the signal processing system 10 will include one or more interfaces enabling software to be conveniently and relatively quickly loaded into the memory 40 . CD and DVD drives and Firewire, USB and Bluetooth ports are examples of the interfaces that can be included for this purpose.
- One or more hosts 42 are included to convert the software plug-ins 44 and instruments 46 in memory 40 to a format suitable for operation by CPU 21 .
- Commercially available hosts 42 such as Real Time Audio Suite (RTAS), Virtual Sound Technology (VST), and Audio Units (AU) that are compatible with commercially available software plug-ins 44 and instruments 46 can be used for this purpose.
- RTAS Real Time Audio Suite
- VST Virtual Sound Technology
- AU Audio Units
- one or more proprietary hosts can be used in connection with proprietary software plug-ins 44 and instruments 46 .
- MIDI interface 48 converts the conventional MIDI protocol trigger signals received from sources such as 16 ( FIG. 1 ) to a format used by CPU 21 .
- Other embodiments of the invention may be configured to receive trigger signals in other protocols (as an alternative and/or in addition to MIDI signals), and these embodiments would include an interface to convert any such trigger signals to the format used by CPU 21 .
- Display screen 24 can be a conventional LCD or LED device providing text and/or graphical displays.
- User controls 26 can be buttons, a key pad, a mouse or other structures that are actuated by a user. Display screen 24 and user controls 26 function together as a graphical user interface 22 , enabling a musician to easily access and operate all the functions available from signal processing system 10 . By way of example, a musician can operate the user interface 22 to select one or more plug-ins 44 and/or one or more virtual instruments 46 .
- the user interface 22 can also be operated to select a processing scheme by which the inputted analog signals, and/or selected virtual instruments 46 will be processed by the plug-ins 44 (and/or combined and/or reprocessed with other analog signals, virtual instruments and/or plug-ins as discussed in greater detail below) to establish a performance arrangement.
- the user interface 22 allows users to store selected plug-ins 44 , virtual instruments 46 and/or processing schemes. The musician can thereby easily select all the parameters required for a previously established performance arrangement. Stored performance arrangement information can also be presented through the user interface 22 as presets stored during the manufacture of the processing system 10 .
- the user interface 22 includes databases of stored information that enable a user to create a certain “sound” without knowing all aspects of the performance arrangement required to achieve that sound.
- the user interface 22 can prompt the musician to input (e.g., select from a menu) a desired output sound (e.g., a celebrity musician or band).
- a desired output sound e.g., a celebrity musician or band.
- the user interface 22 can also prompt the musician to input information representative of the analog source they will be using to provide audio input signals (e.g., what guitar is the musician playing).
- the stored databases will include sufficient information to enable the selection of the plug-ins 44 and/or virtual instruments 46 and the processing schemes that the CPU 21 can implement to achieve a performance arrangement that will produce music signals having the sound desired by the musician.
- FIG. 3 is a flow diagram illustrating the essentially infinite processing schemes that can be implemented with selected plug-ins 44 and selected virtual instruments 46 to achieve an essentially infinite number of performance arrangements.
- inputted audio signals can be processed by selected plug-ins 44 to produce manipulated audio signals.
- virtual instrument sound signals can be generated as a function of the received MIDI trigger signals and the selected virtual instruments 46 as represented by path 62 .
- the virtual instrument sound signals can be processed by selected plug-ins 44 to produce manipulated virtual instrument signals represented at path 64 .
- Any or all manipulated audio signals from path 60 can be combined with any or all manipulated virtual instrument signals from path 64 , as represented by summing node 68 .
- the “unprocessed” audio signals (e.g., from path 66 ) and/or the “unprocessed” virtual instrument sound signals (e.g., from paths 62 and 70 ) can also be combined at node 68 , if desired, with any other signals at the node (e.g., with the manipulated audio signals and/or the manipulated virtual instrument signals as described above).
- the music signals produced by such a first iteration performance arrangement can be outputted from node 68 .
- At least some embodiments of system 10 also have the capability of further processing any or all of the first iteration music signals available from node 68 .
- the music signals from node 68 can be processed by selected plug-ins 44 (that can be the same or different plug-ins than any used in the first iteration) to produce manipulated combined signals.
- the music signals from node 68 can also be recombined with the unprocessed audio signals, the manipulated audio signals, the unprocessed virtual instrument sound signals and/or the manipulated virtual instrument sound signals.
- the music signals produced by such a second iteration performance arrangement can be outputted from node 68 .
- Still other embodiments of system 10 also have the capability of further processing any or all of the second iteration music signals available from node 68 . As indicated by path 82 , any or all of the processing scheme components described above can be repeated with any or all of the signals produced by system 10 . The music signals produced by any such further iteration performance arrangements can be outputted form node 68 .
- Still other embodiments of system 10 offer only subsets of the effectively infinite performance arrangements that can be provided by the embodiments described above. For example, one embodiment of the invention allows only the first iteration performance arrangements. Still other embodiments of system 10 offer only other subsets for the performance arrangements described above.
- One (but not all) embodiment of signal processing system 10 is a limited-functionality device dedicated to use in live performances. This embodiment does not include components typically found in systems used for music recording applications.
- Embodiments of the invention can be implemented using the Rax virtual rack software available from Audiofile Engineering of St. Paul, Minn.
- the Rax software can effectively function as the host 42 of the embodiment of the invention illustrated in FIG. 2 .
- a Manual and other technical information describing the Rax software is available on the Audiofile Engineering website (audiofile-engineering.com), and are incorporated herein by reference in their entirety for all purposes.
- Audiofile Engineering also distributes an audio file editing system known as Wave Editor.
- the Wave Editor file editing system can be incorporated into signal processing system 10 as a system for processing recorded or stored sound files created using the signal processing system, and/or as a system for implementing the signal processing functionality of system 10 .
- the Wave Editor software is described in the Wave Editor User's Guide available on the Audiofile Engineering website, and in the Foust et al. U.S. Patent Application Publication No. 2008/0041220, both of which documents are incorporated herein by reference in their entirety for all purposes.
- signal processing system 10 An important advantage of signal processing system 10 over currently available systems is the high quality of the sound that is produced by the system. Another important advantage provided by signal processing system 10 is its ease of use. All of the functions of the system 10 can be conveniently accessed by a musician through relatively few layers of menu structure in the user interface 22 . Yet another advantage of signal processing system 10 is its relatively compact size. The above-described robust function set of signal processing system 10 is thereby achieved at a relatively inexpensive price.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Electrophonic Musical Instruments (AREA)
Abstract
Description
- This application is a continuation of U.S. application Ser. No. 12/055,903, filed Mar. 26, 2008, which claims the benefit of U.S. Provisional Application Ser. No. 60/921,154, filed on Mar. 30, 2007, and entitled Audio Signal Processing System For Live Music Performance, which are incorporated herein by reference in their entirety.
- The invention is an electronic system for processing audio signals such as those produced by an instrument or a vocalist through a microphone during live music performances in an effectively infinite variety of ways.
- Musicians and vocalists have a wide range of audio signal processing systems available to them during recording sessions. One system widely used in professional recording studios is a workstation with Digidesign's ProTools audio mixer software. These workstations include a wide variety of software sound effects libraries, sampling sequences and other so-called “plug-ins” that can be used to manipulate the audio instrument or vocal source. They also include virtual instrument and vocal libraries that can be “played” and recorded in response to signals (e.g., Musical Instrument Digital Interface (MIDI) trigger signals) inputted into the workstation. Using a “celebrity” guitarist sound effect library, for example, the workstation can manipulate any inputted guitar signal in such a manner as to have the signature sounds of that celebrity guitarist. The sound of vocalists can be enhanced by manipulating dynamics, correcting pitch or by injecting reverberation or digital delay to mask undesirable vocal characteristics, or to enhance appealing ones. Instrument libraries include the notes and other sound features of virtually all commonly used instruments. Using any MIDI-compatible source such as a keyboard, drum pad or stringed instrument, a musician can “play” and record music with any of these instruments. Systems of these types are, however, very complex and require extensive training to be used effectively. They are also relatively expensive. For these reasons they are not suitable for use during live musical and/or vocal performances.
- Audio sound manipulation systems used for live performances are also available, although these systems generally offer relatively limited functionality. Guitarists, for example, commonly use effects pedals or stomp boxes to manipulate the sound of their guitars during live performances. Stomp boxes are special-purpose audio processors connected between the guitar and amplifier that manipulate the clean guitar signal in predetermined manners. Distortion, fuzz, reverberation, and wah-wah are examples of the effects that can be added to the signal produced by the guitar itself before it is amplified and played to the listeners through speakers during a performance. A number of different stomp boxes can be chained together to provide the guitarist the ability to effect the sound in many different ways.
- An effects processor that has the capability of providing greater varieties of plug-ins for live performances is the Plugzilla audio processor available from Manifold Labs. Audio sources interface to the Plugzilla processor through a conventional mixer. The functionality of this processor is, however, relatively limited, and it can be difficult to operate.
- There remains a need for improved audio signal processing systems suitable for use with live performances. Such a system should be capable of providing a large variety of sound manipulation functions. The system should be relatively easy to use and operate. To be commercially viable, it should also be relatively inexpensive.
- The invention is an improved signal processing system for generating and/or manipulating sound in real time. The system includes one or more audio inputs for receiving audio signals, one or more trigger inputs for receiving virtual instrument trigger signals, and memory for storing sound effects plug-ins and libraries of virtual instruments. A graphical user interface enables a musician to select one or more of the sound effects plug-ins and/or virtual instruments from the memory. A digital processor coupled to the audio inputs, trigger inputs, memory and user interface processes the signals in real time. Music signals produced by the processor are outputted in real time through one or more audio outputs. Functions that can be provided by the processor include: (1) manipulating the received audio signals as a function of the selected sound effects plug-ins to produce manipulated audio signals, and/or (2) generating virtual instrument sound signals as a function of the received trigger signals and the selected virtual instruments and/or (3) manipulating the virtual instrument sound signals as a function of the selected sound effect plug-ins to produce manipulated virtual instrument signals, and/or (4) combining the received audio signals and/or the manipulated audio signals and/or the virtual instrument sound signals and/or the manipulated virtual instrument signals to produce combined signals, and/or (5) manipulating any or all of the combined signals to produce manipulated combined signals, and/or (6) repeating operations (4) and/or (5) with any or all of the combined signals and/or with any or all of the manipulated combined signals to produce iteratively processed signals, and (7) producing in real time as one or more output music signals the received audio signals and/or the manipulated audio signals and/or the virtual instrument signals and/or the manipulated virtual instrument signals and/or the combined signals and/or the manipulated combined signals and/or the iteratively processed signals.
-
FIG. 1 is a block diagram of a live music performance system including an audio signal processing system in accordance with the present invention. -
FIG. 2 is a detailed block diagram of one embodiment of the signal processing system shown inFIG. 1 . -
FIG. 3 is a flow diagram illustrating the music processing schemes that can be implemented with the signal processing system shown inFIG. 2 . -
FIG. 1 is a livemusic performance system 8 including an audiosignal processing system 10 in accordance with the present invention. As shown,system 8 includes one ormore audio sources 12 and one or more musical instrument digital interface (MIDI)trigger sources 16 connected tosignal processing system 10.Audio sources 12 are also connected to thesignal processing system 10 through aconventional audio mixer 14 in the illustrated embodiment. Other embodiments of the invention (not shown) do not includemixer 14.Audio sources 12 can be any source of electrical signals representative of audible sound such as guitars, keyboards, or other electric instruments and microphones (for providing vocal sound signals). Alternatively,audio sources 12 can be recorded or stored files of electrical signals that are operated to play back the electrical signals in real time.MIDI trigger sources 16 can be any sources of MIDI protocol electrical trigger signals such as keyboards, drum pads and guitars. Alternatively,trigger sources 16 can be stored files of such trigger signals that are executed to generate the trigger signals. As described in greater detail below,signal processing system 10 includes a wide variety of software sound effects and other plug-ins, software instrument libraries and software vocal libraries. A musician or other operator can use thesignal processing system 10 to select and generate sound or “play” any of the instruments or vocals from the libraries in response to theMIDI trigger sources 16. The musician can also select any of the plug-ins and cause the sound of the instruments and/or vocals to be manipulated by the plug-ins. Alternatively or in addition to the playing of instruments and vocals, the musician can select plug-ins that are used to manipulate the sound of theaudio sources 12. After they are generated and/or manipulated by thesignal processing system 10, the audio signals are outputted to aconventional audio amplifier 18 which drives one ormore speakers 20. A listener (not shown) can then hear in real time or substantially real time the live music performance as it is created by the musician. -
FIG. 2 is a detailed block diagram of one embodiment of thesignal processing system 10. As shown,signal processing system 10 includes a central processing unit (CPU) 21 coupled to a graphic user interface 22 having adisplay screen 24 and user actuatedcontrols 26. Analog audio signals from the audio sources (FIG. 1 ) are inputted to thesignal processing system 10 throughaudio inputs 28 and converted into digital form by A/D (analog-to-digital)converters 30. Anaudio interface 32 couples the digital audio signals from A/D converter 30 toCPU 21. Although not separately shown,CPU 21 includes memory (e.g., random access memory) for storing data and signals such as the analog audio signals during the processing operations. Processed digital audio output signals produced byCPU 21 are converted to analog form by digital-to-analog (D/A)converter 34 and outputted from thesignal processing system 10 throughaudio outputs 36. As shown,audio interface 32 couples theCPU 21 to the A/D converter 34.CPU 21 is controlled by anoperating system 38. Random access memory (RAM) 40 is coupled to theCPU 21 through anaudio host 42. As shown,memory 40 includes sound effect plug-ins 44 and libraries ofvirtual instruments 46. Trigger signals from a MIDI source (FIG. 1 ) are coupled to theCPU 21 through aMIDI interface 48. -
Audio inputs 28 andaudio outputs 36 can be conventional analog devices such as commonly-used ¼″ balanced or unbalanced jacks. One embodiment of the invention includes an 8-channel audio input 28 and an 8-channel audio output 36, although other embodiments have greater or fewer channels. A/D converters 30 and D/A converters 34 can be conventional devices operating at conventional sampling frequencies. By way of example,converters D converters 30 and/or D/A converters 36, and instead are configured to receive and output digital audio signals. In these embodiments of the invention, theaudio inputs 28 andaudio outputs 36 can be conventional ADAT or S/PDIF jacks. -
Audio interface 32 converts the format of the digital signals provided by A/D converter 30 (or received from digitalaudio inputs 28 in the embodiments with no built-in A/D converter) to a format suitable for inputting intoCPU 21. Similarly, theaudio interface 32 converts the format of the digital audio signals outputted fromCPU 21 to a format suitable for inputting into D/A converter 34 (or to digitalaudio inputs 28 in the embodiments with no built-in D/A converter). -
CPU 21 includes one or more high speed microprocessors and associated random access memory. Theoperating system 38 run byCPU 21 can be a commercially available operating system such as OSX, Windows XP, Vista or Linux. Alternatively, theoperating system 38 can be a proprietary system. -
Memory 40 is high-capacity, high-speed random access memory (RAM). One embodiment of the invention includes 5 Gb of memory, although other embodiments include greater or lesser amounts. In general, the greater the amount of memory, the greater the number and the higher the quality of the sound effect plug-ins 44 and thevirtual instruments 46 that can be stored in thememory 40.Memory 40 can be included within the same housing or enclosure as other components ofsignal processing system 10, or in a separate enclosure that is connected to the other components of the signal processing system by a conventional interface. - Preferably stored within
memory 40 is a large number and wide variety of software plug-ins 44 that can be used byCPU 21 to manipulate the audio signals. By way of example, sound effects plug-ins and sampling sequences can be stored inmemory 40. These plug-ins 44 can be commercially available software and/or proprietary software. Similarly, preferred embodiments of the invention include a large number and a wide variety of softwarevirtual instruments 46 that can be used byCPU 21 to generate audio signals in response to MIDI trigger sources. Examples of virtual instruments of these types include vocal and synthetic sounds as well as those producing conventional instrument sounds. Thevirtual instruments 46 withinmemory 40 can be commercially available software and/or proprietary software. Although not shown inFIG. 2 , preferred embodiments of thesignal processing system 10 will include one or more interfaces enabling software to be conveniently and relatively quickly loaded into thememory 40. CD and DVD drives and Firewire, USB and Bluetooth ports are examples of the interfaces that can be included for this purpose. - One or
more hosts 42 are included to convert the software plug-ins 44 andinstruments 46 inmemory 40 to a format suitable for operation byCPU 21. Commerciallyavailable hosts 42 such as Real Time Audio Suite (RTAS), Virtual Sound Technology (VST), and Audio Units (AU) that are compatible with commercially available software plug-ins 44 andinstruments 46 can be used for this purpose. Alternatively, or in addition to the commerciallyavailable hosts 42, one or more proprietary hosts can be used in connection with proprietary software plug-ins 44 andinstruments 46. -
MIDI interface 48 converts the conventional MIDI protocol trigger signals received from sources such as 16 (FIG. 1 ) to a format used byCPU 21. Other embodiments of the invention may be configured to receive trigger signals in other protocols (as an alternative and/or in addition to MIDI signals), and these embodiments would include an interface to convert any such trigger signals to the format used byCPU 21. -
Display screen 24 can be a conventional LCD or LED device providing text and/or graphical displays. User controls 26 can be buttons, a key pad, a mouse or other structures that are actuated by a user.Display screen 24 anduser controls 26 function together as a graphical user interface 22, enabling a musician to easily access and operate all the functions available fromsignal processing system 10. By way of example, a musician can operate the user interface 22 to select one or more plug-ins 44 and/or one or morevirtual instruments 46. The user interface 22 can also be operated to select a processing scheme by which the inputted analog signals, and/or selectedvirtual instruments 46 will be processed by the plug-ins 44 (and/or combined and/or reprocessed with other analog signals, virtual instruments and/or plug-ins as discussed in greater detail below) to establish a performance arrangement. In one embodiment of the invention the user interface 22 allows users to store selected plug-ins 44,virtual instruments 46 and/or processing schemes. The musician can thereby easily select all the parameters required for a previously established performance arrangement. Stored performance arrangement information can also be presented through the user interface 22 as presets stored during the manufacture of theprocessing system 10. - In another embodiment of the invention the user interface 22 includes databases of stored information that enable a user to create a certain “sound” without knowing all aspects of the performance arrangement required to achieve that sound. In this embodiment, for example, the user interface 22 can prompt the musician to input (e.g., select from a menu) a desired output sound (e.g., a celebrity musician or band). In a similar manner the user interface 22 can also prompt the musician to input information representative of the analog source they will be using to provide audio input signals (e.g., what guitar is the musician playing). The stored databases will include sufficient information to enable the selection of the plug-
ins 44 and/orvirtual instruments 46 and the processing schemes that theCPU 21 can implement to achieve a performance arrangement that will produce music signals having the sound desired by the musician. -
Signal processing system 10 is used by a musician to generate and/or manipulate sound during the live or real-time performance of music. Audio sounds can be generated and/or manipulated in essentially infinite numbers ofways using system 10.FIG. 3 is a flow diagram illustrating the essentially infinite processing schemes that can be implemented with selected plug-ins 44 and selectedvirtual instruments 46 to achieve an essentially infinite number of performance arrangements. As indicated bypath 60, inputted audio signals can be processed by selected plug-ins 44 to produce manipulated audio signals. Alternatively or in addition to the inputted audio signal processing described above, virtual instrument sound signals can be generated as a function of the received MIDI trigger signals and the selectedvirtual instruments 46 as represented bypath 62. The virtual instrument sound signals can be processed by selected plug-ins 44 to produce manipulated virtual instrument signals represented atpath 64. Any or all manipulated audio signals frompath 60 can be combined with any or all manipulated virtual instrument signals frompath 64, as represented by summingnode 68. The “unprocessed” audio signals (e.g., from path 66) and/or the “unprocessed” virtual instrument sound signals (e.g., frompaths 62 and 70) can also be combined atnode 68, if desired, with any other signals at the node (e.g., with the manipulated audio signals and/or the manipulated virtual instrument signals as described above). The music signals produced by such a first iteration performance arrangement can be outputted fromnode 68. - At least some embodiments of
system 10 also have the capability of further processing any or all of the first iteration music signals available fromnode 68. As represented bypath 72, the music signals fromnode 68 can be processed by selected plug-ins 44 (that can be the same or different plug-ins than any used in the first iteration) to produce manipulated combined signals. As represented bypaths node 68 can also be recombined with the unprocessed audio signals, the manipulated audio signals, the unprocessed virtual instrument sound signals and/or the manipulated virtual instrument sound signals. The music signals produced by such a second iteration performance arrangement can be outputted fromnode 68. - Still other embodiments of
system 10 also have the capability of further processing any or all of the second iteration music signals available fromnode 68. As indicated bypath 82, any or all of the processing scheme components described above can be repeated with any or all of the signals produced bysystem 10. The music signals produced by any such further iteration performance arrangements can be outputtedform node 68. - Still other embodiments of
system 10 offer only subsets of the effectively infinite performance arrangements that can be provided by the embodiments described above. For example, one embodiment of the invention allows only the first iteration performance arrangements. Still other embodiments ofsystem 10 offer only other subsets for the performance arrangements described above. - One (but not all) embodiment of
signal processing system 10 is a limited-functionality device dedicated to use in live performances. This embodiment does not include components typically found in systems used for music recording applications. - Embodiments of the invention can be implemented using the Rax virtual rack software available from Audiofile Engineering of St. Paul, Minn. In particular, the Rax software can effectively function as the
host 42 of the embodiment of the invention illustrated inFIG. 2 . A Manual and other technical information describing the Rax software is available on the Audiofile Engineering website (audiofile-engineering.com), and are incorporated herein by reference in their entirety for all purposes. Audiofile Engineering also distributes an audio file editing system known as Wave Editor. The Wave Editor file editing system can be incorporated intosignal processing system 10 as a system for processing recorded or stored sound files created using the signal processing system, and/or as a system for implementing the signal processing functionality ofsystem 10. The Wave Editor software is described in the Wave Editor User's Guide available on the Audiofile Engineering website, and in the Foust et al. U.S. Patent Application Publication No. 2008/0041220, both of which documents are incorporated herein by reference in their entirety for all purposes. - An important advantage of
signal processing system 10 over currently available systems is the high quality of the sound that is produced by the system. Another important advantage provided bysignal processing system 10 is its ease of use. All of the functions of thesystem 10 can be conveniently accessed by a musician through relatively few layers of menu structure in the user interface 22. Yet another advantage ofsignal processing system 10 is its relatively compact size. The above-described robust function set ofsignal processing system 10 is thereby achieved at a relatively inexpensive price. - Although the present invention has been described with reference to preferred embodiments, those skilled in the art will recognize that changes can be made in form and detail without departing from the spirit and scope of the invention.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/471,225 US20120269357A1 (en) | 2007-03-30 | 2012-05-14 | Audio signal processing system for live music performance |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US92115407P | 2007-03-30 | 2007-03-30 | |
US12/055,903 US8180063B2 (en) | 2007-03-30 | 2008-03-26 | Audio signal processing system for live music performance |
US13/471,225 US20120269357A1 (en) | 2007-03-30 | 2012-05-14 | Audio signal processing system for live music performance |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/055,903 Continuation US8180063B2 (en) | 2007-03-30 | 2008-03-26 | Audio signal processing system for live music performance |
Publications (1)
Publication Number | Publication Date |
---|---|
US20120269357A1 true US20120269357A1 (en) | 2012-10-25 |
Family
ID=39794413
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/055,903 Active 2030-12-14 US8180063B2 (en) | 2007-03-30 | 2008-03-26 | Audio signal processing system for live music performance |
US13/471,225 Abandoned US20120269357A1 (en) | 2007-03-30 | 2012-05-14 | Audio signal processing system for live music performance |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/055,903 Active 2030-12-14 US8180063B2 (en) | 2007-03-30 | 2008-03-26 | Audio signal processing system for live music performance |
Country Status (2)
Country | Link |
---|---|
US (2) | US8180063B2 (en) |
WO (1) | WO2008121650A1 (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130319208A1 (en) * | 2011-03-15 | 2013-12-05 | David Forrest | Musical learning and interaction through shapes |
WO2015066204A1 (en) * | 2013-10-30 | 2015-05-07 | Music Mastermind, Inc. | System and method for enhancing audio, conforming an audio input to a musical key, and creating harmonizing tracks for an audio input |
US9177540B2 (en) | 2009-06-01 | 2015-11-03 | Music Mastermind, Inc. | System and method for conforming an audio input to a musical key |
US9251776B2 (en) | 2009-06-01 | 2016-02-02 | Zya, Inc. | System and method creating harmonizing tracks for an audio input |
US9263021B2 (en) | 2009-06-01 | 2016-02-16 | Zya, Inc. | Method for generating a musical compilation track from multiple takes |
US9310959B2 (en) | 2009-06-01 | 2016-04-12 | Zya, Inc. | System and method for enhancing audio |
Families Citing this family (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2008121650A1 (en) * | 2007-03-30 | 2008-10-09 | William Henderson | Audio signal processing system for live music performance |
DE102007016274B4 (en) * | 2007-04-04 | 2013-02-07 | Lawo Ag | Device and method for using audio plug-ins in a mixing console |
US8706247B2 (en) * | 2009-10-23 | 2014-04-22 | Advanced Bionics | Remote audio processor module for auditory prosthesis systems |
DE102011003976B3 (en) * | 2011-02-11 | 2012-04-26 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Sound input device for use in e.g. music instrument input interface in electric guitar, has classifier interrupting output of sound signal over sound signal output during presence of condition for period of sound signal passages |
WO2013166188A1 (en) * | 2012-05-01 | 2013-11-07 | Orshan Jesse Harris | Virtual audio effects package and corresponding network |
US8957297B2 (en) | 2012-06-12 | 2015-02-17 | Harman International Industries, Inc. | Programmable musical instrument pedalboard |
US9160294B2 (en) * | 2013-03-05 | 2015-10-13 | Thomas Alexander Allen | Virtual pre-amplifier and effects system and methods for customizing and using the same in live performances |
CN103646656B (en) * | 2013-11-29 | 2016-05-04 | 腾讯科技(成都)有限公司 | Sound effect treatment method, device, plugin manager and audio plug-in unit |
US10741155B2 (en) | 2013-12-06 | 2020-08-11 | Intelliterran, Inc. | Synthesized percussion pedal and looping station |
US9905210B2 (en) | 2013-12-06 | 2018-02-27 | Intelliterran Inc. | Synthesized percussion pedal and docking station |
US9495947B2 (en) | 2013-12-06 | 2016-11-15 | Intelliterran Inc. | Synthesized percussion pedal and docking station |
US20150302086A1 (en) | 2014-04-22 | 2015-10-22 | Gracenote, Inc. | Audio identification during performance |
JP6467887B2 (en) * | 2014-11-21 | 2019-02-13 | ヤマハ株式会社 | Information providing apparatus and information providing method |
JP6801225B2 (en) | 2016-05-18 | 2020-12-16 | ヤマハ株式会社 | Automatic performance system and automatic performance method |
US9542919B1 (en) * | 2016-07-20 | 2017-01-10 | Beamz Interactive, Inc. | Cyber reality musical instrument and device |
JP7193167B2 (en) | 2017-08-29 | 2022-12-20 | インテリテラン,インク. | Apparatus, system and method for recording and rendering multimedia |
US11789689B2 (en) * | 2018-01-19 | 2023-10-17 | Microsoft Technology Licensing, Llc | Processing digital audio using audio processing plug-ins executing in a distributed computing environment |
JP2022041553A (en) * | 2020-09-01 | 2022-03-11 | ヤマハ株式会社 | Communication control method |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5850455A (en) * | 1996-06-18 | 1998-12-15 | Extreme Audio Reality, Inc. | Discrete dynamic positioning of audio signals in a 360° environment |
US20030091204A1 (en) * | 1992-04-27 | 2003-05-15 | Gibson David A. | Method and apparatus for using visual images to mix sound |
US6931134B1 (en) * | 1998-07-28 | 2005-08-16 | James K. Waller, Jr. | Multi-dimensional processor and multi-dimensional audio processor system |
Family Cites Families (78)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE3463306D1 (en) | 1983-01-18 | 1987-05-27 | Matsushita Electric Ind Co Ltd | Wave generating apparatus |
US4961364A (en) | 1987-02-25 | 1990-10-09 | Casio Computer Co., Ltd. | Musical tone generating apparatus for synthesizing musical tone signal by combining component wave signals |
US5092216A (en) * | 1989-08-17 | 1992-03-03 | Wayne Wadhams | Method and apparatus for studying music |
US5225618A (en) * | 1989-08-17 | 1993-07-06 | Wayne Wadhams | Method and apparatus for studying music |
JP2768168B2 (en) | 1992-09-18 | 1998-06-25 | ヤマハ株式会社 | Music synthesizer |
US5331111A (en) | 1992-10-27 | 1994-07-19 | Korg, Inc. | Sound model generator and synthesizer with graphical programming engine |
US5376752A (en) * | 1993-02-10 | 1994-12-27 | Korg, Inc. | Open architecture music synthesizer with dynamic voice allocation |
JP3072452B2 (en) | 1993-03-19 | 2000-07-31 | ヤマハ株式会社 | Karaoke equipment |
JP3206619B2 (en) | 1993-04-23 | 2001-09-10 | ヤマハ株式会社 | Karaoke equipment |
US5393926A (en) | 1993-06-07 | 1995-02-28 | Ahead, Inc. | Virtual music system |
JPH096350A (en) | 1993-11-02 | 1997-01-10 | Yamaha Corp | Effect adding device and electronic musical instrument using the same |
US5511000A (en) * | 1993-11-18 | 1996-04-23 | Kaloi; Dennis M. | Electronic solid-state record/playback device and system |
TW281747B (en) | 1994-03-31 | 1996-07-21 | Yamaha Corp | |
EP0743631B1 (en) * | 1995-05-19 | 2002-03-06 | Yamaha Corporation | Tone generating method and device |
US5740260A (en) * | 1995-05-22 | 1998-04-14 | Presonus L.L.P. | Midi to analog sound processor interface |
JP2998612B2 (en) | 1995-06-06 | 2000-01-11 | ヤマハ株式会社 | Music generator |
JP3235409B2 (en) | 1995-06-07 | 2001-12-04 | ヤマハ株式会社 | Music system, sound source and tone synthesis method |
US5663517A (en) | 1995-09-01 | 1997-09-02 | International Business Machines Corporation | Interactive system for compositional morphing of music in real-time |
JP2861881B2 (en) | 1995-09-04 | 1999-02-24 | ヤマハ株式会社 | Chorus effect imparting device |
US5848164A (en) | 1996-04-30 | 1998-12-08 | The Board Of Trustees Of The Leland Stanford Junior University | System and method for effects processing on audio subband data |
US5781188A (en) | 1996-06-27 | 1998-07-14 | Softimage | Indicating activeness of clips and applying effects to clips and tracks in a timeline of a multimedia work |
US7297856B2 (en) | 1996-07-10 | 2007-11-20 | Sitrick David H | System and methodology for coordinating musical communication and display |
SG67993A1 (en) | 1996-08-30 | 1999-10-19 | Yamaha Corp | Sound source system based on computer software and method of generating acoustic waveform data |
US5952597A (en) | 1996-10-25 | 1999-09-14 | Timewarp Technologies, Ltd. | Method and apparatus for real-time correlation of a performance to a musical score |
US5850628A (en) * | 1997-01-30 | 1998-12-15 | Hasbro, Inc. | Speech and sound synthesizers with connected memories and outputs |
US5913258A (en) * | 1997-03-11 | 1999-06-15 | Yamaha Corporation | Music tone generating method by waveform synthesis with advance parameter computation |
US5930158A (en) | 1997-07-02 | 1999-07-27 | Creative Technology, Ltd | Processor with instruction set for audio effects |
US5928342A (en) | 1997-07-02 | 1999-07-27 | Creative Technology Ltd. | Audio effects processor integrated on a single chip with a multiport memory onto which multiple asynchronous digital sound samples can be concurrently loaded |
EP0913808B1 (en) | 1997-10-31 | 2004-09-29 | Yamaha Corporation | Audio signal processor with pitch and effect control |
WO1999037032A1 (en) | 1998-01-15 | 1999-07-22 | Mackie Designs Inc. | Digital signal mixing architecture |
US6839441B1 (en) | 1998-01-20 | 2005-01-04 | Showco, Inc. | Sound mixing console with master control section |
US6103964A (en) | 1998-01-28 | 2000-08-15 | Kay; Stephen R. | Method and apparatus for generating algorithmic musical effects |
US5986199A (en) * | 1998-05-29 | 1999-11-16 | Creative Technology, Ltd. | Device for acoustic entry of musical data |
FR2783630B1 (en) | 1998-09-23 | 2000-12-15 | Pierre Guilmette | SOUND SYNTHESIS SYSTEM FOR PROVIDING A SUIT OF ELECTRICAL SAMPLES |
AU5009399A (en) * | 1998-09-24 | 2000-05-04 | Sony Corporation | Impulse response collecting method, sound effect adding apparatus, and recording medium |
US6087578A (en) | 1999-01-28 | 2000-07-11 | Kay; Stephen R. | Method and apparatus for generating and controlling automatic pitch bending effects |
FR2792747B1 (en) | 1999-04-22 | 2001-06-22 | France Telecom | DEVICE FOR ACQUIRING AND PROCESSING SIGNALS FOR CONTROLLING AN APPARATUS OR A PROCESS |
US6327367B1 (en) | 1999-05-14 | 2001-12-04 | G. Scott Vercoe | Sound effects controller |
US6392133B1 (en) * | 2000-10-17 | 2002-05-21 | Dbtech Sarl | Automatic soundtrack generator |
US7078609B2 (en) * | 1999-10-19 | 2006-07-18 | Medialab Solutions Llc | Interactive digital music recorder and player |
US7176372B2 (en) * | 1999-10-19 | 2007-02-13 | Medialab Solutions Llc | Interactive digital music recorder and player |
US6757573B1 (en) | 1999-11-02 | 2004-06-29 | Microsoft Corporation | Method and system for authoring a soundscape for a media application |
JP3680691B2 (en) | 2000-03-15 | 2005-08-10 | ヤマハ株式会社 | Remix device and storage medium |
JP4389330B2 (en) | 2000-03-22 | 2009-12-24 | ヤマハ株式会社 | Performance position detection method and score display device |
US6882891B2 (en) | 2000-12-06 | 2005-04-19 | Microsoft Corporation | Methods and systems for mixing digital audio signals |
US6664460B1 (en) | 2001-01-05 | 2003-12-16 | Harman International Industries, Incorporated | System for customizing musical effects using digital signal processing techniques |
JP4423790B2 (en) | 2001-01-11 | 2010-03-03 | ソニー株式会社 | Demonstration system, demonstration method via network |
US7107110B2 (en) | 2001-03-05 | 2006-09-12 | Microsoft Corporation | Audio buffers with audio effects |
US6888999B2 (en) | 2001-03-16 | 2005-05-03 | Magix Ag | Method of remixing digital information |
JP4267925B2 (en) * | 2001-04-09 | 2009-05-27 | ミュージックプレイグラウンド・インコーポレーテッド | Medium for storing multipart audio performances by interactive playback |
US7119267B2 (en) * | 2001-06-15 | 2006-10-10 | Yamaha Corporation | Portable mixing recorder and method and program for controlling the same |
US7314994B2 (en) * | 2001-11-19 | 2008-01-01 | Ricoh Company, Ltd. | Music processing printer |
EP1326228B1 (en) * | 2002-01-04 | 2016-03-23 | MediaLab Solutions LLC | Systems and methods for creating, modifying, interacting with and playing musical compositions |
JP3726757B2 (en) | 2002-02-07 | 2005-12-14 | ヤマハ株式会社 | Effect imparting method, effect imparting apparatus, and program |
US7742609B2 (en) | 2002-04-08 | 2010-06-22 | Gibson Guitar Corp. | Live performance audio mixing system with simplified user interface |
US6967275B2 (en) | 2002-06-25 | 2005-11-22 | Irobot Corporation | Song-matching system and method |
US20040016338A1 (en) | 2002-07-24 | 2004-01-29 | Texas Instruments Incorporated | System and method for digitally processing one or more audio signals |
AU2002334362A1 (en) | 2002-09-12 | 2004-04-30 | Musicraft Ltd | Computer-generated expression in music production |
AU2003275089A1 (en) | 2002-09-19 | 2004-04-08 | William B. Hudak | Systems and methods for creation and playback performance |
US7596234B2 (en) | 2003-06-26 | 2009-09-29 | Microsoft Corporation | Method and apparatus for playback of audio files |
JP3960278B2 (en) | 2003-08-13 | 2007-08-15 | ヤマハ株式会社 | Connection setting program |
ES2325859T3 (en) | 2003-10-08 | 2009-09-22 | Nokia Corporation | AUDIO PROCESSING SYSTEM. |
US20110064233A1 (en) * | 2003-10-09 | 2011-03-17 | James Edwin Van Buskirk | Method, apparatus and system for synthesizing an audio performance using Convolution at Multiple Sample Rates |
US7541536B2 (en) * | 2004-06-03 | 2009-06-02 | Guitouchi Ltd. | Multi-sound effect system including dynamic controller for an amplified guitar |
US8818537B2 (en) | 2004-10-06 | 2014-08-26 | Avid Technology, Inc. | Software monitor and control application for use with a digital audio studio suite |
US7420113B2 (en) | 2004-11-01 | 2008-09-02 | Yamaha Corporation | Rendition style determination apparatus and method |
US20060159291A1 (en) | 2005-01-14 | 2006-07-20 | Fliegler Richard H | Portable multi-functional audio sound system and method therefor |
US7916060B2 (en) * | 2005-01-27 | 2011-03-29 | Electro Industries/Gauge Tech. | Intelligent electronic device having circuitry for noise reduction for analog-to-digital converters |
EP2410681A3 (en) | 2005-03-31 | 2012-05-02 | Yamaha Corporation | Control apparatus for music system comprising a plurality of equipments connected together via network, and integrated software for controlling the music system |
EP1904816A4 (en) | 2005-07-18 | 2014-12-24 | Diego Giuseppe Tognola | A signal process and system |
US7869892B2 (en) | 2005-08-19 | 2011-01-11 | Audiofile Engineering | Audio file editing system and method |
US20070098368A1 (en) | 2005-11-02 | 2007-05-03 | Thomas Carley | Mobile recording studio system |
GB2449823A (en) * | 2006-02-27 | 2008-12-03 | John Grigsby | Controlling and/or configuring an electronic audio recorder, player, processor and/or synthesizer |
EP1841108B1 (en) | 2006-03-28 | 2015-11-18 | Yamaha Corporation | Music processing apparatus and management method therefor |
US7326849B2 (en) * | 2006-04-06 | 2008-02-05 | Fender Musical Instruments Corporation | Foot-operated docking station for electronic modules used with musical instruments |
WO2008064230A2 (en) * | 2006-11-20 | 2008-05-29 | Personics Holdings Inc. | Methods and devices for hearing damage notification and intervention ii |
WO2008121650A1 (en) * | 2007-03-30 | 2008-10-09 | William Henderson | Audio signal processing system for live music performance |
GB2473266A (en) * | 2009-09-07 | 2011-03-09 | Nokia Corp | An improved filter bank |
-
2008
- 2008-03-26 WO PCT/US2008/058262 patent/WO2008121650A1/en active Application Filing
- 2008-03-26 US US12/055,903 patent/US8180063B2/en active Active
-
2012
- 2012-05-14 US US13/471,225 patent/US20120269357A1/en not_active Abandoned
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030091204A1 (en) * | 1992-04-27 | 2003-05-15 | Gibson David A. | Method and apparatus for using visual images to mix sound |
US5850455A (en) * | 1996-06-18 | 1998-12-15 | Extreme Audio Reality, Inc. | Discrete dynamic positioning of audio signals in a 360° environment |
US6931134B1 (en) * | 1998-07-28 | 2005-08-16 | James K. Waller, Jr. | Multi-dimensional processor and multi-dimensional audio processor system |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9177540B2 (en) | 2009-06-01 | 2015-11-03 | Music Mastermind, Inc. | System and method for conforming an audio input to a musical key |
US9251776B2 (en) | 2009-06-01 | 2016-02-02 | Zya, Inc. | System and method creating harmonizing tracks for an audio input |
US9263021B2 (en) | 2009-06-01 | 2016-02-16 | Zya, Inc. | Method for generating a musical compilation track from multiple takes |
US9293127B2 (en) | 2009-06-01 | 2016-03-22 | Zya, Inc. | System and method for assisting a user to create musical compositions |
US9310959B2 (en) | 2009-06-01 | 2016-04-12 | Zya, Inc. | System and method for enhancing audio |
US20130319208A1 (en) * | 2011-03-15 | 2013-12-05 | David Forrest | Musical learning and interaction through shapes |
US9147386B2 (en) * | 2011-03-15 | 2015-09-29 | David Forrest | Musical learning and interaction through shapes |
US9378652B2 (en) * | 2011-03-15 | 2016-06-28 | David Forrest | Musical learning and interaction through shapes |
WO2015066204A1 (en) * | 2013-10-30 | 2015-05-07 | Music Mastermind, Inc. | System and method for enhancing audio, conforming an audio input to a musical key, and creating harmonizing tracks for an audio input |
Also Published As
Publication number | Publication date |
---|---|
US20080240454A1 (en) | 2008-10-02 |
WO2008121650A1 (en) | 2008-10-09 |
US8180063B2 (en) | 2012-05-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8180063B2 (en) | Audio signal processing system for live music performance | |
US11132984B2 (en) | Automatic multi-channel music mix from multiple audio stems | |
US20060159291A1 (en) | Portable multi-functional audio sound system and method therefor | |
US6818815B2 (en) | Phonograph turntable with MIDI output | |
US20020144587A1 (en) | Virtual music system | |
US6881891B1 (en) | Multi-channel nonlinear processing of a single musical instrument signal | |
JP4097726B2 (en) | Electronic sound equipment | |
d'Escrivan | Music technology | |
CN1993730A (en) | Automatic playing and recording apparatus for acoustic/electric guitar | |
EP3709290B1 (en) | Acoustic device and acoustic control program | |
US20030164085A1 (en) | Surround sound system | |
Ciesla | More Effects and Digital Audio Workstations | |
White | Desktop Digital Studio | |
Derakhshan | Non-Directed Time | |
JP2017073590A (en) | Program for sound signal processing device | |
Ciesla | MIDI and Composing in the Digital Age | |
Hildebrand | Computer Aided Composition, Timbral Synthesis, & Electro-Acoustic Super-Instruments: An Exploration of Heaviness in Electro-Acoustic Chamber Music | |
Mead et al. | Composing, Recording and Producing with Historical Equipment and Instrument Models | |
MIDI | Products of Interest | |
WO2023235676A1 (en) | Enhanced music delivery system with metadata | |
US9418641B2 (en) | Swap Divisi process | |
Kortier | The MIDI harp: what is a MIDI harp? | |
Hansen | The Fabulous Party Boys and Adapting the Tuba for Modern Popular Contexts | |
Rincón | Music technology | |
Aramburu | Expanding guitar production techniques: building the guitar application toolkit (GATK) |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: AUDIOFILE ENGINEERING LLC, MINNESOTA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HENDERSON, WILLIAM;REEL/FRAME:028543/0751 Effective date: 20110504 |
|
AS | Assignment |
Owner name: AUDIOFILE ENGINEERING LLC, MINNESOTA Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE RECEIVING PARTY(IES) ADDRESS PREVIOUSLY RECORDED ON REEL 028543 FRAME 0751. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNOR:HENDERSON, WILLIAM;REEL/FRAME:028719/0905 Effective date: 20110504 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |