US11893211B2 - Display method and display device - Google Patents

Display method and display device Download PDF

Info

Publication number
US11893211B2
US11893211B2 US17/750,296 US202217750296A US11893211B2 US 11893211 B2 US11893211 B2 US 11893211B2 US 202217750296 A US202217750296 A US 202217750296A US 11893211 B2 US11893211 B2 US 11893211B2
Authority
US
United States
Prior art keywords
display
group
groups
scene
display mode
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US17/750,296
Other languages
English (en)
Other versions
US20220374111A1 (en
Inventor
Kosuke Saito
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Yamaha Corp
Original Assignee
Yamaha Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Yamaha Corp filed Critical Yamaha Corp
Assigned to YAMAHA CORPORATION reassignment YAMAHA CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SAITO, KOSUKE
Publication of US20220374111A1 publication Critical patent/US20220374111A1/en
Application granted granted Critical
Publication of US11893211B2 publication Critical patent/US11893211B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04847Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0482Interaction with lists of selectable items, e.g. menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04886Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/165Management of the audio stream, e.g. setting of volume, audio stream path
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04803Split screen, i.e. subdividing the display area or the window area into separate subareas
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04HBROADCAST COMMUNICATION
    • H04H60/00Arrangements for broadcast applications with a direct linking to broadcast information or broadcast space-time; Broadcast-related systems
    • H04H60/02Arrangements for generating broadcast information; Arrangements for generating broadcast-related information with a direct linking to broadcast information or to broadcast space-time; Arrangements for simultaneous generation of broadcast information and broadcast-related information
    • H04H60/04Studio equipment; Interconnection of studios

Definitions

  • the present invention relates to a data display method and a data display device.
  • content data are grouped and displayed hierarchically, such as by year, month, and day. That is, in the display method of Japanese Patent Number 5086960, content data are displayed on a yearly, monthly, or daily basis.
  • an object of one aspect of the present disclosure is to provide a display method with which it is possible to easily ascertain to which group, out of all groups, certain data (components of a group) belong, and which group the group related to the data is.
  • a display method comprises displaying a list of a plurality of groups in a first area which is a first partial area of a display, and a name of a component belonging to at least one of the plurality of groups in a second area which is a second partial area of the display and is different from the first area.
  • switching a display mode between a first display mode and a second display mode is performed in the second area.
  • the first display mode is a mode in which all of the plurality of groups are displayed and a group, out of the plurality of groups, to which the component belongs is displayed in a display appearance different from that of a different group of the plurality of groups.
  • the second display mode is a mode in which, out of the plurality of groups, only the group to which the component belongs is displayed.
  • FIG. 1 is a block diagram showing a configuration of an audio mixer.
  • FIG. 2 is a function block diagram of signal processing.
  • FIG. 3 is a diagram showing an input patch setting screen.
  • FIG. 4 is a diagram showing a processing configuration of a given input channel 1 .
  • FIG. 5 is a diagram showing a configuration of an operation panel of an audio mixer 1 .
  • FIG. 6 is a diagram showing one example of a scene group management screen in a first display mode.
  • FIG. 7 is a diagram showing one example of a scene group management screen in a second display mode.
  • FIG. 8 is a flowchart showing a display control operation of a CPU 16 .
  • FIG. 9 is a diagram showing one example of the scene group management screen in the second display mode.
  • FIG. 10 is a diagram showing one example of a scene group management screen when switching a display mode between a first display mode and a second display mode for each piece of scene data.
  • FIG. 1 is a block diagram showing a configuration of an audio mixer 1 .
  • the audio mixer 1 is one example of a display device of this disclosure.
  • the audio mixer 1 has a display (display unit) 11 , an operation unit 12 , an audio I/O (Input/Output) 13 , a signal processing unit 14 , a communication interface (I/F) 15 , a CPU 16 , a flash memory 17 , and a RAM (Random Access Memory) 18 .
  • the display 11 , the operation unit 12 , the audio I/O (Input/Output) 13 , the signal processing unit 14 , the communication interface (I/F) 15 , the CPU 16 , the flash memory 17 , and the RAM 18 are connected to each other via a bus 25 .
  • the audio I/O 13 is an interface for receiving input of a sound signal to be processed by the signal processing unit 14 .
  • the audio I/O 13 is provided with an input port such as an analog input port or a digital input port, that receives input of the sound signal. Further, the audio I/O 13 is an interface for outputting a sound signal that has been processed by the signal processing unit 14 .
  • the audio I/O 13 is provided with an output port such as an analog output port or a digital output port, that outputs the sound signal.
  • the communication I/F 15 is an interface for communicating with other devices.
  • an external PC is connected to the communication I/F 15 .
  • the communication I/F 15 sends and receives a sound signal that conforms to a standard, such as Dante (registered trademark), via a network.
  • Dante registered trademark
  • the CPU 16 is an electronic controller (control unit) that controls the operation of the audio mixer 1 .
  • the electronic controller can include one or a plurality of processors and is configured from one or more types of processors, such as a CPU (Central Processing Unit), a GPU (Graphics Processing Unit), a DSP (Digital Signal Processor), an FPGA (Field Programmable Gate Array), an ASIC (Application Specific Integrated Circuit), and the like.
  • the term “electronic controller” as used herein refers to hardware, and does not include a human.
  • the CPU 16 reads a prescribed program stored in the flash memory 19 as a storage unit (computer memory) into the RAM 20 to perform various operations.
  • the CPU 16 functions as a display control unit of the display 11 by the program.
  • the display 11 displays various information in accordance with the control of the CPU 16 .
  • the display 11 is configured from, for example, an LCD (Liquid Crystal Display) or a light-emitting diode (LED).
  • LCD Liquid Crystal Display
  • LED light-emitting diode
  • the operation unit 12 is a user operable input and receives an operation of the audio mixer 1 from a user.
  • the operation unit 12 includes various keys, buttons, rotary encoders, sliders, and the like. Further, the operation unit 12 can be a touch panel overlaid on an LCD serving as the display 11 .
  • the signal processing unit 14 includes a DSP (Digital Signal Processor) for performing various signal processes, such as mixing or effects processing.
  • the signal processing unit 14 performs signal processing, such as mixing or equalizing, of a sound signal supplied from the audio I/O 13 or the communication I/F 15 .
  • the signal processing unit 14 outputs a digital audio signal that has been subjected to signal processing via the audio I/O 13 or the communication I/F 15 .
  • FIG. 2 is a function block diagram of the signal processing. As shown in FIG. 2 , the signal processing is performed by an input patch 301 , an input channel 302 , a bus 303 , an output channel 304 , and an output patch 305 .
  • the input patch 301 inputs audio signals from a plurality of input ports (for example, analog input ports or digital input ports) of the audio I/O 13 , and assigs one of the plurality of ports to at least one of a plurality of input channels (for example, 32 ch).
  • FIG. 3 is a diagram showing an example of a setting screen of the input patch 301 .
  • the setting screen of the input patch 301 is displayed on the display 11 .
  • the setting screen of the input patch 301 displays the channel names and port numbers in matrix format.
  • a touch panel is overlaid on the display 11 . The user selects each port number in the matrix, thereby specifying the channel to be assigned to each port. In this way, the audio signal is supplied to each channel of the input channel 302 .
  • FIG. 4 is a block diagram showing the functional configuration of the signal processing of a given input channel i.
  • each input channel of the input channel 302 performs signal processing of an attenuator (ATT), an equalizer (EQ), a gate (GATE), a compressor (COMP), or the like on the audio signal supplied from the input patch 301 .
  • ATT attenuator
  • EQ equalizer
  • GATE gate
  • COMP compressor
  • the audio signal is subjected to level adjustment in a fader section (FADER) 352 , and then sent to a downstage bus 303 via a pan section (PAN) 353 .
  • the pan section 353 adjusts the balance of the signal supplied to a stereo bus (a 2-channel bus serving as a master output) 3031 of the bus 303 .
  • the audio signal on which the signal processing has been performed is subjected to level adjustment in a feed section (MIX1-MIX16) 355 via a selector (SEL) 354 , and then sent to the downstage bus 303 .
  • the SEL 354 can input, to the downstage feed section 355 , by the user's selecting operation, either a signal output from the signal processing block 351 or a signal after being subjected to level adjustment by the fader section 352 .
  • the feed section 355 determines the level of the signal supplied to each MIX bus 3032 in accordance with each feed amount set by the user.
  • the output channel 304 has, for example, 16 channels. Each of the channels of the output channel 304 applies various signal processing to the audio signal that is input, in the same manner as the input channel. Each channel of the output channel 304 sends the audio signal after signal processing to the output patch 305 .
  • the output patch 305 assigns each channel to one of a plurality of analog output ports or digital output ports. The audio signal after being subjected to signal processing is thus supplied to the audio I/O 13 .
  • the signal processing described above is controlled based on the set values of various parameters.
  • the CPU 16 stores the currently set values (current data) of the various parameters in the RAM 20 .
  • the CPU 16 updates the current data when the user operates the operation unit 12 .
  • FIG. 5 is a diagram showing a configuration of an operation panel of the audio mixer 1 .
  • the display 11 a channel strip 63 , a store button 72 , a recall button 73 , increase/decrease buttons 74 , and the like, are provided on the operation panel of the audio mixer 1 .
  • the display 11 is a display on which a touch panel has been overlaid, which is one aspect of the operation unit 12 , and displays a GUI (Graphical User Interface) screen for receiving user operations.
  • the GUI of the display 11 , the channel strip 63 , the store button 72 , the recall button 73 , and the increase/decrease buttons 74 correspond to the operation unit 12 .
  • the channel strip 63 is an area in which a plurality of operators for receiving an operation with respect to one channel are arranged vertically.
  • FIG. 5 only one fader and one knob are displayed for each channel as the operators, but in practice, numerous knobs, switches, etc., are provided.
  • a plurality (for example, 16) of faders and knobs arranged on the left side of the channel strip 63 correspond to the input channel.
  • the two faders and knobs arranged on the right side are operators that correspond to the master output (2-channel bus).
  • the store button 72 is a button for an instruction to store scene data.
  • the user can thus operate the store button 72 to record (store) the current data in the flash memory 19 as one piece of scene data.
  • a plurality of scene data are stored in the flash memory 19 .
  • the user can operate the increase/decrease buttons 74 to select, from among the plurality of scene data, the scene data to be saved or recalled.
  • the user can operate the recall button 73 to recall the necessary scene data in order to recall the set values of the various parameters.
  • These buttons can be configured by a GUI using the touch panel overlaid on the display 11 .
  • the audio mixer 1 of the present embodiment can group and manage the plurality of scene data.
  • the user designates scene data of the same song, the same singer, the same performer, or the like, as the same scene group.
  • the user edits and stores certain scene data and carries out an operation to copy the stored scene data as certain scene data.
  • the user can collectively copy certain arbitrary scene data as a grouped plurality of scene data.
  • the user can collectively copy a certain arbitrary parameter as a grouped plurality of scene data.
  • FIG. 6 is a diagram showing one example of a scene group management screen in a first display mode, which is displayed on the display 11 .
  • the scene group management screen 70 has a first area 81 and a second area 82 .
  • the first display mode is a mode in which, in the second area 82 , all of the plurality of groups are displayed and one or more groups, out of the plurality of groups, to which the scene data belong are displayed in a display appearance different from that of at one or more different groups of the plurality of groups.
  • the first area 81 displays a list of a plurality of scene groups.
  • the first area 81 includes a group selection tab 51 , a group selection tab 52 , and a plurality of group selection icons 71 .
  • a list of scene groups 1-16 is displayed in the first area.
  • a list of scene groups 17-32 is displayed in the first area.
  • the number of scene groups is not limited to the number shown in the present embodiment.
  • the list of scene groups 1-16 is displayed in the first area.
  • the user selects any scene group from the group selection icons 71 .
  • the user has selected scene group 1.
  • the scene group 1 includes scene data 1 and scene data 3.
  • the second area 82 has a title display portion 91 that displays at least one a name (TITLE) of the scene data corresponding to a component belonging to the scene group, and a group display portion 92 that displays all of the scene groups.
  • TITLE a name
  • the group display portion 92 further displays the scene group(s) to which each piece of the scene data belongs in a display appearance that is different from that of the other scene groups.
  • the group display portion 92 inverts color of the scene groups to which the scene data belong.
  • the display appearance can be any appearance, such as changing of color or hatching.
  • the user has selected scene group 1 in the first area 81 .
  • the scene group 1 includes scene data 1 and scene data 3. Therefore, in the second area 82 , the scene data 1 and the scene data 3 appear emphasized. Further, the scene data 1 belong to the scene group 1, the scene group 3, and the scene group 4. As a result, in the group display portion 92 for the scene data 1, the scene group 1, the scene group 3, and the scene group 4 are color inverted.
  • the scene data 2 belong to the scene group 2, the scene group 3, and the scene group 4. As a result, in the group display portion 92 for the scene data 2, the scene group 2, the scene group 3, and the scene group 4 are color inverted.
  • the scene data 3 belong to the scene group 1, the scene group 2, and the scene group 4.
  • the scene group 1, the scene group 2, and the scene group 4 are color inverted.
  • the scene data 4 belong to the scene group 4.
  • the scene group 4 is color inverted in the group display portion 92 for the scene data 4.
  • the user can easily ascertain to which scene group, out of all the scene groups, each piece of the scene data belongs.
  • FIG. 7 is a diagram showing one example of a scene group management screen in a second display mode.
  • the second display mode is a mode in which, out of the plurality of groups, only one or more groups to which the scene data belong are displayed in the second area 82 .
  • the configurations that are the same as those in FIG. 6 have been assigned to the same reference numerals, and their descriptions have been omitted.
  • the second area 82 has a belonging group display portion 93 instead of the group display portion 92 that displays all the scene groups.
  • the belonging group display portion 93 displays, out of the plurality of scene groups, only the scene group(s) to which each piece of the scene data belongs.
  • the scene data 1 belong to the scene group 1, the scene group 3, and the scene group 4.
  • the belonging group display portion 93 for the scene data only the scene group 1, the scene group 3, and the scene group 4 are displayed.
  • the belonging group display portion 93 for the scene data 2 only the scene group 2, the scene group 3, and the scene group 4 are displayed.
  • the belonging group display portion 93 for the scene data 3 only the scene group 1, the scene group 2, and the scene group 4 are displayed.
  • the belonging group display portion 93 for the scene data 4 only the scene group 4 is displayed.
  • the belonging group display portion 93 displays a name of the scene group.
  • the belonging group display portion 93 for the scene data 1 only the name (SONG 1 ) of the scene group 1, the name (SONG 3 ) of the scene group 3, and the name (SONG 4 ) of the scene group 4 are displayed.
  • the user can easily and specifically grasp which scene group(s) the scene group(s) that corresponds to each piece of the scene data is.
  • the CPU 16 when the CPU 16 displays the list of the plurality of scene groups in the first area 81 , and the names of the scene data belonging to at least one of the plurality of groups in the second area 82 in the display 11 , the CPU 16 switches a display mode between the first display mode and the second display mode in the second area 82 .
  • FIG. 8 is a flowchart showing a display control operation of the CPU 16 .
  • the CPU 16 determines whether the display mode is the first display mode or the second display mode (S 11 ). If it is determined to be the first display mode (S 11 : Yes), the CPU 16 displays all of the scene groups, and displays the scene group(s) to which each piece of scene data belongs in a display appearance that is different from that of the other scene groups (S 12 ). On the other hand, if it is determined to be the second display mode (S 11 : No), the CPU 16 displays, out of the plurality of scene groups, only the scene group(s) to which each piece of the scene data belongs (S 13 ).
  • the CPU 16 determines whether mode switching has been received (S 14 ).
  • the mode switching is received by a touch operation with respect to the title display portion 91 of the second area 82 , for example.
  • the manner in which mode switching is received is not limited to this example.
  • the audio mixer 1 can comprise a mode switching button in the configuration of the operation panel shown in FIG. 5 and receive the mode switching via the mode switching button.
  • the CPU 16 switches the display mode (S 15 ). If the current display mode is the first display mode, the CPU 16 switches to the second display mode. If the current display mode is the second display mode, the CPU 16 switches to the first display mode.
  • the audio mixer 1 in the first display mode, the user can easily ascertain to which scene groups, out of all the scene groups, each piece of the scene data belongs. Further, in the second display mode, the user can easily ascertain which scene group a scene group related to each piece of the scene data is. As a result of the audio mixer 1 switching between the first display mode and the second display mode, the user can easily ascertain to which scene group, out of all the scene groups, each piece of the scene data belongs, and for each piece of the scene data, easily ascertain which scene group a scene group related to each piece of the scene data is, even in a small display area.
  • the second display mode if the number of scene groups to which each piece of the scene data belongs is greater than a prescribed number, the displaying of the scene group that exceeds the prescribed number can be omitted, and an indication that such omission is made can be displayed.
  • the groups to which the scene data 1 belong are the scene group 1, the scene group 2, the scene group 3, and the scene group 4, the CPU 16 omits the display of the scene group 4, and displays an omitted display icon 94 .
  • the prescribed number is three so that three scene groups are displayed in the belonging group display portion 93 , but the prescribed number can be one, two, or four or more.
  • the audio mixer 1 can omit the display of some scene groups in order to display a large amount of information even if the number of scene groups to which each piece of scene data belongs increases.
  • the CPU 16 further functions as a reception unit that receives a change operation of the scene group to which each piece of the scene data belongs in response to, for example, a touch operation with respect to the operation unit 12 . Operation to change the scene group to which each piece of the scene data belongs is preferably carried out in a separate setting screen as the operation unit 12 . Accordingly, if the CPU 16 receives a touch operation with respect to the second area 82 , the CPU 16 preferably does not accept a change of the scene group to which each piece of the scene data belongs.
  • the operation to change the scene group to which each piece of the scene data belongs can be carried out on a separate setting screen, even if a touch operation with respect to the second area 82 is received, the CPU 16 does not change the display contents of the first area 81 . As a result, the user does not confuse the setting screen of the input patch 301 shown in FIG. 3 with the scene group management screen.
  • the first display mode and the second display mode are switched for all scene data.
  • the switching between the first display mode and the second display mode can be performed in accordance with each piece of the scene data.
  • the CPU 16 can display only the scene data 1 in the second display mode and display the other scene data 2-4 in the first display mode.
  • the CPU 16 receives a touch operation with respect to the title display portion 91 of each piece of the scene data, and switches the display mode for each piece of the scene data.
  • the scene data are illustrated as one example of components, and the scene groups are shown as plurality of groups.
  • the components of this disclosure are not limited to the scene data, and the groups are not limited to the scene groups.
  • the components can be the input channels of the input channel 302 .
  • the groups can be MIX buses.
  • the CPU 16 displays a list (for example, numbers between 1-16) of all the MIX buses, and displays the MIX buses (the numbers of the MIX buses) for which the transmission of sound signals from each input channel is set to ON in a display appearance that is different than that of the numbers of the other MIX buses.
  • the CPU 16 displays the number of the MIX buses for which the gain value of the volume of the input channel is set to be larger than ⁇ in a display appearance that is different from that of the numbers of the other MIX buses.
  • the CPU 16 displays only the numbers of the MIX buses for which the transmission of the input channel is set to ON, or for which the gain value is set to be larger than ⁇ .
  • the CPU 16 can display the names of the MIX buses for which the transmission is set to ON, or for which the gain value is set to be larger than ⁇ .
  • the components can be MIX buses, and the groups to which each of the components belongs can be the input channels of the input channel 302 .
  • the CPU 16 can display the names of the MIX buses in the tide display portion 91 , the numbers of the input channels in the group display portion 92 in the first display mode, and the names of the input channels in the belonging group display portion 93 in the second display mode.
  • the components can be input channels and the groups can be auto mixers.
  • the components can be input channels and the groups can be channel groups.
  • the components can be faders and the groups can be DCA (Digital Controlled Amplifier) faders.
  • the components can be channel groups, and the groups to which each of the components belongs can be input channels.
  • the CPU 16 can display the names of the channel groups in the title display portion 91 , the numbers of the input channels in the group display portion 92 in the first display mode, and the names of the input channels in the belonging group display portion 93 in the second display mode.
  • the components are constituent elements related to audio processing, but the components may be constituent elements of other types of processing besides audio processing.
  • the components may be photograph data and the groups may be specific people.
  • a CPU of an information processing device displays a list (for example, the initials) of all persons and displays the person to which each photograph belongs in a display appearance that is different from that of the other persons.
  • the CPU displays only the person to which each photograph belongs.
  • the CPU may display details related to each person, such as their full name.
  • the components may be parameters of equipment in a factory, and the groups may be buildings.
  • the CPU displays all of the buildings and the building (for example, its number) to which each piece of equipment belongs in a display appearance that is different from that of the other buildings.
  • the CPU displays only the building to which each equipment belongs.
  • the CPU may display details, such as the names of buildings.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Circuit For Audible Band Transducer (AREA)
US17/750,296 2021-05-24 2022-05-20 Display method and display device Active US11893211B2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2021086980A JP2022180072A (ja) 2021-05-24 2021-05-24 表示方法および表示装置
JP2021-086980 2021-05-24

Publications (2)

Publication Number Publication Date
US20220374111A1 US20220374111A1 (en) 2022-11-24
US11893211B2 true US11893211B2 (en) 2024-02-06

Family

ID=84102766

Family Applications (1)

Application Number Title Priority Date Filing Date
US17/750,296 Active US11893211B2 (en) 2021-05-24 2022-05-20 Display method and display device

Country Status (2)

Country Link
US (1) US11893211B2 (ja)
JP (1) JP2022180072A (ja)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090028359A1 (en) * 2007-07-23 2009-01-29 Yamaha Corporation Digital Mixer
US20090307633A1 (en) * 2008-06-06 2009-12-10 Apple Inc. Acceleration navigation of media device displays
JP5086960B2 (ja) 2008-09-26 2012-11-28 キヤノン株式会社 データ閲覧装置及びデータ閲覧装置の制御方法
US20140293349A1 (en) 2013-03-29 2014-10-02 Brother Kogyo Kabushiki Kaisha Image display control apparatus, image display control method, and non-transitory storage medium storing instructions executable by image display control apparatus
US20170288798A1 (en) * 2016-04-04 2017-10-05 Yamaha Corporation Method for controlling audio signal processing device, audio signal processing device, and storage medium
US20190222332A1 (en) * 2016-09-28 2019-07-18 Yamaha Corporation Mixer, Control Method of Mixer, and Program
US20200092642A1 (en) * 2018-09-13 2020-03-19 Yamaha Corporation Audio signal processing apparatus, audio signal processing method, and storage medium

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090028359A1 (en) * 2007-07-23 2009-01-29 Yamaha Corporation Digital Mixer
US20090307633A1 (en) * 2008-06-06 2009-12-10 Apple Inc. Acceleration navigation of media device displays
JP5086960B2 (ja) 2008-09-26 2012-11-28 キヤノン株式会社 データ閲覧装置及びデータ閲覧装置の制御方法
US20140293349A1 (en) 2013-03-29 2014-10-02 Brother Kogyo Kabushiki Kaisha Image display control apparatus, image display control method, and non-transitory storage medium storing instructions executable by image display control apparatus
JP6115256B2 (ja) 2013-03-29 2017-04-19 ブラザー工業株式会社 画像表示制御装置、画像表示制御方法および画像表示制御プログラム
US20170288798A1 (en) * 2016-04-04 2017-10-05 Yamaha Corporation Method for controlling audio signal processing device, audio signal processing device, and storage medium
US20190222332A1 (en) * 2016-09-28 2019-07-18 Yamaha Corporation Mixer, Control Method of Mixer, and Program
US20200092642A1 (en) * 2018-09-13 2020-03-19 Yamaha Corporation Audio signal processing apparatus, audio signal processing method, and storage medium

Also Published As

Publication number Publication date
JP2022180072A (ja) 2022-12-06
US20220374111A1 (en) 2022-11-24

Similar Documents

Publication Publication Date Title
US8098850B2 (en) Digital mixer
US9570058B2 (en) Audio signal processing device and parameter adjusting method
US8050427B2 (en) Digital mixer and display control method therefor
US20100309153A1 (en) Audio mixing apparatus
JP4591696B2 (ja) デジタルミキサおよびプログラム
US8170240B2 (en) Audio device
US10536231B2 (en) Mixer, control method of mixer, and program
EP2228923A2 (en) Audio signal mixer
JP5387472B2 (ja) ミキシング装置
US11893211B2 (en) Display method and display device
JP5182188B2 (ja) オーディオミキシング装置及びオーディオミキシング用プログラム
US11178502B2 (en) Audio mixer and control method of audio mixer
JP5182187B2 (ja) オーディオミキシング装置及びオーディオミキシング用プログラム
JP3772803B2 (ja) 信号処理装置及び該装置の制御プログラム
US10887688B2 (en) Audio signal processing apparatus, audio signal processing method, and storage medium that update currently used parameter upon receiving change operation instruction
US11188291B2 (en) Audio signal processing apparatus, method for processing audio signal, and storage medium storing program
US10534572B2 (en) Control device, control method, and storage medium storing a program
US20230315379A1 (en) Audio mixer, and method of controlling audio mixer
CN110992918B (zh) 声音信号处理装置、声音信号处理方法以及存储介质
US20230017387A1 (en) Signal processing method and signal processing device
JP2007074668A (ja) デジタルミキサ
JP2018142931A (ja) 音制御システム及び方法
JP2014187668A (ja) 音響信号処理装置

Legal Events

Date Code Title Description
AS Assignment

Owner name: YAMAHA CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SAITO, KOSUKE;REEL/FRAME:059977/0231

Effective date: 20220513

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE