WO2016051693A1 - Service monitoring system and service monitoring method - Google Patents

Service monitoring system and service monitoring method Download PDF

Info

Publication number
WO2016051693A1
WO2016051693A1 PCT/JP2015/004661 JP2015004661W WO2016051693A1 WO 2016051693 A1 WO2016051693 A1 WO 2016051693A1 JP 2015004661 W JP2015004661 W JP 2015004661W WO 2016051693 A1 WO2016051693 A1 WO 2016051693A1
Authority
WO
WIPO (PCT)
Prior art keywords
customer service
utterance
customer
employee
predetermined
Prior art date
Application number
PCT/JP2015/004661
Other languages
French (fr)
Japanese (ja)
Inventor
久裕 田中
昭年 泉
正成 宮本
信一 重永
亮太 藤井
広志 田中
寿嗣 辻
Original Assignee
パナソニックIpマネジメント株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP2015171556A external-priority patent/JP6624368B2/en
Application filed by パナソニックIpマネジメント株式会社 filed Critical パナソニックIpマネジメント株式会社
Priority to US15/513,622 priority Critical patent/US10706448B2/en
Publication of WO2016051693A1 publication Critical patent/WO2016051693A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/06Buying, selling or leasing transactions
    • GPHYSICS
    • G07CHECKING-DEVICES
    • G07GREGISTERING THE RECEIPT OF CASH, VALUABLES, OR TOKENS
    • G07G1/00Cash registers
    • G07G1/12Cash registers electronically operated
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis

Definitions

  • the present disclosure relates to a customer service monitoring system and a customer service monitoring method for monitoring a situation during customer service using an employee's voice.
  • a method for measuring customer satisfaction is a method of covering investigation by an investigator.
  • the survey results obtained depending on the environment where the store survey is conducted are different, and the improvement advice based on the survey results may not be accurate.
  • Differences in the environment in which surveys are conducted include, for example, the degree of congestion and the number of store staff (employees) in the stores subject to store surveys for each time period during which the store surveys were conducted, the survey staff's survey technology, and the store staff receiving the store surveys This is the difference in the skill level of customer service (for example, the difference in working years).
  • Patent Document 1 a service evaluation diagnosis system that presents advice information based on the reality based on the corrected survey result after correcting the influence on the survey result based on the difference in the environment in which the store survey is performed.
  • a survey result of a store survey (for example, a store survey) is input by an investigator who operates a portable information terminal.
  • Another example of a method for measuring customer satisfaction is to recognize store clerk's emotion and customer's emotion based on the store clerk's voice and customer's voice included in the store clerk's customer conversation, and based on this recognition result.
  • a customer service data recording device for calculating store clerk satisfaction data and customer satisfaction data see, for example, Patent Document 2.
  • the customer service data recording device shown in Patent Document 2 records customer service data in which a store clerk satisfaction data, a customer satisfaction data, and a sales record of a store clerk for a customer are associated with each other in a database.
  • Patent Document 2 unlike Patent Document 1, an input operation by an investigator who operates a portable information terminal is not necessary.
  • This disclosure does not use human resources like investigators, and protects customer privacy extensively, and monitors customer utterances during various customer service events for customers in the store. It is an object of the present invention to provide a customer monitoring system and a customer monitoring method for accurately and objectively evaluating the situation.
  • the customer service monitoring system includes a sound collection unit that collects employee voices in a predetermined sound collection area, a first storage unit that stores customer event data including determination conditions for each predetermined customer event, and a predetermined storage unit.
  • a second storage unit that stores the terminal operation history data indicating the operation history of the employee for the business terminal and the voice data of the employee collected by the sound collection unit in association with each other, and the first storage unit
  • a detection unit Based on the customer service event data stored in and the terminal operation history data stored in the second storage unit, a detection unit for detecting the customer service event of the employee, and the detection unit detected by the detection unit
  • a service utterance evaluation value corresponding to a predetermined utterance keyword at the time of operation of the business terminal is calculated based on the voice data of the employee stored in the second storage unit.
  • the service utterance evaluation value calculated by the determination unit and the calculation unit is stored in association with the employee's voice data specified by the employee identification information, the customer service position and the customer service time.
  • an output unit is
  • the customer service monitoring method of the present disclosure is a customer service monitoring method in a customer service monitoring system including a sound collection unit that collects an employee's voice in a predetermined sound collection region, and includes a determination condition for each predetermined customer service event.
  • the customer service event data is stored in the first storage unit, and the terminal operation history data indicating the operation history of the employee with respect to the predetermined business terminal is associated with the voice data of the employee collected by the sound collecting unit.
  • the customer service event is detected based on the customer service event data stored in the second memory and stored in the first memory and the terminal operation history data stored in the second memory.
  • a predetermined utterance keyword at the time of operating the business terminal The customer service utterance evaluation value corresponding to the employee is calculated, and the calculated customer service utterance evaluation value is associated with the employee's voice data specified by the employee identification information, the employee service position, and the customer service time.
  • FIG. 1 is a diagram illustrating an example of an image in a store where a customer service monitoring system according to the present embodiment is installed.
  • FIG. 2 is a block diagram illustrating a first system configuration example of the customer service monitoring system according to the present embodiment.
  • FIG. 3 is a block diagram illustrating a second system configuration example of the customer service monitoring system according to the present embodiment.
  • FIG. 4 is a block diagram illustrating a third system configuration example of the customer service monitoring system according to the present embodiment.
  • FIG. 5 is a block diagram illustrating a fourth system configuration example of the customer service monitoring system according to the present embodiment.
  • FIG. 6 is a flowchart for explaining an example of the overall operation procedure of the customer service monitoring system of this embodiment.
  • FIG. 1 is a diagram illustrating an example of an image in a store where a customer service monitoring system according to the present embodiment is installed.
  • FIG. 2 is a block diagram illustrating a first system configuration example of the customer service monitoring system according to the present embodiment.
  • FIG. 3 is
  • FIG. 7 is a flowchart for explaining an example of a detailed operation procedure of the event detection availability determination process.
  • FIG. 8 is a flowchart for explaining an example of a detailed operation procedure of the customer service event detection process.
  • FIG. 9 is a flowchart illustrating an example of a detailed operation procedure of customer service event information processing.
  • FIG. 10 is a diagram illustrating an example of the customer service situation DB.
  • FIG. 11 is a diagram illustrating an example of a customer service event information DB corresponding to the customer service monitoring system illustrated in FIG. 2.
  • FIG. 12 is a diagram illustrating an example of a customer service event information DB corresponding to the customer service monitoring system illustrated in FIG. 3.
  • FIG. 13 is a diagram illustrating an example of a customer service event information DB corresponding to the customer service monitoring system illustrated in FIG. 4.
  • FIG. 14 is a diagram showing an example of a customer service event information DB corresponding to the customer service monitoring system shown in FIG.
  • FIG. 15 is a flowchart for explaining an example of the operation procedure of the customer service utterance evaluation process.
  • FIG. 16 is a flowchart for explaining an example of the operation procedure of the noise level determination process.
  • FIG. 17 is a flowchart for explaining an example of the operation procedure of the service keyword utterance determination process.
  • FIG. 18 is a flowchart for explaining an example of the operation procedure of the scoring process.
  • FIG. 19A is a flowchart for explaining an example of an operation procedure of speech length determination processing.
  • FIG. 19A is a flowchart for explaining an example of an operation procedure of speech length determination processing.
  • FIG. 19B is a flowchart illustrating an example of an operation procedure of a frequency characteristic determination process.
  • FIG. 20 is a diagram illustrating a specific example of an utterance length determination process using model voice data.
  • FIG. 21 is a diagram showing a specific example of the frequency characteristic determination process using the fundamental frequency of each phoneme of the model voice data.
  • FIG. 22A is a diagram illustrating an example of an utterance assumption keyword table constituting a part of the customer service utterance evaluation DB.
  • FIG. 22B is a diagram illustrating an example of a service utterance model list that forms part of the service utterance evaluation DB.
  • FIG. 23 is a flowchart illustrating an example of an operation procedure of a browsing process by a limited viewer or a service utterance evaluation value correction process.
  • FIG. 24 is a flowchart for explaining an example of a detailed operation procedure of the service utterance evaluation value correction process.
  • FIG. 25 is a flowchart for explaining the continuation of the detailed operation procedure of the modification process of the customer service utterance evaluation value shown in FIG.
  • FIG. 26A is a diagram illustrating an example of a viewer DB.
  • FIG. 26B is a diagram illustrating an example of a customer service DB.
  • FIG. 27 is a diagram illustrating an example of a login screen to the customer service status DB to be browsed in the customer service monitoring system.
  • FIG. 28 is a diagram illustrating an example of a totaling result of customer service utterance evaluation values of all customer service customers per day as a customer service status display screen.
  • FIG. 24 is a flowchart for explaining an example of a detailed operation procedure of the service utterance evaluation value correction process.
  • FIG. 25 is a flowchart for explaining the continuation of the detailed operation procedure of the modification process of the customer service utterance evaluation value
  • FIG. 29 is a diagram illustrating an example of a totaling result of customer service utterance evaluation values of all customers with respect to customers in a day time zone.
  • FIG. 30A is a diagram illustrating an example of a totaling result of customer utterance evaluation values of one customer for each time slot of one day.
  • FIG. 30B is a diagram illustrating an example of a total result of customer service utterance evaluation values for each customer service per day.
  • FIG. 31 is a diagram illustrating an example of a totaling result of customer service utterance evaluation values for each store on a day.
  • FIG. 32 is a diagram illustrating a specific example of each record displayed on the detail display screen of the customer service situation DB.
  • FIG. 33 is a diagram illustrating an example of a modification operation for the customer utterance evaluation value of a specific record displayed on the detail display screen of the customer service status DB.
  • FIG. 34 is a diagram illustrating an example of the customer utterance evaluation value after correction of a specific record displayed on the detail display screen of the customer service situation DB.
  • FIG. 35 is a diagram illustrating an example of a customer service position correcting operation on the customer service status preview screen.
  • FIG. 36 is a diagram illustrating an example of coordinates of the customer service position after correction of a specific record displayed on the detail display screen of the customer service DB.
  • FIG. 37 is a diagram illustrating an example of the relationship between the customer microphone device, the microphone device, the microphone array device, and the privacy protection mark.
  • FIG. 38 is an explanatory diagram illustrating an example of shift cut-out processing of monitoring data.
  • FIG. 39 is an explanatory diagram illustrating an example of a variation of the monitoring data cut-out process.
  • FIG. 40 is a flowchart for explaining another example of the detailed operation procedure of customer service event information processing.
  • FIG. 41 is a diagram illustrating an example of an utterance assumption section table for each service event that constitutes a part of the service utterance evaluation DB.
  • the customer service monitoring system is installed in a store (for example, a retail store, a wholesale store, a department store, a convenience store, a supermarket, a restaurant, or a bank) where customer service is performed, and serves a customer of a store clerk (employee) to a customer.
  • a store for example, a retail store, a wholesale store, a department store, a convenience store, a supermarket, a restaurant, or a bank
  • customer service serves a customer of a store clerk (employee) to a customer.
  • the situation is monitored, and the customer service attitude (hospitality) of the store clerk who performs various customer service events (for example, store entrance / exit greetings, accounting start greetings, etc., see below for details) is objectively evaluated.
  • a quantitative index (value) as a result of objectively evaluating the customer service attitude (customer service situation) of the store clerk in the customer service monitoring system is referred to as a “customer utterance evaluation value”.
  • this embodiment demonstrates the evaluation of the customer service situation with respect to the customer of the clerk, the customer service monitoring system and the customer service monitoring method according to the present disclosure are a customer service to an employee other than the clerk (for example, a bank clerk, an employee, etc.) It is also applicable to situational assessment.
  • this indication is a method including each operation
  • FIG. 1 is a diagram illustrating an example of an image in a store where a customer service monitoring system 100 according to the present embodiment is installed.
  • a customer service monitoring system 100 for example, in a store, two employees each serve a plurality of customers while operating the POS terminal 5 at a cashier counter in which two POS terminals 5 are installed.
  • At the store at least one microphone device M1 for picking up sound around the cashier counter, which is a sound pickup area, is installed, and at least one camera device C1 for taking an image so as to include the cashier counter is included.
  • a sensor device S1 for detecting the entrance and exit of the customer to the store is installed near the entrance of the store.
  • the customer service evaluation device 3 for example, PC :
  • the customer service evaluation device 3 for example, PC :
  • the customer service evaluation device 3 for example, PC :
  • the customer service evaluation device 3 for example, PC :
  • the customer service evaluation device 3 for example, PC :
  • the customer service situation in the backyard for example, monitoring room
  • the store manager for example, store manager
  • the customer service monitoring system 100 detects, for example, data indicating the operation history of a store clerk who operates the POS terminal 5 (hereinafter referred to as “POS operation history data”) and presence / absence of a customer service event for each of various customer service events to be described later.
  • a customer service event by a store clerk is detected based on customer service event data including a customer service event determination condition.
  • the POS operation history data includes, for example, a customer ID of the customer (store clerk) (a store clerk identification information), a customer ID input operation history for inputting the customer ID into the POS terminal 5, and the age of the customer (customer). Includes input history, transaction completion operation history, and the like.
  • a predetermined utterance when operating the POS terminal 5 A service utterance evaluation value corresponding to the assumed keyword (utterance keyword) (in other words, a service utterance evaluation value for a clerk who utters the utterance assumed keyword) is calculated.
  • the customer service monitoring system 100 stores the calculated customer service utterance evaluation value in association with the clerk's voice data specified by the clerk's identification information, the clerk's customer service position, and the customer service time.
  • the “salesperson (employee) voice data” here refers to, for example, voice recorded by the microphone device M1, the customer microphone device SM1, or the microphone array device AM1 of the voice uttered by the store clerk at the position of the store clerk. Data.
  • FIG. 2 is a block diagram illustrating a first system configuration example of the customer service monitoring system 100 according to the present embodiment.
  • FIG. 3 is a block diagram illustrating a second system configuration example of the customer service monitoring system 100A according to the present embodiment.
  • FIG. 4 is a block diagram illustrating a third system configuration example of the customer service monitoring system 100B according to the present embodiment.
  • FIG. 5 is a block diagram illustrating a fourth system configuration example of the customer service monitoring system 100C according to the present embodiment. 2 to FIG. 5, the customer service monitoring system 100 shown in FIG.
  • FIG. 2 will be described in detail, and the structures of the customer service monitoring systems 100A, 100B, and 100C shown in FIG. 3, FIG. 4, and FIG. Differences from the customer service monitoring system 100 will be described, and the same components will be denoted by the same reference numerals, and description thereof will be simplified or omitted.
  • the customer service monitoring system 100 shown in FIG. 2 includes at least one microphone device M1,..., ML (L is an integer of 1 or more; the same applies hereinafter), a management server 2, a customer service evaluation device 3, a recorder device 4, The POS terminal 5 is included.
  • a network NW may be a wired network (for example, an intranet or the Internet), or may be a wireless network (for example, a wireless LAN (Local Area Network)).
  • At least one microphone device M1,..., ML as an example of a sound collection unit is installed on a predetermined installation surface (for example, a ceiling surface) of a predetermined sound collection region (for example, a cash register counter of a store), and The voice of the employee (store clerk) is collected, and the voice data of the clerk obtained by the sound collection is transmitted to the recorder device 4.
  • the directivities of the microphone devices M1,..., ML are already determined by the design specifications from the time of manufacture, including the non-directivity having no directivity, and cannot be changed.
  • the microphone devices M1,..., ML do not collect only the voice of the employee (store clerk), but for example, the voice spoken by the customer when the customer and the employee (clerk) are talking is leaked. It is possible to pick up sound.
  • the microphone devices M1,..., ML may be provided for customer service for evaluating the customer service situation of the clerk's customer, or may be provided for monitoring such as crime prevention in the store (see FIG. 37). .
  • the microphone device for monitoring is installed in a place where it is difficult to see in the store, a place far away from the cashier counter in the store, and the like.
  • FIG. 37 is a diagram showing an example of the relationship between the customer microphone devices SM1,..., SML, microphone devices M1,... ML, microphone array devices AM1,.
  • the privacy protection mark is an example of predetermined information indicating protection of the privacy of the customer, for example.
  • the customer's voice may be mixed with the clerk's voice data and collected. This is information indicating in advance that it is not a target for customer service evaluation using the voice data.
  • the microphone device provided for the customer service purpose stores information indicating that the microphone device is provided for the customer service purpose in advance, and the microphone device provided for the monitor purpose is provided for the monitor purpose. Information indicating this is held in advance.
  • the privacy protection mark is given to the clerk's voice data obtained by the sound collection of the microphone device provided for monitoring purposes by the processing of the microphone device. However, the privacy protection mark is not given to the clerk's voice data obtained by collecting sound from the microphone device provided for customer service.
  • the management server 2 as an example of the first storage unit calculates the customer service utterance evaluation value of the store clerk for each customer service event in the customer service evaluation device 3, or the customer service utterance of the store clerk for each customer service event calculated by the customer service evaluation device 3.
  • Various data necessary for browsing the evaluation value is stored (saved) as a management DB (Database) 2a.
  • the management server 2 stores the customer service utterance evaluation value of the store clerk for each customer service event calculated by the customer service evaluation device 3 in the management DB 2a.
  • the management DB 2a includes a customer service event information DB, a viewer DB, a customer service DB, a customer service utterance evaluation DB, and a customer service situation DB. Details of the contents of each DB will be described later.
  • the management server 2 may not be disposed in the store itself where the customer service monitoring system 100 is installed, and may be, for example, online storage (for example, storage used in a cloud service) connected via the network NW.
  • the customer service evaluation device 3 detects various customer service events in a predetermined sound collection area (for example, in a store), and further, based on the detected clerk's voice data during the customer service event, The customer utterance evaluation value corresponding to the utterance assumption keyword is calculated.
  • the customer service evaluation device 3 includes, for example, a data communication device such as a PC (including a laptop and a desktop), a smartphone, a tablet terminal, a mobile phone, or a PDA (Personal Digital Assistant), and includes an operation unit 31, a memory 32, and the like.
  • the operation unit 31 is a user interface (UI: User Interface) for notifying the customer utterance evaluation unit 33 or the output unit 34 of the operation contents of the user (for example, the store manager), for example, a pointing device such as a mouse or a keyboard. It is. Moreover, the operation part 31 may be comprised using the touch panel or touchpad which is arrange
  • UI User Interface
  • the memory 32 is configured by using, for example, a RAM (Random Access Memory), functions as a work memory when each part of the customer service evaluation device 3 operates, and further stores data necessary when each part of the customer service evaluation device 3 operates. .
  • a RAM Random Access Memory
  • the customer service utterance evaluation unit 33 includes, for example, a CPU (Central Processing Unit), an MPU (Micro Processing Unit), or a DSP (Digital Signal Processor), and includes a customer service event detection unit 331 and a service utterance evaluation value calculation unit 332. It is the composition which has.
  • a CPU Central Processing Unit
  • MPU Micro Processing Unit
  • DSP Digital Signal Processor
  • the customer service event detection unit 331 as an example of the detection unit is a customer service event information DB (customer service event data, see later) of the management DB 2a of the management server 2 and a store clerk's operation history for the POS terminal 5 as an example of a predetermined business terminal.
  • the customer service event of the store clerk is detected based on the POS operation history data (terminal operation history data). Details of the service event detection method will be described later.
  • the customer service utterance evaluation value calculation unit 332 as an example of a calculation unit is based on the clerk's voice data stored (stored) in the recorder device 4 in the customer service event detected by the customer service event detection unit 331. A service utterance evaluation value corresponding to a predetermined utterance assumed keyword at the time of operation is calculated. Details of the method for calculating the customer service utterance evaluation value will be described later.
  • the output unit 34 is configured using, for example, a CPU, MPU, or DSP, and the customer service utterance evaluation value calculated by the customer service utterance evaluation value calculation unit 332 is printed on identification information of the store clerk (for example, on a name tag worn by the store clerk, The data is stored in the memory 32 or the management DB 2a of the management server 2 in association with the clerk's voice data specified by the clerk's customer service position (for example, coordinate information) and the customer service time.
  • the output unit 34 has a communication function (wired communication function, wireless communication function) via the network NW with each device of the customer service monitoring system 100, and controls the operation of the display device 35 and the speaker device 36.
  • Various screens related to the customer service monitoring system 100 are displayed on the display device 35 according to a predetermined input operation by the user, or voice packets transmitted from the microphone devices M1,... ML are received and output from the speaker device 36.
  • the display device 35 as an example of the display unit is configured by using, for example, an LCD (Liquid Crystal Display) or an organic EL (Electroluminescence), for example, customer monitoring under the control of the output unit 34 according to a user input operation.
  • Various screens (see later) related to the system 100 are displayed.
  • the speaker device 36 as an example of the audio output unit outputs audio data included in an audio packet transmitted from the microphone devices M1, ..., ML.
  • the display device 35 and the speaker device 36 may be configured integrally with the customer service evaluation device 3 or may be different devices.
  • the monitoring data extracting unit 38 calculates the customer service utterance evaluation value of the store clerk from the monitoring data 4a stored in the recorder device 4 by the customer service utterance evaluation value calculator 332 of the customer service utterance evaluation unit 33.
  • the monitoring data 4ak (k is an integer of 1 or more, for example, see FIG. 38) is extracted every predetermined time interval (for example, about 10 to 20 seconds) necessary for the extraction. Details of the monitoring data extraction processing in the monitoring data extraction unit 38 will be described later with reference to FIGS.
  • the recorder device 4 as an example of the second storage unit stores, as monitoring data 4a, voice data of a clerk uttered at a customer service event and POS operation history data indicating an operation history of the POS terminal 5, for example.
  • the clerk's voice data is voice data collected by at least one microphone device M1, ..., ML and transmitted to the recorder device 4.
  • the POS operation history data is, for example, data acquired at the POS terminal 5 and further acquired by the management server 2 or the customer service evaluation device 3 from the POS terminal 5 and transmitted to the recorder device 4.
  • a POS terminal 5 as an example of a business terminal is installed at a cash register counter in a store and has an input device 51, a display device 52, and a memory 53.
  • the POS terminal 5 stores, for example, store sales information and information regarding the price of each product in the memory 53. 2 to 5, only one POS terminal 5 is illustrated, but a plurality of POS terminals 5 may be connected via the network NW.
  • the input device 51 is a user interface (UI) for receiving an input operation of a user (for example, a store clerk) and notifying the POS terminal 5, and is, for example, a pointing device such as a mouse or a keyboard. Further, the input device 51 may be configured using a touch panel or a touch pad that is arranged corresponding to the screen of the display device 52 and can be operated by a user's finger FG or stylus pen, for example.
  • UI user interface
  • the display device 52 is configured by using an LCD (Liquid Crystal Display) or an organic EL (Electroluminescence) in the same manner as the display device 35, and according to a user input operation, store sales information, information on the price of each product, and products Displays a screen related to the settlement of
  • the memory 53 is configured by using a RAM (Random Access Memory) similarly to the memory 32, functions as a work memory when each part of the POS terminal 5 is operated, and further stores data necessary when each part of the POS terminal 5 is operated.
  • RAM Random Access Memory
  • At least one microphone device M1,..., ML of the customer service monitoring system 100 shown in FIG. 2 is replaced with at least one customer microphone device SM1,.
  • the other configuration is the same as that of the customer service monitoring system 100 shown in FIG.
  • the directivities of the customer microphone devices SM1,..., SML are already determined by design specifications from the time of manufacture, and cannot be changed.
  • the at least one customer microphone device SM1,..., SML as an example of the sound collection unit is configured by using, for example, a pin microphone and is individually attached to each store clerk in the store to collect the sound of the corresponding store clerk. Then, the clerk's voice data obtained by sound collection is transmitted to the recorder device 4. Note that the customer microphone devices SM1,..., SML do not pick up only the voice of the employee (store clerk), for example, the voice spoken by the customer when the customer and the employee (clerk) are talking. It is possible to leak and pick up sound.
  • the customer service microphone devices SM1,..., SML of this embodiment hold information indicating that they are provided for customer service purposes in advance. Therefore, as shown in FIG. 37, the privacy protection mark is not given to the clerk's voice data obtained by the sound collection of the customer microphone devices SM1,..., SML.
  • the monitoring data 4a stored (stored) in the recorder device 4 is the same as the customer service monitoring system 100 shown in FIG. This is voice data of a clerk picked up by a device (for example, a customer microphone device SM1).
  • a device for example, a customer microphone device SM1
  • At least one camera device C1,..., CM and at least one sensor device S1 are the same as those of the customer service monitoring system 100 shown in FIG.
  • At least one camera device C1,..., CM (M is an integer of 1 or more) as an example of an imaging unit is installed fixedly on a ceiling surface of a store, for example, and has a function as a surveillance camera or a security camera.
  • the camera devices C1, By remote operation from the customer service evaluation device 3 connected to the network NW, using the zoom function (for example, zoom-in process, zoom-out process) and the optical axis movement function (pan, tilt), respectively, the camera devices C1,. Captures an image within the angle of view of the CM.
  • the installation positions and directions of the respective camera devices C1,..., CM are registered in advance in, for example, the memory 32 of the customer service evaluation device 3, and control information related to pan / tilt / zoom is transmitted to the customer service evaluation device 3 as needed to display images.
  • the positional relationship between each image position and the directing direction is always associated with each other.
  • each of the camera devices C1,..., CM is an omnidirectional camera, for example, predetermined video data (that is, omnidirectional video data) indicating omnidirectional video in the sound collection area or omnidirectional video data is predetermined.
  • the plane image data generated by performing the panorama conversion by performing the distortion correction process is transmitted to the customer service evaluation device 3 via the network NW. Note that the angle of view and the optical axis of each of the camera devices C1, ..., CM may be fixed.
  • the output unit 34 causes the display device 35 to display video data transmitted from any one of the camera devices C1,..., CM according to, for example, a user input operation.
  • the at least one sensor device S1,..., SN as an example of the customer detection unit detects the appearance or exit of the customer from the store (in other words, the customer enters or exits the store), and information on the detection result is obtained as sensor data.
  • a plurality of sensor devices S1,..., SN may be provided according to the type and number of customer service events that can be detected by the customer service monitoring system 100.
  • a microphone device that picks up a predetermined position (preset position) in the store and a camera device that images the predetermined position are associated in advance. ing. For this reason, the preset ID that is the identification information of the preset position and the camera ID that is the identification information of the camera device that captures the preset position are associated in advance.
  • the monitoring data 4b stored (stored) in the recorder device 4 includes the POS operation history data for each store clerk and the microphone device (for example, the microphone device) in the customer service monitoring system 100 shown in FIG.
  • the microphone device for example, the microphone device
  • the sensor data transmitted from is further added.
  • the customer monitoring system 100C shown in FIG. 2 replaces at least one microphone device M1,..., ML of the customer monitoring system 100 with at least one microphone array device AM1,.
  • a directivity control unit 37 is added, and at least one camera device C1,..., CM and at least one sensor device S1,..., SN are further added. This is the same as the customer service monitoring system 100 shown in FIG.
  • the at least one microphone array device AM1,..., AML as an example of the sound collection unit is installed on a predetermined installation surface (for example, a ceiling surface) of a predetermined sound collection region (for example, a cash register counter of a store). Pick up the clerk's voice.
  • at least one microphone array device AM1,..., AML includes a plurality of microphones as an example of sound collection elements, and a plurality of microphones are used to center the installation positions of the microphone array devices AM1,.
  • a voice for example, a clerk's voice
  • the microphone array devices AM1,..., AML do not pick up only the voice of the employee (store clerk), and for example, the voice spoken by the customer when the customer and the employee (clerk) are talking is leaked. It is possible to pick up sound.
  • the microphone array devices AM1,..., AML may be provided for customer service for evaluating the customer service situation of the clerk's customer, or may be provided for surveillance such as store crime prevention (see FIG. 37). ).
  • the microphone array device for monitoring is installed in a place where it is difficult to see in the store, a place far away from the cashier counter in the store, and the like.
  • the microphone array device provided for customer service holds information indicating that it has been provided for customer service, and the array microphone device provided for monitoring is provided for monitoring purposes. Information indicating that it has been received is held in advance.
  • a privacy protection mark is given to the sound data of the store clerk obtained by the sound collection of the microphone array device provided for monitoring purposes by the processing of the array microphone device. However, the privacy protection mark is not given to the clerk's voice data obtained by collecting sound from the microphone array device provided for customer service.
  • the at least one microphone array device AM1,..., AML transmits a voice packet including the voice collected by each microphone as voice data to the recorder device 4 via the network NW.
  • the operation unit 31 designates an image on the screen displayed on the display device 35 (for example, an image captured by any one of the camera devices C1,..., CM; the same applies hereinafter) by a user operation.
  • the coordinate data indicating the position of the image on the screen is acquired and output to the customer service utterance evaluation unit 33 or the output unit 34.
  • Each camera device C1,..., CM is designated when an arbitrary position is designated by the user's finger FG or stylus pen in a state in which the video data captured by each camera device is displayed on the screen of the display device 35.
  • the coordinate data of the position is received from the customer service evaluation device 3, and the distance and direction (horizontal angle and vertical) from each camera device to the position in the real space corresponding to the designated position (hereinafter simply referred to as “voice position”). Data including corners, and so on) is calculated and transmitted to the customer service evaluation device 3.
  • voice position data including corners, and so on
  • the directivity control unit 37 designates from one of the microphone array devices associated with the camera device that captured the video in accordance with the user's position designation operation from the video displayed on the screen of the display device 35. Coordinates indicating the directivity direction toward the voice position corresponding to the position are calculated. Since the calculation method of the coordinate which shows the directivity direction of the directivity control part 37 is a well-known technique, detailed description is abbreviate
  • the directivity control unit 37 acquires, for example, data on the distance and direction from the installation position of the camera device C1 to the audio position from the camera device C1, and uses these data, for example, the microphone array device AM1 (for example, the camera device C1 and the like).
  • the coordinates indicating the directivity direction from the installation position of the microphone array apparatus AM1 to the voice position are calculated. For example, when the housing of the microphone array device AM1 and the camera device C1 are integrally attached so as to surround the housing of the camera device C1, the direction from the camera device C1 to the sound position (horizontal angle, vertical angle) ) Can be used as coordinates indicating the directivity direction from the microphone array device AM1 to the voice position.
  • the directivity control unit 37 calculates calibration parameter data calculated in advance and the camera device C1.
  • the coordinates indicating the directivity direction from the microphone array device AM1 to the voice position are calculated using the data from the direction to the voice position (horizontal angle, vertical angle).
  • the calibration is an operation for calculating or acquiring a predetermined calibration parameter required for the directivity control unit 37 of the customer service evaluation device 3C to calculate the coordinates indicating the directivity direction.
  • the coordinates indicating the directivity direction are indicated by a horizontal angle in the directivity direction from the microphone array device AM1 toward the sound position and a vertical angle in the directivity direction from the microphone array device AM1 toward the sound position.
  • the sound position is the actual monitoring target or sound collection target corresponding to the designated position designated by the user's finger FG or stylus pen in the video displayed on the screen of the display device 35 by the operation unit 31. Position (see FIG. 1).
  • the directivity control unit 37 emphasizes the voice data by forming the directivity in the direction indicated by the calculated coordinates using the voice data of the store clerk included in the voice packet transmitted from the microphone array device AM1, for example. Then, the audio data after the enhancement processing is generated and passed to the output unit 34. Note that the enhancement processing in the directivity control unit 37 may be performed by any microphone array device corresponding to the camera device that captured the video selected by the user.
  • the monitoring data 4b stored (stored) in the recorder device 4 includes the POS operation history data for each store clerk and the microphone device (for example, the microphone device) in the customer service monitoring system 100 shown in FIG.
  • the sensor data transmitted from is further added.
  • the microphone array devices AM1,..., AML are connected to the network NW. However, some or all of the microphone devices M1,... ML shown in FIG. A part or all of the customer microphone devices SM1,... SML shown in FIG.
  • FIG. 6 is a flowchart for explaining an example of the overall operation procedure of the customer service monitoring system of this embodiment.
  • the service monitoring system 100, 100A, 100B, 100C basically operates in the same manner, but for the sake of simplicity, for example, the system of the service monitoring system 100 shown in FIG. It demonstrates using a structure, and it demonstrates with reference to the structure of each customer service monitoring system 100A, 100B, 100C shown to FIGS. 3-5 as needed.
  • the customer service evaluation is completed (S1, YES)
  • the operation of the customer service monitoring system 100 shown in FIG. 6 is completed.
  • the case where the customer service evaluation is completed includes, for example, a case where the “end” button of the customer service evaluation application installed in the customer service evaluation device 3 is pressed, or a case where the customer service evaluation device 3 is shut down.
  • the present invention is not limited to these cases.
  • the monitoring data extraction unit 38 of the customer service evaluation device 3 calculates the customer service speech evaluation value calculation unit 332 of the customer service speech evaluation unit 33 to calculate the customer service speech evaluation value of the store clerk. Therefore, the monitoring data 4a acquired from the recorder device 4 is cut out at predetermined time intervals (for example, about 10 to 20 seconds). The length of the monitoring data cut out by the monitoring data extraction unit 38 is based on a customer service keyword (for example, “Thank you very much” after or before or after the customer service event (for example, accounting completion operation) that is assumed in advance. ”)" Can be entered until the end of the utterance.
  • a customer service keyword for example, “Thank you very much” after or before or after the customer service event (for example, accounting completion operation) that is assumed in advance. ”
  • the monitoring data extraction unit 38 of the customer service evaluation device 3 shifts the cutting start time, which is a starting point for cutting the monitoring data 4a, for a predetermined time (for example, about 1 second) from the cutting start time of the immediately preceding monitoring data. (S2, see FIG. 38). However, the monitoring data extraction unit 38 of the customer service evaluation device 3 does not perform a shift process when the monitoring data 4a is cut for the first time.
  • FIG. 38 is an explanatory diagram showing an example of the shift cut-out process of the monitoring data 4a. In the process of step S2, if there is a customer service event at the boundary point of the predetermined time interval of the monitoring data (for example, monitoring data 4a2 shown in FIG.
  • the service keyword of the customer service keyword is extracted from the cut monitoring data 4a2. It is provided in order to avoid that detection becomes difficult.
  • the voice data of the customer service keyword assumed in advance is not interrupted in the middle, but from the beginning to the end. Monitoring data that has been recorded up to all will be included.
  • the POS operation history data indicating that the accounting completion operation occurred at the time t1, and the salesclerk from the time t2 to the time t3, “Thank you. Is assumed to be stored.
  • the monitoring data 4a1 first cut by the monitoring data extraction unit 38 in step S2 does not include the clerk's voice data corresponding to the transaction completion operation, the customer service event cannot be detected accurately. Customer service evaluation is impossible.
  • the monitoring data extraction unit 38 of the customer service evaluation device 3 monitors the cutting start time, which is the starting point for cutting out the monitoring data 4a, by shifting the monitoring data 4a1 by a predetermined time ts from the cutting start time of the monitoring data 4a1.
  • the data size of the monitoring data 4ak (k: an integer equal to or larger than 1) for a predetermined time interval acquired by cutting is the same.
  • the monitoring data 4a2 does not store all the voice data of the clerk's “Thank you” utterance, the customer service event cannot be detected, and the customer service evaluation cannot be performed in the same manner.
  • the monitoring data extraction unit 38 of the customer service evaluation device 3 performs monitoring by shifting the cutting start time, which is a starting point for cutting the monitoring data 4a, by a predetermined time ts from the cutting start time of the monitoring data 4a2.
  • Data 4a3 is extracted.
  • the monitoring data 4a3 all the voice data of the clerk's “thank you” utterance is stored, so that the customer service evaluation device 3 can detect the customer service event.
  • the customer service evaluation device 3 includes the customer service keywords in each of the monitoring data 4a1, 4a2, 4a3,... That is always the same length (that is, the predetermined time ts) extracted by the monitoring data extraction unit 38. For example, it is only necessary to try to detect whether or not it is detected by voice recognition. Processing such as detection of speech start time and speech end time and setting of a range to be recognized by speech are not required. Therefore, it is possible to reliably detect a customer service keyword assumed in advance. Note that variations of the monitoring data cut start time will be described later with reference to FIGS.
  • the method for cutting the monitoring data 4b is the same as the method for cutting the monitoring data 4a, and the description thereof is omitted.
  • the monitoring data extraction unit 38 of the customer service evaluation device 3 acquires monitoring data 4a from the recorder device 4 every predetermined time interval (for example, about 10 seconds) from the start time set in step S2 (S3).
  • the acquired monitoring data 4a (specifically, the clerk's POS operation history data and voice data included in the monitoring data 4a) and the start time and end time of the monitoring data 4a are stored in the memory 32 in association with each other (S4). ).
  • the end time is a time obtained by adding a predetermined time interval from the start time.
  • the customer service utterance evaluation unit 33 of the customer service evaluation device 3 performs an event detection availability determination process (S5), and if the event detection availability flag is set to “permitted” (S6, YES), the customer service is performed.
  • the output unit 34 of the evaluation device 3 receives the monitoring data 4a for each predetermined time interval held in the memory 32 in step S4 (in other words, the monitoring data 4a for each predetermined time interval acquired from the recorder device 4). (I.e., customer service event detector 331, customer service utterance evaluation value calculator 332) (S7).
  • a service event detection process is performed in the service event detection unit 331 of the service utterance evaluation unit 33 of the service evaluation device 3 (S8).
  • step S8 the operation of the customer service monitoring system 100 returns to step S1.
  • FIG. 7 is a flowchart for explaining an example of a detailed operation procedure of the event detection availability determination process.
  • the customer service utterance evaluation unit 33 of the customer service evaluation device 3 performs a predetermined area (for example, a header area, a part of the payload area, or other optional areas) of the monitoring data 4a for each predetermined time interval held in the memory 32 in step S4.
  • a privacy protection mark as an example of predetermined information indicating customer privacy protection (S5-1, YES), an event detection enable / disable flag indicating presence / absence of customer service event detection processing is set to “No”. (That is, omit the service event process without performing the service event process) (S5-2).
  • step S5-2 the operation of the customer service monitoring system 100 proceeds to step S6.
  • the customer service utterance evaluation unit 33 protects the privacy of the customer in a predetermined area (for example, a header area, a part of the payload area, or other optional areas) of the monitoring data 4a held in the memory 32 in step S4 at every predetermined time interval. If it is determined that the privacy protection mark as an example of the predetermined information is not included (S5-1, NO), the customer utterance evaluation unit 33 of the customer service evaluation device 3 uses the monitoring data as the voice of the customer who entered the store. It is determined whether it is not included in 4a (S5-3).
  • a predetermined area for example, a header area, a part of the payload area, or other optional areas
  • the customer service utterance evaluation unit 33 includes keywords in the voice data included in the monitoring data 4a that are likely to be uttered by the customer in the store (more specifically, for example, the voice included in the monitoring data 4a). If it is determined that the word spotting processing result of a keyword that is likely to be spoken by the customer in the store with respect to the data is equal to or higher than a predetermined level), it is determined that the voice of the customer is included in the monitoring data 4a (S5). -4, YES).
  • the customer service utterance evaluation unit 33 collects voices other than the clerk registered in advance in the customer mic device individually attached to the clerk (more specifically, for example, collected voice data If the voice clerk recognition result of the clerk registered in advance is determined to be below a predetermined level), it may be determined that the customer's voice is included in the monitoring data 4a (S5-4, YES). .
  • the customer service utterance evaluation unit 33 of the customer service evaluation device 3C detects a face other than the face image of the store clerk registered in advance by performing image processing on the video data included in the monitoring data 4b, and further detects the detected face. If it is determined that the voice uttered from the position of the face or the voice of the detected face position that forms directivity and is emphasized includes human voice, the voice of the customer is monitored It may be determined that it is included in the data 4a (S5-4, YES).
  • the event The detection enable / disable flag is set to “NO” (S5-5).
  • step S5-6 the operation of the customer service monitoring system 100 proceeds to step S6.
  • each predetermined customer service event corresponding to the customer service monitoring systems 100, 100A, 100B, and 100C shown in FIGS. 2 to 5 is included.
  • An example of a customer service event information DB as an example of customer service event data will be described with reference to FIGS.
  • Each customer service event information DB shown in FIGS. 11 to 14 is stored in the management DB 2 a of the management server 2.
  • FIG. 11 is a diagram showing an example of a customer service event information DB corresponding to the customer service monitoring system 100 shown in FIG.
  • FIG. 12 is a diagram illustrating an example of a customer service event information DB corresponding to the customer service monitoring system 100A illustrated in FIG.
  • FIG. 13 is a diagram illustrating an example of a customer service event information DB corresponding to the customer service monitoring system 100B illustrated in FIG.
  • FIG. 14 is a diagram illustrating an example of a customer service event information DB corresponding to the customer service monitoring system 100C illustrated in FIG.
  • description of contents overlapping with those of FIG. 11 will be omitted, and different contents will be described.
  • the customer service event information DB shown in FIG. 11 includes a customer service event ID, a customer service event name, a customer service event determination condition (that is, a condition for determining whether or not a customer service event is detected in the monitoring data 4a), and a customer service event.
  • the type and type of data corresponding to each item of output information are defined.
  • the customer service event determination condition of the customer service event information DB shown in FIG. 11 stipulates that a customer service event detection trigger is a predetermined operation (POS operation) performed on the POS terminal 5.
  • the customer service event output information shown in FIG. 11 stipulates that a preset ID, customer service ID, and customer service event ID are output.
  • the customer service event detection trigger is different for each customer service event. Specifically, a predetermined operation (POS operation) on the POS terminal 5 is performed, voice data Each of them is specified to contain a specific keyword.
  • the customer service event output information in the customer service event information DB shown in FIG. 12 is different for each customer service event, and specifically, a preset ID (identification information of a predetermined position in the store; the same applies hereinafter) and a customer service ID (store employee identification information. The same applies hereinafter) and the customer service event ID, and only the customer service ID (identification information of the customer microphone device worn by the store clerk; the same applies hereinafter) is defined.
  • the customer service event information DB shown in FIG. 13 includes a customer service event ID, an item indicating whether or not the event is a customer service event, a customer service event name, a customer service event determination condition (that is, a customer service event is detected in the monitoring data 4a). And the type of data corresponding to each item of customer service event output information (that is, information output when a customer service event is detected) are defined.
  • the customer service event detection trigger is different for each customer service event.
  • the sensor device S1 for example, an automatic door installed near the store entrance is opened and closed.
  • a store clerk stays at a predetermined position (preset position) corresponding to a predetermined preset ID and there is a position corresponding to the predetermined visitor position preset ID (that is, there is a customer (customer) at the customer service event)
  • POS operation a predetermined operation
  • the customer service event output information of the customer service event information DB shown in FIG. 13 is different for each customer service event, and specifically includes a combination of a microphone ID (see later), a camera ID (see later), a customer ID, and a customer service event ID.
  • a combination of a preset ID, a customer service ID, and a customer service event ID is defined.
  • the customer service event output information in the customer service event information DB shown in FIG. 14 differs for each customer service event. Specifically, the customer service location coordinates, camera ID, customer service ID, customer service event ID combination, preset ID, and customer service A combination of ID and customer service event ID is defined.
  • the customer service position coordinates are used when the directivity control unit 37 forms the sound directivity in the direction from the microphone array device that picks up the sound data of each store clerk toward each store clerk.
  • FIG. 8 is a flowchart for explaining an example of a detailed operation procedure of the customer service event detection process.
  • FIG. 9 is a flowchart illustrating an example of a detailed operation procedure of customer service event information processing.
  • FIGS. 11 to 14 corresponding to the system configurations of the customer monitoring systems 100, 100A, 100B, and 100C shown in FIGS. A specific description will be given with reference to the contents of each record of the customer service event information DB shown.
  • duplicate customer service event ID records are defined in the customer service event information DB shown in FIG. 11 to FIG. 14, duplicate explanation will be omitted and different contents will be explained.
  • the customer service event detection unit 331 receives the monitoring data 4a from the customer service utterance evaluation unit 33 at predetermined time intervals (for example, about 10 seconds) in which the start time and the end time are determined (S8-1).
  • the customer service event information DB (see FIG. 11) stored in the management DB 2a is read (S8-2).
  • the customer service event detection unit 331 acquires a record (service event ID “EID1”, customer service event name “accounting completion greeting”) in the first line of the customer service event information DB that has not been acquired (S8-3), and receives customer service event information. Start processing.
  • the customer service event detection unit 331 acquires the POS operation history data as the verification target data from the monitoring data 4a (S11), and detects a customer event determination condition detection trigger for the transaction completion operation from the POS operation history data (that is, for the POS terminal 5). It is verified whether or not a predetermined operation (operation) has been satisfied (S12).
  • the customer service event information processing shown in FIG. 9 is terminated, and the service of the customer service event detection unit 331 performs step S8. Proceed to -5.
  • the customer service event detection unit 331 receives the customer service event output information ( Specifically, the corresponding preset ID (corresponding one of 1 to PN), customer service ID (identification information of store clerk: corresponding to 1 to EN), customer service event ID) are stored in the customer service status DB (FIG. 10). (Refer to) is stored (held) (S14).
  • the identification information (customer ID: 1 to EN) of the customer (store clerk) who operated the corresponding POS terminal 5 is, for example, a bar printed on the name tag when the operation of the POS terminal 5 is started.
  • the customer ID obtained by reading the code with the barcode reader is used.
  • the directivity of the microphone devices M1,..., ML is predetermined from the time of manufacture, and the directivity cannot be changed. That is, in the customer service monitoring system 100 shown in FIG. 2, since the voice is not collected by the microphone array devices AM1,..., AML (S15, NO), directivity formation processing is impossible, and the customer service event detection unit 331 is Then, voice data of the clerk corresponding to the detected customer service event is acquired from the monitoring data 4a (S16), and the voice data and customer service event ID are input to the customer service utterance evaluation value calculation unit 332 (S18).
  • the clerk's voice data corresponding to the detected customer service event is, for example, voice data collected by the microphone device associated with the POS terminal 5 in which the customer service event “accounting completion operation” is detected.
  • the customer service utterance evaluation value calculation unit 332 executes the customer service utterance evaluation process shown in FIG. 15 (S19), and stores (holds) the service utterance evaluation output value in the customer service status DB (S20). Thereby, the service event information processing shown in FIG. 9 ends.
  • step S8-4 if not all records in the customer service event information DB have been acquired (S8-5, NO), the process of the customer service event detector 331 returns to step S8-3. On the other hand, when all the records of the customer service event information DB have been acquired (S8-5, YES), the process of the customer service event detector 331 ends.
  • the customer service event detecting unit 331 acquires a record (service customer event ID “EID2”, customer service event name “greeting in store”) in the second line of the customer service event information DB that has not been acquired (S8 -3) Start customer service event information processing.
  • the customer service event detection unit 331 acquires all voice data collected by the customer microphone device as verification target data from the monitoring data 4a (S11), performs voice recognition processing on all the voice data, and identifies the result. It is verified whether or not the keyword “I welcome you” is included (S12).
  • the customer service event detection unit 331 outputs a customer service event in the customer service event information DB shown in FIG. Information (specifically, customer service ID (store employee identification information: one corresponding to 1 to EN), customer service event ID) is stored (held) in the customer service DB (see FIG. 10) (S14).
  • the directivities of the customer microphone devices SM1,..., SML are predetermined from the time of manufacture, and the directivity cannot be changed. That is, in the service monitoring system 100A shown in FIG. 3, since voice is not picked up by the microphone array devices AM1,..., AML (S15, NO), directivity formation processing is impossible, and the service event detection unit 331 Then, voice data of the clerk corresponding to the detected customer service event is acquired from the monitoring data 4a (S16), and the voice data and customer service event ID are input to the customer service utterance evaluation value calculation unit 332 (S18).
  • the salesclerk's voice data corresponding to the detected customer service event is voice data that includes the specific keyword “welcome” in the voice recognition processing result in the customer service event detection unit 331.
  • the customer service utterance evaluation value calculation unit 332 executes the customer service utterance evaluation process shown in FIG. 15 (S19), and stores (holds) the service utterance evaluation output value in the customer service status DB (S20). Thereby, the service event information processing shown in FIG. 9 ends.
  • step S8-4 if not all records in the customer service event information DB have been acquired (S8-5, NO), the process of the customer service event detector 331 returns to step S8-3. On the other hand, when all the records of the customer service event information DB have been acquired (S8-5, YES), the process of the customer service event detector 331 ends.
  • step S8-2 the customer service event detection unit 331 acquires a record (service event ID “EID1”, customer service event name “entrance / exit greeting”) in the first line of the service event information DB that has not been acquired ( S8-3), customer service event information processing is started.
  • a microphone ID and a camera ID capable of grasping all store clerk are output as customer service event output information. The same applies hereinafter.
  • the customer service event detection unit 331 acquires video data as verification target data and detection results (automatic door opening / closing history data) included in the sensor data from the monitoring data 4b (S11), and the automatic door opening / closing history data includes the automatic door opening / closing history data. It is checked whether or not there is an opening / closing operation (S12).
  • the customer service event information processing shown in FIG. 9 ends, and the process of the customer service event detection unit 331 returns to step S8-5.
  • the customer service event detection unit 331 receives the service event output information (specifically, the customer service event information DB shown in FIG. 13).
  • the service event output information specifically, the customer service event information DB shown in FIG. 13.
  • Microphone ID identifying information of microphone device: 1 to MN
  • camera ID identifying information of camera device: 1 to CN
  • customer service ID clerk identification information: 1 to EN
  • the camera ID is output as identification information of a camera device that captures the position where each store clerk is closest when the automatic door opening / closing operation is performed by the customer service event detection unit 331 performing image processing on predetermined video data.
  • the directivity of the microphone devices M1... ML is predetermined from the time of manufacture, and the directivity cannot be changed. That is, in the customer service monitoring system 100B shown in FIG. 4, since the voice is not collected by the microphone array devices AM1,..., AML (S15, NO), the directivity forming process is impossible, and the customer service event detecting unit 331 is Then, voice data of each clerk corresponding to the detected customer service event is acquired from the monitoring data 4b (S16), and the voice data and customer service event ID are input to the customer service utterance evaluation value calculation unit 332 (S18).
  • the voice data of each store clerk corresponding to the detected customer service event is determined to be closest to the position where each store clerk was present during the automatic door opening / closing operation by image processing the predetermined video data by the customer service event detection unit 331.
  • Voice data collected by the microphone device From the microphone device, a microphone ID as the identification information is output.
  • the predetermined video data is, for example, video data or a combination of a plurality of video data captured by at least one camera device necessary for understanding the entire area in the store, and the corresponding camera device is fixed. However, it may be changed as appropriate according to the user's input operation, and so on.
  • the customer service utterance evaluation value calculation unit 332 executes the customer service utterance evaluation process shown in FIG. 15 (S19), and stores (holds) the service utterance evaluation output value in the customer service status DB (S20). Thereby, the service event information processing shown in FIG. 9 ends.
  • step S8-4 if not all records in the customer service event information DB have been acquired (S8-5, NO), the process of the customer service event detector 331 returns to step S8-3.
  • the customer service event detection unit 331 acquires the second line record (customer service event ID “EID2”, customer service event name “accounting start greeting”) of the customer service event information DB that has not been acquired (S8-3). Start event information processing.
  • the customer service event detection unit 331 acquires the predetermined video data as the verification target data from the monitoring data 4b (S11), and performs image processing on the video data so that the store clerk performs a “accounting start greeting”.
  • a position for example, the standing position of the cashier counter
  • the customer visitor
  • a predetermined position for example, a predetermined standby position in front of the cashier counter or in the store
  • a predetermined period for example, about 5 seconds
  • the store clerk does not exist at a predetermined position (for example, the operation standing position of the cashier counter) for performing “accounting start greeting”, or the customer (visitor) has a predetermined position (for example, before the cashier counter)
  • a predetermined position for example, the operation standing position of the cashier counter
  • the customer (visitor) has a predetermined position (for example, before the cashier counter)
  • the customer service event information processing shown in FIG. 9 ends. Then, the process of the customer service event detection unit 331 returns to step S8-5.
  • the store clerk is present at a predetermined position (for example, an operation standing position of the cashier counter) for performing “accounting start greeting”, and a customer (visitor) has a predetermined position (for example, cashier counter). If it is determined that the user has stayed for a predetermined period of time (for example, about 5 seconds) at a predetermined standby position in front of or in the store) (S13, YES), the customer service event detection unit 331 is shown in FIG.
  • Customer service event output information in the customer service event information DB (specifically, a preset ID (identifying information of a predetermined position: applicable to 1 to PN) and a customer service ID (identifying information of a clerk: 1 to EN) Stuff) and customer service event ID) are stored (held) in the customer service situation DB (see FIG. 10) (S14).
  • the directivity of the microphone devices M1... ML is predetermined from the time of manufacture, and the directivity cannot be changed. That is, in the customer service monitoring system 100B shown in FIG. 4, since the voice is not collected by the microphone array devices AM1,..., AML (S15, NO), the directivity forming process is impossible, and the customer service event detecting unit 331 is Then, voice data of the store clerk corresponding to the detected customer service event is acquired from the monitoring data 4b (S16), and the voice data and customer service event ID are input to the customer service utterance evaluation value calculation unit 332 (S18). The clerk's voice data corresponding to the detected customer service event is the clerk's voice data collected by the microphone device associated with the predetermined position (preset position).
  • the customer service utterance evaluation value calculation unit 332 executes the customer service utterance evaluation process shown in FIG. 15 (S19), and stores (holds) the service utterance evaluation output value in the customer service status DB (S20). Thereby, the service event information processing shown in FIG. 9 ends.
  • step S8-4 if not all records in the customer service event information DB have been acquired (S8-5, NO), the process of the customer service event detector 331 returns to step S8-3. On the other hand, when all the records of the customer service event information DB have been acquired (S8-5, YES), the process of the customer service event detector 331 ends.
  • the customer service event detection unit 331 acquires a record (service event ID “EID1”, customer service event name “entrance / exit greeting”) in the first line of the service event information DB that has not been acquired ( S8-3), customer service event information processing is started.
  • the customer service event detection unit 331 acquires video data as verification target data and detection results (automatic door opening / closing history data) included in the sensor data from the monitoring data 4b (S11), and the automatic door opening / closing history data includes the automatic door opening / closing history data. It is checked whether or not there is an opening / closing operation (S12).
  • the customer service event information processing shown in FIG. 9 ends, and the process of the customer service event detection unit 331 returns to step S8-5.
  • the customer service event detection unit 331 receives the service event output information (specifically, the customer service event information DB shown in FIG. 13).
  • the service event output information specifically, the customer service event information DB shown in FIG. 13.
  • the customer service position coordinates are obtained by image processing of predetermined video data by the customer service event detection unit 331, and are output as the coordinates of the position where each store clerk exists in the video data displayed on the screen of the display device 35.
  • the camera ID is output as identification information of a camera device that images the position where each store clerk was closest when the automatic door opening / closing operation is performed by the customer service event detection unit 331 performing image processing on predetermined video data.
  • the microphone ID is associated with the camera ID in advance, the microphone ID is selected and output when the camera ID is selected.
  • the unit 331 inputs, to the directivity control unit 37, the data on the customer position coordinates of each store clerk corresponding to the detected customer service event and the voice data of each store clerk corresponding to the customer service event included in the monitoring data 4b.
  • the directivity control unit 37 acquires voice data after directivity is formed in the direction from the microphone array device closest to each store clerk toward each store clerk with respect to the voice data of each store clerk (S17).
  • the customer service event detection unit 331 obtains in the service customer utterance evaluation value calculation unit 332 the data on the customer service staff position coordinates (for example, the coordinates of the position of the sales staff displayed on the screen of the display device 35) and the data in step S17.
  • the received voice data and customer service event ID are input (S18).
  • the customer service utterance evaluation value calculation unit 332 executes the customer service utterance evaluation process shown in FIG. 15 (S19), and stores (holds) the service utterance evaluation output value in the customer service status DB (S20). Thereby, the service event information processing shown in FIG. 9 ends.
  • step S8-4 if not all records in the customer service event information DB have been acquired (S8-5, NO), the process of the customer service event detector 331 returns to step S8-3.
  • FIG. 10 is a diagram illustrating an example of the customer service situation DB.
  • the customer service status DB shown in FIG. 10 includes a customer service status data ID, a customer service utterance evaluation value, an event start time, an event end time, a customer service ID, a customer service event ID, a customer service location (preset), and a customer service. Data corresponding to each item of the person position (outside the preset) is defined.
  • the customer service utterance evaluation value is V11... V1n, and the customer service position is detected because the customer service (store clerk) is not at the default position (preset position). And the coordinates indicating the position of the sales clerk on the video data displayed on the screen of the display device 35 (coordinate position on the screen).
  • the camera device with the camera ID “C1” may be an omnidirectional camera device, a camera device having a fixed angle of view, or a PTZ camera device having a pan / tilt / zoom function.
  • n is an integer of 1 or more. For example, when a plurality of customer service events having the same customer service ID are detected in the monitoring data 4a and 4b of about 10 seconds, n is an integer of 2 or more.
  • the customer service utterance evaluation value is V21... V2m, and m is an integer of 1 or more as in the case of n described above.
  • the customer service position is configured by a preset ID indicating the default position (preset position) since it is detected that the customer (store clerk) is present at the default position (preset position).
  • the customer service monitoring systems 100, 100A, 100B, and 100C shown in FIGS. 2 to 5 have the customer service situation DB shown in FIG. 10 as described above, so that the voice data from the start time to the end time of the corresponding customer service event.
  • video data can be output (reproduced) by the customer service evaluation device 3, and a store manager (for example, a store manager) can carefully observe and review the customer service status of the store clerk at the customer service event with sound and video. it can. Since the voice data is stored (stored) in the recorder device 4, the customer service evaluation device 3 obtains the voice data collected when the customer service event corresponding to the customer service event ID is detected from the recorder device 4. Output (play).
  • FIG. 15 is a flowchart for explaining an example of the operation procedure of the customer service utterance evaluation process.
  • the customer service utterance evaluation value calculation unit 332 acquires the voice data and customer service event ID passed from the customer service event detection unit 331 in step S18 (S21), performs a noise level determination process (S22), and receives the customer service.
  • a keyword utterance determination process is performed (S23).
  • the customer service utterance evaluation value calculation unit 332 determines whether or not the flag of the detection state (see later) is “1” (S24). When the flag of the detection state is “1” (S24, YES), the customer service utterance evaluation value calculation unit 332 performs a scoring process (S25). On the other hand, when the flag of the detection state is not “1” (S24, NO), the service utterance evaluation value calculation unit 332 sets the service utterance evaluation value to a zero point or deducts a predetermined score (S26).
  • the service utterance evaluation value calculation unit 332 outputs the detected keyword ID (described later) and the service utterance evaluation value as scoring data to the service utterance evaluation unit 33 (S27).
  • FIG. 16 is a flowchart for explaining an example of the operation procedure of the noise level determination process.
  • the customer service utterance evaluation value calculation unit 332 determines whether or not the noise level around the sound collection area (for example, a store) acquired by the customer service evaluation device 3 is equal to or less than a predetermined value x [dB] (S22). -1).
  • the noise level is collected by, for example, any one of a microphone device, a customer service microphone device, and a microphone array device, and is transmitted to the customer service evaluation device 3.
  • the customer service utterance evaluation value calculation unit 332 determines that the noise level is equal to or less than the predetermined value x [dB] (S22-1, YES)
  • the utterance determination threshold (see later) is determined to be ⁇ 1 (S22-). 2).
  • the customer service utterance evaluation value calculation unit 332 determines whether the noise level is equal to or lower than the predetermined value y (> x) [dB]. It is determined whether or not (S22-3). If the customer service utterance evaluation value calculation unit 332 determines that the noise level is equal to or lower than the predetermined value y [dB] (S22-3, YES), it determines the utterance determination threshold (see later) to ⁇ 2 (S22-). 4).
  • the customer service utterance evaluation value calculation unit 332 determines ⁇ 3 as the utterance determination threshold (see below) (S22). -5).
  • FIG. 17 is a flowchart for explaining an example of the operation procedure of the service keyword utterance determination process.
  • the customer service utterance evaluation value calculation unit 332 sets the detection state flag to “0” (S23-1).
  • the flag of the detection state is information indicating a state in which an exemplary utterance assumption keyword (see FIG. 22A) that is likely to be uttered by a store clerk in a customer service event or to be uttered is uttered.
  • the service utterance evaluation value calculation unit 332 inputs the voice data acquired in step S21 to the speech recognition engine of the service utterance evaluation value calculation unit 332 (S23-2), and further, all utterance assumption keywords corresponding to the service event ID And a keyword ID for identifying each utterance assumed keyword is acquired from the customer service utterance evaluation DB of the management DB 2a of the management server 2 (S23-3).
  • the customer service utterance evaluation value calculation unit 332 determines whether the speech utterance keyword acquired in step S23-3 is included in the speech recognition result by the speech recognition engine (S23-4). If it is determined that the speech recognition keyword is not included in the speech recognition result by the speech recognition engine (S23-4, NO), the processing of the customer utterance evaluation value calculation unit 332 shown in FIG. 17 ends.
  • the speech recognition It is determined whether or not the evaluation value of the processing result is greater than or equal to the utterance determination threshold (any one of ⁇ 1, ⁇ 2, and ⁇ 3) determined in step S22-2, step S22-4, or step S22-5 (S23-). 5).
  • the evaluation value of the speech recognition processing result is less than the utterance determination threshold value (any of ⁇ 1, ⁇ 2, and ⁇ 3) (S23-5, NO)
  • the customer service utterance evaluation value calculation unit 332 determines that the evaluation value of the speech recognition processing result is equal to or greater than the utterance determination threshold (any one of ⁇ 1, ⁇ 2, and ⁇ 3) (S23-5, YES)
  • the detection state flag is changed to “1” (S23-6)
  • the voice data acquired in step S21 is cut only to the utterance portion of the keyword corresponding to the utterance assumed keyword, and is updated and saved (S23-). 7). Even if there are extra noises before and after the utterance part, by cutting out only the utterance part of the keyword, the noise sound before and after it will be cut, which improves the accuracy of voice recognition.
  • the accuracy of the scoring process in the subsequent step S25 is also ensured.
  • FIG. 18 is a flowchart for explaining an example of the operation procedure of the scoring process.
  • the customer service utterance evaluation value calculation unit 332 performs utterance length determination processing using the voice data updated in step S23-7 (S25-1), and further performs frequency characteristic determination processing (S25). -2). Further, the customer service utterance evaluation value calculation unit 332 obtains scoring data (specifically, the keyword detected from the voice data updated in step S23-7) as a result of the utterance length determination process and the frequency characteristic determination process. A set of a keyword ID for identifying the same utterance assumed keyword and a customer service utterance evaluation value) is stored in the memory 32 (S25-3).
  • FIG. 19A is a flowchart for explaining an example of an operation procedure of speech length determination processing.
  • FIG. 19B is a flowchart illustrating an example of an operation procedure of a frequency characteristic determination process.
  • the service utterance evaluation value calculation unit 332 refers to the service utterance evaluation DB of the management DB 2a of the management server 2, and model voice data specified by the service utterance model ID corresponding to the service event ID acquired in step S21. Is acquired from the management DB 2a of the management server 2 (S31).
  • the model voice data is an example of keyword voice data including voice data of an utterance assumed keyword for each predetermined customer service event.
  • the customer service utterance evaluation value calculation unit 332 determines whether or not the length of the voice data (for example, the clerk's utterance portion) updated in step S23-7 is within a typical predetermined range (S32).
  • FIG. 20 is a diagram showing a specific example of the utterance length determination process using the model voice data.
  • the horizontal axis indicates time. For example, “Welcome” with an utterance length l0 of an exemplary predetermined range of “Welcome” uttered in a customer service event of “greeting in store”, and a predetermined range “I welcome you” (refer to No. 1 shown in FIG. 20) and “I welcome you” (refer to No. 2 shown in FIG. 20) having an utterance length l2.
  • the service utterance evaluation value calculation unit 332 determines that the utterance length of the voice data updated in step S23-7 exceeds a predetermined range (for example, 10%) from the utterance length of the model voice data (utterance length l0) ( For example, in No. 1 and No. 2 shown in FIG. 20 (S32, NO), a predetermined score is subtracted from the customer service utterance evaluation value (S34).
  • a predetermined range for example, 10%
  • the utterance length of the uttered “Ira-saisei” is shorter than the utterance length of “Ira-saisei” in the model voice data by a predetermined range.
  • the service utterance evaluation value calculation unit 332 has a predetermined number of points.
  • “100 ⁇ (0.910 ⁇ 11) / 10” is deducted.
  • 11 corresponds to No. 11 shown in FIG.
  • the utterance length of “I welcome you” uttered in case 1 is shown. More specifically, FIG.
  • the utterance length of the uttered “Islaimase” is longer than the utterance length of the “speaking” of the model voice data by a predetermined range.
  • the customer utterance evaluation value calculation unit 332 As a result, “100 ⁇ (l2 ⁇ 1.1l0) / l0” is deducted.
  • l2 is No. 12 shown in FIG.
  • the utterance length of “I welcome you” uttered in case 2 is shown. More specifically, FIG. 20 shows that the utterance length of the model voice data “Irasementation” is 1 second and the predetermined range is ⁇ 10% of the utterance length of the model voice data “Irasementation”. No.
  • the service utterance evaluation value calculation unit 332 determines that the utterance length of the voice data updated in step S23-7 exceeds a predetermined range (for example, 10%) from the utterance length of the model voice data (utterance length l0). If not (S32, YES) or after step S34, the scoring data (specifically, the keyword ID for identifying the keyword detected in the voice data updated in step S23-7 and the points deducted in step S34)
  • the initial value (for example, a set with the customer utterance evaluation value that is 100 points) that was not deducted after or in step S34 is held in the memory 32 (S33).
  • the service utterance evaluation value calculation unit 332 refers to the service utterance evaluation DB of the management DB 2a of the management server 2, and model voice data specified by the service utterance model ID corresponding to the service event ID acquired in step S21. Is acquired from the management DB 2a of the management server 2 (S41).
  • the service utterance evaluation value calculation unit 332 has a frequency characteristic (for example, frequency) for each phoneme (sound of each word) of the voice data updated in step S23-7 within a predetermined range from a basic frequency for each phoneme. It is determined whether it is within (S42).
  • FIG. 21 is a diagram showing a specific example of frequency characteristic determination processing using the fundamental frequency of each phoneme of the model voice data.
  • the horizontal axis indicates time
  • the dotted circle indicates the fundamental frequencies f1 to f7 of each phoneme of the model speech data
  • the solid circle indicates the frequency f of each phoneme of the speech data updated in step S23-7.
  • '1 to f'7 for example, a predetermined range for each of the fundamental frequencies f1 to f7 of the phoneme of “I welcome you” uttered at the customer service event “greeting at the store” (each of the linear types shown in FIG. 21) (See solid arrows).
  • the service level evaluation value calculation unit 332 For each phoneme of the voice data updated in step S23-7, the service level evaluation value calculation unit 332 sets the frequency characteristics (for example, frequency) of each phoneme to a predetermined range (for example, 60) for each frequency characteristic of each phoneme of the model voice data. [Hz]) If determined to exceed (NO in S42), the predetermined score is subtracted from the customer utterance evaluation value according to the number of phonemes exceeding the predetermined range (S44).
  • the customer utterance evaluation value calculation is performed. If the frequency difference (for example,
  • the utterance of the utterance assumed keyword is uttered in an upward tone (refer to the alternate long and short dash line shown in FIG. 21).
  • the utterance of the utterance assumed keyword is uttered in an upward tone (refer to the alternate long and short dash line shown in FIG. 21).
  • the utterance of the utterance assumed keyword is uttered in an upward tone (refer to the alternate long and short dash line shown in FIG. 21).
  • the utterance of the utterance assumed keyword is uttered in an upward tone (refer to the alternate long and short dash line shown in FIG. 21).
  • the ending or the ending May be used by increasing the value of the fundamental frequency.
  • the service utterance evaluation value calculation unit 332 sets the frequency characteristics (for example, frequency) of each phoneme to a predetermined range (for each frequency characteristic of each phoneme of the model voice data) for each phoneme of the voice data updated in step S23-7. For example, when it is determined that the frequency does not exceed 60 [Hz] (S42, YES) or after step S44, it is detected in the scoring data (specifically, the audio data updated in step S23-7)
  • the memory ID is stored in the memory 32 with the keyword ID for identifying the selected keyword and the customer utterance evaluation value, which is the initial value (for example, 100 points) that has been deducted in step S44 or not deducted in step S44 (S43).
  • FIG. 22A is a diagram showing an example of an utterance assumption keyword table constituting a part of the customer service utterance evaluation DB.
  • FIG. 22B is a diagram illustrating an example of a list of service utterance models that constitute a part of the service utterance evaluation DB.
  • utterance assumption keyword table shown in FIG. 22A, data corresponding to each item of the service event ID, the service event name, the keyword ID, the utterance assumption keyword, and the service utterance model ID are defined.
  • the keyword ID identifies an utterance assumption keyword.
  • the customer service utterance model ID is associated with model voice data as shown in FIG. 22B.
  • one or more utterance assumption keywords may be defined corresponding to one customer utterance model ID (see the record of customer service event ID “EID2” shown in FIG. 22A).
  • FIG. 23 is a flowchart illustrating an example of an operation procedure of a browsing process by a limited viewer or a service utterance evaluation value correction process.
  • the login ID is input by the input operation (for example, touch operation with the finger FG) of the customer who requests the customer service status DB.
  • the password are input, and the login button LGI is pressed (S51).
  • the customer service evaluation device 3 refers to the viewer DB (see FIG. 26A) of the management DB 2a of the management server 2 and inputs information (specifically, entered on the login screen WD1 of FIG. 27) by the browse requester.
  • the access right, authority level, and password are acquired from (password) (S51).
  • FIG. 26A is a diagram illustrating an example of a viewer DB.
  • FIG. 26B is a diagram illustrating an example of a customer service DB.
  • the type and type of data for each item of the viewer ID, password, browser authority, and authority level are defined.
  • the password may be, for example, an actually input password or a hash value (digest) of the input password.
  • the data type and type for each item of the customer ID indicating the store clerk's identification information, the store ID indicating the store identification information, and the name of the store clerk are defined.
  • FIG. 27 is a diagram showing an example of a login screen to the customer service status DB to be browsed in the customer service monitoring system.
  • the customer service evaluation device 3 determines whether or not the browsing requester input on the login screen in step S51 has the access right of the browsing operation defined in the browser DB and the passwords match (S52). If it is determined that there is no access right for the browsing operation or that the passwords do not match (S52, NO), the processing of the customer service evaluation device 3 shown in FIG. 23 ends.
  • the customer service evaluation device 3 determines that the user has access rights for the browsing operation defined in the viewer DB and the passwords match (S52, YES)
  • the customer service status of the management DB 2a of the management server 2 is determined.
  • the DB is accessed, and for example, the total result of the customer service utterance evaluation values of all the store clerks for the customer (customer) per day is displayed on the display device 35 as a customer service status display screen WD2 (see FIG. 28) (S53).
  • FIG. 28 is a diagram illustrating an example of a totaling result of customer service utterance evaluation values of all customer service customers per day as a customer service status display screen WD2.
  • the number of customer service events detected for a customer who visited the store is 255
  • the number of utterances of the utterance assumed keyword of “I welcome you” is 230.
  • the ratio compared to the total number of 255 is 90%
  • the number of utterances of the “Thank you” utterance assumed keyword was executed in the customer service event of closing greetings, and the ratio compared to the total number of 255 was 76 %.
  • the number of customer service events related to cashier reception detected for customers who visited the store is 180, of which the point card confirmation omission (that is, the customer service event for prompting the presentation of the point card is not detected). 8), the ratio compared to the total number 180 is 4%, and the point card confirmation of the store clerk confirms that the presentation rate of the customer's point card is 10%. It is shown that the number of missed warming confirmations (that is, a customer service event for confirming that the lunch box is warmed in the microwave oven was not detected) was 3, and 2% of the total number was 180%. ing.
  • the output unit 34 of the customer service evaluation apparatus 3 aggregates and redisplays the data of each item shown in FIG. 28 by switching to a week or month instead of a day according to a predetermined input operation. May be.
  • the customer service evaluation device 3 closes (closes) all the browsing screens displayed on the display device 35 (S55).
  • the logout button LGO is not selected (S54, NO)
  • the detail display button IDT on the customer service status display screen WD2 shown in FIG. 28 is selected, and the corresponding access right (authority level L1 that allows a correction operation).
  • FIG. 32 is a diagram illustrating a specific example of each record displayed on the detail display screen WD7 of the customer service situation DB.
  • step S58 the process of the customer service evaluation device 3 shown in FIG. 23 returns to step S54.
  • FIG. 24 is a flowchart for explaining an example of a detailed operation procedure of the service utterance evaluation value correction process.
  • FIG. 25 is a flowchart for explaining the continuation of the detailed operation procedure of the modification process of the customer service utterance evaluation value shown in FIG.
  • FIG. 24 in the state where the detailed display screen WD7 shown in FIG. 32 is displayed on the display device 35, for example, the record RC1 of the customer service situation data ID (see FIG. 33) that the user (correction requester) wants to correct is the user's finger FG. (S58-1).
  • FIG. 33 is a diagram illustrating an example of a modification operation for the customer utterance evaluation value of the specific record RC1 displayed on the detail display screen WD7 of the customer service situation DB.
  • the customer service utterance evaluation unit 33 of the customer service evaluation device 3 accesses the customer service status DB of the management DB 2a of the management server 2 via the output unit 34, and the event start time corresponding to the customer service event ID specified in step S58-1.
  • the event end time is extracted, and video data and audio data corresponding to the event start time and event end time are acquired from the recorder device 4 and passed to the output unit 34 (S58-2).
  • FIG. 35 is a diagram illustrating an example of a customer service position correcting operation on the customer service status preview screen.
  • the customer service utterance evaluation unit 33 acquires the customer service location data corresponding to the customer service status data ID from the customer service status DB, and sends it to the directivity control unit 37 together with the voice data. hand over.
  • the directivity control unit 37 uses the voice data and the customer service location data to transmit the voice in the direction from the microphone array device that picks up the customer ID of the customer service ID corresponding to the customer service status data ID to the store employee closest. Directivity is formed and passed to the output unit 34 (S58-2).
  • the customer service utterance evaluation unit 33 is in an active state so that the position change button of the customer service status preview screen WD8 shown in FIG. 35 can be selected, and the customer service status preview is performed by the user's (correction requester) finger FG. It is determined whether or not the stop button on the screen WD8 is selected and there is an access right (authority level L1 at which correction operation is possible) corresponding to the user (correction requester) (S58-3).
  • the microphone array device is not used, so the position change button on the customer service situation preview screen WD8 shown in FIG. Is in a non-selectable state (inactive state).
  • step S58-8 When the position change button of the customer service situation preview screen WD8 is in an inactive state, when the stop button of the customer service situation preview screen WD8 is not selected by the finger FG of the user (correction requester) or access corresponding to the user (correction requester) If there is no right, the process proceeds to step S58-8 (see FIG. 25).
  • the position change button of the customer service status preview screen WD8 is in an active state that can be selected, and the stop button of the customer service status preview screen WD8 is selected by the user's (correction requester) finger FG, and the corresponding access right (correction) If it is determined that there is an authority level L1 that can be operated (S58-3, YES), the pointing direction is designated by the user (correction requester) finger FG on the customer service status preview screen WD8 ( In step S58-4, the directivity control unit 37 changes the directivity direction for forming the sound directivity in the direction specified in step S58-4 (S58-5). The output unit 34 causes the speaker device 36 to output the voice data after the directivity formation with the directivity direction changed. This voice data is confirmed by the user (correction requester) (S58-5).
  • FIG. 36 is a diagram illustrating an example of coordinates of the customer service position after correction of a specific record displayed on the detail display screen of the customer service DB.
  • a cell CL2 of an arbitrary coordinate that is, a coordinate indicating a designated position on the screen of the display device 35 in which the directing direction is changed
  • the customer service position outside the preset of the record of the customer service status data ID “4” Is shown to have changed.
  • the customer service utterance evaluation unit 33 receives the customer selected by the finger FG of the user (correction requester) on the detail display screen WD7 (see FIG. 32) displayed by the display device 35. Whether or not the service utterance evaluation value cell CL1 (see FIG. 34) of the record of the situation data ID (see step S58-1) is double-tapped and there is a corresponding access right (authorization level L1 that can be modified). Is determined (S58-8).
  • FIG. 34 is a diagram illustrating an example of a customer utterance evaluation value after correction of a specific record displayed on the detail display screen WD7 of the customer service situation DB.
  • the cell CL1 of the customer utterance evaluation value of the record of the customer service status data ID (see step S58-1) selected by the user (correction requester) finger FG on the detailed display screen WD7 (see FIG. 32) is double-tapped. And when it is determined that there is a corresponding access right (authority level L1 at which a correction operation can be performed) (YES in S58-8), for example, the operation is performed by using the finger FG of the correction requester to double tap. If the service utterance evaluation value of the cell CL1 has been corrected (changed) (S58-9, YES), the service utterance evaluation unit 33 receives the corrected (changed) service utterance evaluation value (see FIG. 34). The situation DB is overwritten and saved (stored) (S58-10).
  • FIG. 29 is a diagram showing an example of a totaling result of customer utterance evaluation values of all customers with respect to customers in a day time zone.
  • FIG. 30A is a diagram illustrating an example of a totaling result of customer utterance evaluation values of one customer for each time slot of one day.
  • FIG. 30B is a diagram illustrating an example of a total result of customer service utterance evaluation values for each customer service per day.
  • FIG. 31 is a diagram illustrating an example of a totaling result of customer service utterance evaluation values for each store on a day.
  • the output unit 34 of the customer service evaluation device 3 is shown in FIG. 29 from the customer service status display screen WD2 shown in FIG. 28 by a predetermined input operation of the user (that is, a person having at least the authority level L1 shown in FIG. 26A). It may be switched to the customer service status display screen WD3 and displayed on the display device 35.
  • the horizontal axis shows the time of the day
  • the vertical axis shows the number of customers (customers) (see black bars) and the number of salesclerks who could greet the customers (in other words, , The number of customer service event names “sales greetings” properly executed by store clerk, see white bars).
  • the output unit 34 of the customer service evaluation device 3 is shown in FIG. 30A from the customer service status display screen WD3 shown in FIG. 29 by a predetermined input operation of the user (that is, a person having at least the authority level L1 shown in FIG. 26A).
  • the service status display screen WD4 limited to a specific time zone may be switched to display on the display device 35.
  • the horizontal axis indicates the time of the day
  • the vertical axis indicates the number of customers (see black bars) and the number of cases that the store clerk could greet the customer (in other words, , The number of customer service event names “sales greetings” properly executed by store clerk, see white bars).
  • the output unit 34 of the customer service evaluation device 3 is shown in FIG. 30B from the customer service status display screen WD2 shown in FIG. 28 by a predetermined input operation of the user (that is, a person having at least the authority level L1 shown in FIG. 26A). It may be switched to the customer service status display screen WD5 and displayed on the display device 35.
  • the greeting rate, the average score, and the number of cashiers corresponding to each store clerk are shown in comparison on a daily basis.
  • the output unit 34 of the customer service evaluation device 3 aggregates and redisplays the data of each item shown in FIG. 30B by switching to a weekly or monthly unit instead of a daily unit according to a predetermined input operation. May be.
  • the output unit 34 of the customer service evaluation device 3 is shown in FIG. 31 from the customer service status display screen WD2 shown in FIG. 28 by a predetermined input operation of a user (that is, a person having at least the authority level L1 shown in FIG. 26A). It may be switched to the customer service status display screen WD6 and displayed on the display device 35.
  • a predetermined input operation of a user that is, a person having at least the authority level L1 shown in FIG. 26A. It may be switched to the customer service status display screen WD6 and displayed on the display device 35.
  • the number of customers visiting the store, the greeting rate, the average score, and the number of cashiers corresponding to each store are shown in comparison on a daily basis.
  • the output unit 34 of the customer service evaluation device 3 switches the data for each item shown in FIG. 31 and redisplays the data by switching to a week or month instead of a day according to a predetermined input operation. May be.
  • the customer service monitoring systems 100, 100A, 100B, and 100C of the present embodiment have the customer service event information DB (customer service event data) including customer service event determination conditions for each predetermined customer service event and the POS terminal 5 (predetermined business terminal).
  • the customer service event of the clerk is detected, and based on the voice data of the clerk included in the monitoring data 4a or the monitoring data 4b, A service utterance evaluation value corresponding to a predetermined utterance keyword when the POS terminal 5 is operated is calculated.
  • the customer service monitoring systems 100, 100A, 100B, and 100C determine the calculated customer service utterance evaluation values based on the customer service ID (sales employee identification information), the customer service position (sales staff service location), and the customer service time. Are stored in association with the audio data.
  • the customer service monitoring systems 100, 100A, 100B, and 100C can protect the customer's customer's privacy extensively without using human resources like an investigator, as in the prior art.
  • the objectivity of the customer utterance content of the customer can be obtained as a customer utterance evaluation value. It is possible to accurately and objectively evaluate the customer service situation of employees.
  • the customer service monitoring systems 100, 100A, 100B, and 100C have data (that is, privacy protection marks) indicating predetermined information indicating customer privacy protection in employee voice data stored in the recorder device 4 as the second storage unit. ) Is omitted, the detection of employee service events is omitted, so that customer privacy is more clearly protected when customer service events are detected by excluding customer service events that involve customers. can do.
  • the customer service monitoring systems 100, 100A, 100B, and 100C further store in the management DB 2a of the management server 2 the utterance assumption keyword table (keyword data) of the customer utterance evaluation DB including the utterance assumption keywords for each predetermined customer service event. If the speech utterance keyword corresponding to the customer service event is not included in the clerk's voice data, the customer service utterance evaluation value is set to zero or a predetermined score is deducted from the customer service utterance evaluation value. It is possible to accurately evaluate the customer service situation for a store clerk who does not speak.
  • the customer service monitoring systems 100, 100A, 100B, and 100C further store in the management DB 2a of the management server 2 the utterance assumption keyword table (keyword data) of the customer utterance evaluation DB including the utterance assumption keywords for each predetermined customer service event. If the store clerk's voice data includes an utterance assumption keyword corresponding to a customer service event, the clerk's voice data is cut out only to the utterance part of the keyword corresponding to the utterance assumption keyword, and is overwritten and saved, so unnecessary noise The scoring process accuracy can be improved by cutting the sound, and the volume of the clerk's voice data can be reduced. Furthermore, it can contribute to the calculation of an accurate customer utterance evaluation value.
  • the customer service monitoring systems 100, 100A, 100B, and 100C use the customer service utterance model list (keyword speech data) of the customer service utterance evaluation DB including the speech data of the utterance assumed keywords for each predetermined customer service event as a management DB 2a of the management server 2.
  • the predetermined score is subtracted from the customer utterance evaluation value Therefore, it is possible to accurately evaluate the customer service situation for a store clerk who utters an utterance assumption keyword that deviates from an exemplary utterance length at the customer service event.
  • the customer service monitoring systems 100, 100A, 100B, and 100C use the customer service utterance model list (keyword speech data) of the customer service utterance evaluation DB including the speech data of the utterance assumed keywords for each predetermined customer service event as a management DB 2a of the management server 2.
  • a predetermined score is obtained from the customer utterance evaluation value. Since the points are deducted, it is possible to accurately evaluate the customer service situation with respect to the store clerk who utters by deviating from the typical basic frequency at the customer service event.
  • the customer service monitoring system 100A detects the customer service event of the store clerk based on the voice data and customer service event data of each store clerk picked up by the customer microphone devices SM1,..., SML individually attached to the store clerk. Compared with the case where the store clerk is far away from the microphone device other than the customer microphone device (for example, a microphone installed on the ceiling surface), the store clerk's voice can be clearly picked up. A customer service event can be accurately detected.
  • the customer service monitoring systems 100B and 100C further store in the recorder device 4 video data at a predetermined position in a predetermined sound collection area (for example, a store) obtained by imaging of the camera devices C1, ..., CM as monitoring data 4b. Since the customer service event of the store clerk is detected based on this video data, the video data having a predetermined position near the POS terminal 5 where the customer service event of the accounting start greeting of a predetermined sound collection area (for example, store) is performed is processed. By doing so, it is possible to accurately evaluate whether or not the customer service event of the accounting start greeting is properly performed.
  • a predetermined sound collection area for example, a store
  • the customer service monitoring systems 100B and 100C monitor the detection results of the appearance or exit of a customer (for example, the customer entering or leaving the store) in a predetermined sound collection area obtained by the sensor devices S1,. Further, the data is recorded in the recorder device 4 as data 4b, and a customer service event of a store clerk is detected based on the detection result. Therefore, a sensor that gives a motivation for a customer service event for entering and leaving a store in a predetermined sound collection area (for example, a store) Depending on the detection result of the device (for example, an automatic door that opens and closes), it is possible to accurately evaluate whether or not the customer service event for entering and leaving the store is properly performed.
  • a predetermined sound collection area for example, a store
  • the customer service monitoring system 100C has a voice directivity from any one of the microphone array devices AM1,..., AML to a predetermined directivity direction (for example, a fixed position (for example, a cashier counter) at the time of customer service).
  • a predetermined directivity direction for example, a fixed position (for example, a cashier counter) at the time of customer service.
  • the customer service monitoring systems 100, 100A, 100B, and 100C display authority information for the browsing operation on the detailed display screen WD7 (service customer utterance evaluation value display screen) of the customer service situation DB in the customer utterance evaluation value display device 35 for each customer service event.
  • the authority level (authority data) that is included is further stored in the viewer DB of the management DB 2a of the management server 2, and the authority information of the browsing requester (user) on the service utterance evaluation value display screen is included in the authority data. If the condition is satisfied, the customer service utterance evaluation value display screen can be displayed on the display device 35.
  • the customer service monitoring systems 100, 100A, 100B, and 100C further include, as authority data, authority information for correcting the customer service utterance evaluation value for the service status DB detail display screen WD7 (customer service utterance evaluation value display screen). If the authority information of the requester for correction of the evaluation value (user) satisfies the authority information of the correction operation included in the authority data, the customer utterance evaluation value on the customer service utterance evaluation value display screen according to the service utterance evaluation value correction operation Can be updated (corrected).
  • the customer service monitoring systems 100B and 100C further store video data at a predetermined position in a predetermined sound collection area obtained by imaging of the camera devices C1,..., CM as monitoring data 4b in the recorder device 4 for browsing operation and
  • the customer service utterance evaluation value on the customer service utterance evaluation value display screen is updated in accordance with the operation of correcting the customer service position of the customer service event of the store clerk while the video data is output by the operation of the user who has the authority information for both corrective operations ( Correction).
  • the customer service monitoring systems 100, 100A, 100B, and 100C set the service utterance evaluation value for each customer service event on the service utterance evaluation value display screen according to a predetermined input operation by a user who has at least browsing authority information.
  • Each display can be displayed on the display device 35 in a comparative manner, and comparison for each predetermined item can be easily performed.
  • the shift process of step S2 shown in FIG. 6 is not limited to the method of shifting the cut start time that is the starting point for cutting the monitoring data 4a by a predetermined time ts. Variations of the shift process will be described with reference to FIGS.
  • FIG. 39 is an explanatory diagram showing an example of a variation of the cutting process of the monitoring data 4a.
  • FIG. 40 is a flowchart for explaining another example of the detailed operation procedure of customer service event information processing.
  • FIG. 41 is a diagram illustrating an example of an utterance assumption section table for each service event that constitutes a part of the service utterance evaluation DB. In the description of FIG. 40, the same processes as those shown in FIG. 9 are assigned the same step numbers, and the description is simplified or omitted.
  • the monitoring data extraction unit 38 of the customer service evaluation device 3 receives the audio data for a predetermined time interval (that is, the cutout range RG0) from the head of the monitoring data 4a.
  • the monitoring data 4a1 cut out from the POS operation history data is extracted.
  • the monitoring data 4a1 does not include the clerk's voice data corresponding to the transaction completion operation. Therefore, the customer service event cannot be detected, and accurate customer service evaluation is impossible.
  • the monitoring data extraction unit 38 of the customer service evaluation device 3 acquires information related to the utterance assumption section corresponding to the customer service event ID with reference to the utterance assumption section table shown in FIG. 41 after step S14 shown in FIG.
  • the monitoring data 4a2 ′ is reacquired from the monitoring data 4a1 by changing the range of the audio data to be cut (S14-1 shown in FIG. 40).
  • the processing after step S14-1 is the same as the processing after step S15 shown in FIG.
  • step S14-1 the details of step S14-1 will be described with reference to FIG.
  • a service event ID is “accounting completion greeting” with “EID1”
  • the section assumed to be uttered by the store clerk is 10 seconds from the time when the customer service event (that is, the accounting completion greeting) is detected.
  • the monitoring data extraction unit 38 of the customer service evaluation device 3 uses the customer service event ID as a starting point for cutting the audio data of the monitoring data 4a when the customer service event (for example, accounting completion greeting) is detected.
  • the monitoring data extraction unit 38 of the customer service evaluation device 3 updates the monitoring data 4a1 acquired at the first time to the monitoring data 4a2 ′, so that the audio data corresponding to the voice section spoken by the store clerk at the customer service event is obtained. As a result, the customer service event detection accuracy is improved.
  • the cut-out range of the voice data is different for each customer service event, and a customer service event (that is, a customer service start greeting) is detected in the customer service event ID “EID2” “account start greeting”.
  • the total time is 10 seconds over 5 seconds before and after the starting point (see, for example, the cutout range RG3 shown in FIG. 39).
  • the cut-out range of the audio data is the customer service event (that is, the customer greeting) similar to the “accounting completion greeting” with the customer service event ID “EID1”. It is 10 seconds from the time when is detected.
  • the cut-out range of the audio data is a period of 10 seconds before the customer service event (that is, the salutation greeting) is detected (for example, FIG. 39).
  • the monitoring data extraction unit 38 of the customer service evaluation device 3 holds, as an utterance assumption interval table, information related to a speech segment that is assumed to include the speech of the clerk at the customer service event corresponding to the detected customer service event ID. By using this, it is possible to cut out and extract the voice data of the optimum voice section spoken by the store clerk for each customer service event as monitoring data, and to perform accurate customer service evaluation.
  • This disclosure does not use human resources like investigators, and protects customer privacy extensively, and monitors customer utterances during various customer service events for customers in the store. It is useful as a service monitoring system and service monitoring method that improves the convenience of accurately and objectively evaluating the situation.

Abstract

A service monitoring system is provided with: a sound collection unit that collects the sound of an employee's voice in a predetermined sound collection region; a storage unit in which service event data including a determination condition for each predetermined service event, terminal operation history data indicating an employee operation history with respect to a predetermined work terminal, and employee voice data are stored in association with one another; a detection unit that detects an employee's service event on the basis of the service event data and the terminal operation history data; a calculation unit that calculates a service speech evaluation value corresponding to a predetermined speech keyword on the basis of the employee voice data at the time of the detected event; and an output unit that stores the service speech evaluation value in association with employee identification information and the employee voice data identified by the employee's service position and service time.

Description

接客モニタリングシステム及び接客モニタリング方法Service monitoring system and service monitoring method
 本開示は、従業員の音声を用いて接客時の状況をモニタリングする接客モニタリングシステム及び接客モニタリング方法に関する。 The present disclosure relates to a customer service monitoring system and a customer service monitoring method for monitoring a situation during customer service using an employee's voice.
 従来、各種の接客業務では、顧客満足度が売上実績に大きく影響することが知られており、店舗の責任者(例えば店長)には、顧客満足度を定量的に測定することが必要とされるようになっている。 Conventionally, it has been known that customer satisfaction greatly affects sales performance in various customer service operations, and store managers (for example, store managers) are required to measure customer satisfaction quantitatively. It has become so.
 顧客満足度を測定する方法の一例として、調査員による覆面調査という手法がある。調査員による臨店調査では、臨店調査が行われる環境の違いにより得られる調査結果が異なり、調査結果に基づいて行われる改善アドバイスが的確ではない可能性がある。調査が行われる環境の違いとは、例えば臨店調査が行われた時間帯毎の臨店対象店舗内の混雑度や店員(従業員)数の違いや、調査員の調査技術や臨店調査を受ける店員の接客技術の熟練度(例えば勤務年数の長短による差異)の違いである。 の 一 One example of a method for measuring customer satisfaction is a method of covering investigation by an investigator. In the store survey by the investigator, the survey results obtained depending on the environment where the store survey is conducted are different, and the improvement advice based on the survey results may not be accurate. Differences in the environment in which surveys are conducted include, for example, the degree of congestion and the number of store staff (employees) in the stores subject to store surveys for each time period during which the store surveys were conducted, the survey staff's survey technology, and the store staff receiving the store surveys This is the difference in the skill level of customer service (for example, the difference in working years).
 そこで、特許文献1では、臨店調査が実施された環境の違いに基づく臨店調査結果への影響を修正した上で、修正後の調査結果に基づく現実に即したアドバイス情報を提示するサービス評価診断システムが開示されている。特許文献1では、店舗調査(例えば臨店調査)の調査結果は携帯情報端末を操作する調査員により入力される。 Therefore, in Patent Document 1, a service evaluation diagnosis system that presents advice information based on the reality based on the corrected survey result after correcting the influence on the survey result based on the difference in the environment in which the store survey is performed. Is disclosed. In Patent Document 1, a survey result of a store survey (for example, a store survey) is input by an investigator who operates a portable information terminal.
 また、顧客満足度を測定する他の方法の一例として、店員と顧客の会話に含まれる店員の音声及び顧客の音声を基に、店員の感情及び顧客の感情を認識し、この認識結果に基づいて店員満足度データ及び顧客満足度データをそれぞれ算定する接客データ記録装置が知られている(例えば特許文献2参照)。 Another example of a method for measuring customer satisfaction is to recognize store clerk's emotion and customer's emotion based on the store clerk's voice and customer's voice included in the store clerk's customer conversation, and based on this recognition result. There is known a customer service data recording device for calculating store clerk satisfaction data and customer satisfaction data (see, for example, Patent Document 2).
 特許文献2に示す接客データ記録装置は、店員満足度データと顧客満足度データと店員の顧客に対する売上実績とを関連付けた接客データをデータベースに記録する。特許文献2では、特許文献1とは異なり、携帯情報端末を操作する調査員による入力操作は必要無い。 The customer service data recording device shown in Patent Document 2 records customer service data in which a store clerk satisfaction data, a customer satisfaction data, and a sales record of a store clerk for a customer are associated with each other in a database. In Patent Document 2, unlike Patent Document 1, an input operation by an investigator who operates a portable information terminal is not necessary.
 本開示は、調査員のような人的リソースを用いず、顧客のプライバシーを広範に保護した上で、店舗内の顧客に対する各種の接客イベント時における該当接客者の発話内容を監視することで接客状況を的確かつ客観的に評価する接客モニタリングシステム及び接客モニタリング方法を提供することを目的とする。 This disclosure does not use human resources like investigators, and protects customer privacy extensively, and monitors customer utterances during various customer service events for customers in the store. It is an object of the present invention to provide a customer monitoring system and a customer monitoring method for accurately and objectively evaluating the situation.
特許第5336675号公報Japanese Patent No. 5336675 特許第5533219号公報Japanese Patent No. 5533219
 本開示の接客モニタリングシステムは、所定の収音領域における従業員の音声を収音する収音部と、所定の接客イベント毎の判定条件を含む接客イベントデータを記憶する第1記憶部と、所定の業務端末に対する従業員の操作履歴を示す端末操作履歴データと、前記収音部により収音された前記従業員の音声データとを対応付けて記憶する第2記憶部と、前記第1記憶部に記憶される前記接客イベントデータと前記第2記憶部に記憶される前記端末操作履歴データとを基に、前記従業員の前記接客イベントを検出する検出部と、前記検出部により検出された前記接客イベントにおいて、前記第2記憶部に記憶される前記従業員の音声データを基に、前記業務端末の操作時における所定の発話キーワードに対応する接客発話評価値を算定する算定部と、前記算定部により算定された前記接客発話評価値を、前記従業員の識別情報と前記従業員の接客位置及び接客時刻により特定される前記従業員の音声データに対応付けて記憶する出力部と、を備える。 The customer service monitoring system according to the present disclosure includes a sound collection unit that collects employee voices in a predetermined sound collection area, a first storage unit that stores customer event data including determination conditions for each predetermined customer event, and a predetermined storage unit. A second storage unit that stores the terminal operation history data indicating the operation history of the employee for the business terminal and the voice data of the employee collected by the sound collection unit in association with each other, and the first storage unit Based on the customer service event data stored in and the terminal operation history data stored in the second storage unit, a detection unit for detecting the customer service event of the employee, and the detection unit detected by the detection unit In a service event, a service utterance evaluation value corresponding to a predetermined utterance keyword at the time of operation of the business terminal is calculated based on the voice data of the employee stored in the second storage unit. The service utterance evaluation value calculated by the determination unit and the calculation unit is stored in association with the employee's voice data specified by the employee identification information, the customer service position and the customer service time. And an output unit.
 また、本開示の接客モニタリング方法は、所定の収音領域における従業員の音声を収音する収音部を含む接客モニタリングシステムにおける接客モニタリング方法であって、所定の接客イベント毎の判定条件を含む接客イベントデータを第1記憶部に記憶し、所定の業務端末に対する従業員の操作履歴を示す端末操作履歴データと、前記収音部により収音された前記従業員の音声データとを対応付けて第2記憶部に記憶し、前記第1記憶部に記憶される前記接客イベントデータと前記第2記憶部に記憶される前記端末操作履歴データとを基に、前記従業員の前記接客イベントを検出し、検出された前記接客イベントにおいて、前記第2記憶部に記憶される前記従業員の音声データを基に、前記業務端末の操作時における所定の発話キーワードに対応する接客発話評価値を算定し、算定された前記接客発話評価値を、前記従業員の識別情報と前記従業員の接客位置及び接客時刻により特定される前記従業員の音声データに対応付けて記憶する。 Further, the customer service monitoring method of the present disclosure is a customer service monitoring method in a customer service monitoring system including a sound collection unit that collects an employee's voice in a predetermined sound collection region, and includes a determination condition for each predetermined customer service event. The customer service event data is stored in the first storage unit, and the terminal operation history data indicating the operation history of the employee with respect to the predetermined business terminal is associated with the voice data of the employee collected by the sound collecting unit. The customer service event is detected based on the customer service event data stored in the second memory and stored in the first memory and the terminal operation history data stored in the second memory. Then, in the detected customer service event, based on the employee's voice data stored in the second storage unit, a predetermined utterance keyword at the time of operating the business terminal The customer service utterance evaluation value corresponding to the employee is calculated, and the calculated customer service utterance evaluation value is associated with the employee's voice data specified by the employee identification information, the employee service position, and the customer service time. Remember.
図1は、本実施形態の接客モニタリングシステムが設置された店舗内のイメージの一例を示す図である。FIG. 1 is a diagram illustrating an example of an image in a store where a customer service monitoring system according to the present embodiment is installed. 図2は、本実施形態の接客モニタリングシステムの第1のシステム構成例を示すブロック図である。FIG. 2 is a block diagram illustrating a first system configuration example of the customer service monitoring system according to the present embodiment. 図3は、本実施形態の接客モニタリングシステムの第2のシステム構成例を示すブロック図である。FIG. 3 is a block diagram illustrating a second system configuration example of the customer service monitoring system according to the present embodiment. 図4は、本実施形態の接客モニタリングシステムの第3のシステム構成例を示すブロック図である。FIG. 4 is a block diagram illustrating a third system configuration example of the customer service monitoring system according to the present embodiment. 図5は、本実施形態の接客モニタリングシステムの第4のシステム構成例を示すブロック図である。FIG. 5 is a block diagram illustrating a fourth system configuration example of the customer service monitoring system according to the present embodiment. 図6は、本実施形態の接客モニタリングシステムの全体の動作手順の一例を説明するフローチャートである。FIG. 6 is a flowchart for explaining an example of the overall operation procedure of the customer service monitoring system of this embodiment. 図7は、イベント検出可否判定処理の詳細な動作手順の一例を説明するフローチャートである。FIG. 7 is a flowchart for explaining an example of a detailed operation procedure of the event detection availability determination process. 図8は、接客イベント検出処理の詳細な動作手順の一例を説明するフローチャートである。FIG. 8 is a flowchart for explaining an example of a detailed operation procedure of the customer service event detection process. 図9は、接客イベント情報処理の詳細な動作手順の一例を説明するフローチャートである。FIG. 9 is a flowchart illustrating an example of a detailed operation procedure of customer service event information processing. 図10は、接客状況DBの一例を示す図である。FIG. 10 is a diagram illustrating an example of the customer service situation DB. 図11は、図2に示す接客モニタリングシステムに対応した接客イベント情報DBの一例を示す図である。FIG. 11 is a diagram illustrating an example of a customer service event information DB corresponding to the customer service monitoring system illustrated in FIG. 2. 図12は、図3に示す接客モニタリングシステムに対応した接客イベント情報DBの一例を示す図である。FIG. 12 is a diagram illustrating an example of a customer service event information DB corresponding to the customer service monitoring system illustrated in FIG. 3. 図13は、図4に示す接客モニタリングシステムに対応した接客イベント情報DBの一例を示す図である。FIG. 13 is a diagram illustrating an example of a customer service event information DB corresponding to the customer service monitoring system illustrated in FIG. 4. 図14は、図5に示す接客モニタリングシステムに対応した接客イベント情報DBの一例を示す図である。FIG. 14 is a diagram showing an example of a customer service event information DB corresponding to the customer service monitoring system shown in FIG. 図15は、接客発話評価処理の動作手順の一例を説明するフローチャートである。FIG. 15 is a flowchart for explaining an example of the operation procedure of the customer service utterance evaluation process. 図16は、騒音レベル判定処理の動作手順の一例を説明するフローチャートである。FIG. 16 is a flowchart for explaining an example of the operation procedure of the noise level determination process. 図17は、接客キーワードの発話判定処理の動作手順の一例を説明するフローチャートである。FIG. 17 is a flowchart for explaining an example of the operation procedure of the service keyword utterance determination process. 図18は、スコアリング処理の動作手順の一例を説明するフローチャートである。FIG. 18 is a flowchart for explaining an example of the operation procedure of the scoring process. 図19Aは、発話長さの判定処理の動作手順の一例を説明するフローチャートである。FIG. 19A is a flowchart for explaining an example of an operation procedure of speech length determination processing. 図19Bは、周波数特性の判定処理の動作手順の一例を説明するフローチャートである。FIG. 19B is a flowchart illustrating an example of an operation procedure of a frequency characteristic determination process. 図20は、モデル音声データを用いた発話長さの判定処理の具体例を示す図である。FIG. 20 is a diagram illustrating a specific example of an utterance length determination process using model voice data. 図21は、モデル音声データの各音素の基本周波数を用いた周波数特性の判定処理の具体例を示す図である。FIG. 21 is a diagram showing a specific example of the frequency characteristic determination process using the fundamental frequency of each phoneme of the model voice data. 図22Aは、接客発話評価DBの一部を構成する発話想定キーワード表の一例を示す図である。FIG. 22A is a diagram illustrating an example of an utterance assumption keyword table constituting a part of the customer service utterance evaluation DB. 図22Bは、接客発話評価DBの一部を構成する接客発話モデル一覧表の一例を示す図である。FIG. 22B is a diagram illustrating an example of a service utterance model list that forms part of the service utterance evaluation DB. 図23は、限定された閲覧者による閲覧処理又は接客発話評価値の修正処理の動作手順の一例を示すフローチャートである。FIG. 23 is a flowchart illustrating an example of an operation procedure of a browsing process by a limited viewer or a service utterance evaluation value correction process. 図24は、接客発話評価値の修正処理の詳細な動作手順の一例を説明するフローチャートである。FIG. 24 is a flowchart for explaining an example of a detailed operation procedure of the service utterance evaluation value correction process. 図25は、図24に示す接客発話評価値の修正処理の詳細な動作手順の続きを説明するフローチャートである。FIG. 25 is a flowchart for explaining the continuation of the detailed operation procedure of the modification process of the customer service utterance evaluation value shown in FIG. 図26Aは、閲覧者DBの一例を示す図である。FIG. 26A is a diagram illustrating an example of a viewer DB. 図26Bは、接客者DBの一例を示す図である。FIG. 26B is a diagram illustrating an example of a customer service DB. 図27は、接客モニタリングシステムにおける閲覧対象となる接客状況DBへのログイン画面の一例を示す図である。FIG. 27 is a diagram illustrating an example of a login screen to the customer service status DB to be browsed in the customer service monitoring system. 図28は、1日単位の来店客に対する全接客者の接客発話評価値の集計結果の一例を接客状況表示画面として示す図である。FIG. 28 is a diagram illustrating an example of a totaling result of customer service utterance evaluation values of all customer service customers per day as a customer service status display screen. 図29は、1日の時間帯における来店客に対する全接客者の接客発話評価値の集計結果の一例を示す図である。FIG. 29 is a diagram illustrating an example of a totaling result of customer service utterance evaluation values of all customers with respect to customers in a day time zone. 図30Aは、1日の時間帯毎の1人の接客者の接客発話評価値の集計結果の一例を示す図である。FIG. 30A is a diagram illustrating an example of a totaling result of customer utterance evaluation values of one customer for each time slot of one day. 図30Bは、1日の接客者毎の接客発話評価値の集計結果の一例を示す図である。FIG. 30B is a diagram illustrating an example of a total result of customer service utterance evaluation values for each customer service per day. 図31は、1日の店舗毎の接客発話評価値の集計結果の一例を示す図である。FIG. 31 is a diagram illustrating an example of a totaling result of customer service utterance evaluation values for each store on a day. 図32は、接客状況DBの詳細表示画面に表示された各レコードの具体例を示す図である。FIG. 32 is a diagram illustrating a specific example of each record displayed on the detail display screen of the customer service situation DB. 図33は、接客状況DBの詳細表示画面に表示された特定のレコードの接客発話評価値の修正操作の一例を示す図である。FIG. 33 is a diagram illustrating an example of a modification operation for the customer utterance evaluation value of a specific record displayed on the detail display screen of the customer service status DB. 図34は、接客状況DBの詳細表示画面に表示された特定のレコードの修正後の接客発話評価値の一例を示す図である。FIG. 34 is a diagram illustrating an example of the customer utterance evaluation value after correction of a specific record displayed on the detail display screen of the customer service situation DB. 図35は、接客状況プレビュー画面における接客位置の修正操作の一例を示す図である。FIG. 35 is a diagram illustrating an example of a customer service position correcting operation on the customer service status preview screen. 図36は、接客状況DBの詳細表示画面に表示された特定のレコードの修正後の接客位置の座標の一例を示す図である。FIG. 36 is a diagram illustrating an example of coordinates of the customer service position after correction of a specific record displayed on the detail display screen of the customer service DB. 図37は、接客者マイク装置、マイク装置、マイクアレイ装置とプライバシー保護マークとの関係の一例を示す図である。FIG. 37 is a diagram illustrating an example of the relationship between the customer microphone device, the microphone device, the microphone array device, and the privacy protection mark. 図38は、モニタリングデータのシフト切り出し処理の一例を示す説明図である。FIG. 38 is an explanatory diagram illustrating an example of shift cut-out processing of monitoring data. 図39は、モニタリングデータの切り出し処理のバリエーションの一例を示す説明図である。FIG. 39 is an explanatory diagram illustrating an example of a variation of the monitoring data cut-out process. 図40は、接客イベント情報処理の詳細な動作手順の他の一例を説明するフローチャートである。FIG. 40 is a flowchart for explaining another example of the detailed operation procedure of customer service event information processing. 図41は、接客発話評価DBの一部を構成する接客イベントごとの発話想定区間表の一例を示す図である。FIG. 41 is a diagram illustrating an example of an utterance assumption section table for each service event that constitutes a part of the service utterance evaluation DB.
 以下、本開示に係る接客モニタリングシステム及び接客モニタリング方法を具体的に開示した実施形態(以下、「本実施形態」という)について、図面を参照して説明する。本実施形態の接客モニタリングシステムは、接客業務が行われる店舗(例えば小売店、卸売店、百貨店、コンビニエンスストア、スーパーマーケット、飲食店、銀行)に設置され、店舗の店員(従業員)の顧客に対する接客状況を監視し、店舗における各種の接客イベント(例えば入退店挨拶、会計開始挨拶等。詳細は後述参照)を行う店員の接客態度(おもてなし)を客観的に評価する。以下の説明において、接客モニタリングシステムにおいて店員の顧客に対する接客態度(接客状況)を客観的に評価した結果の定量的な指標(値)を「接客発話評価値」という。なお、本実施形態では、店員の顧客に対する接客状況の評価について説明するが、本開示に係る接客モニタリングシステム及び接客モニタリング方法は、店員以外の従業員(例えば銀行員、職員など)の顧客に対する接客状況の評価にも適用可能である。 Hereinafter, an embodiment (hereinafter referred to as “the present embodiment”) that specifically discloses the customer service monitoring system and the customer service monitoring method according to the present disclosure will be described with reference to the drawings. The customer service monitoring system according to the present embodiment is installed in a store (for example, a retail store, a wholesale store, a department store, a convenience store, a supermarket, a restaurant, or a bank) where customer service is performed, and serves a customer of a store clerk (employee) to a customer. The situation is monitored, and the customer service attitude (hospitality) of the store clerk who performs various customer service events (for example, store entrance / exit greetings, accounting start greetings, etc., see below for details) is objectively evaluated. In the following description, a quantitative index (value) as a result of objectively evaluating the customer service attitude (customer service situation) of the store clerk in the customer service monitoring system is referred to as a “customer utterance evaluation value”. In addition, although this embodiment demonstrates the evaluation of the customer service situation with respect to the customer of the clerk, the customer service monitoring system and the customer service monitoring method according to the present disclosure are a customer service to an employee other than the clerk (for example, a bank clerk, an employee, etc.) It is also applicable to situational assessment.
 なお、本開示は、接客モニタリングシステムを構成する各装置(例えば後述する接客評価装置)、指向性制御システムを構成する各装置(例えば接客評価装置)が行う各動作を含む方法、又は当該方法をコンピュータである接客評価装置に実行させるためのプログラムとして表現することも可能である。 In addition, this indication is a method including each operation | movement which each apparatus (for example, customer service evaluation apparatus mentioned later) which comprises a customer service monitoring system, each apparatus (for example, customer service evaluation apparatus) which comprises a directivity control system, or the said method It can also be expressed as a program to be executed by the customer service evaluation device which is a computer.
 (接客モニタリングシステムの概略)
 図1は、本実施形態の接客モニタリングシステム100が設置された店舗内のイメージの一例を示す図である。図1では、例えば店舗において、2個のPOS端末5が設置されたレジカウンターにそれぞれ2名の従業員がPOS端末5を操作しながら複数の顧客を接客している。店舗には、収音領域であるレジカウンターの周囲の音声を収音するためのマイク装置M1が少なくとも1個設置され、レジカウンターを画角に含むように撮像するカメラ装置C1が少なくとも1個設置され、更に、店舗への顧客の入退店を検出するためのセンサ装置S1が店舗の入り口付近に設置されている。また、接客モニタリングシステム100では、店舗のバックヤード(例えば監視室)において、従業員(店員)の接客状況の監視(モニタリング)により接客発話評価値を算定するための接客評価装置3(例えばPC:Personal Computer)が設置されており、例えば店舗の責任者(例えば店長)により操作される。
(Outline of customer service monitoring system)
FIG. 1 is a diagram illustrating an example of an image in a store where a customer service monitoring system 100 according to the present embodiment is installed. In FIG. 1, for example, in a store, two employees each serve a plurality of customers while operating the POS terminal 5 at a cashier counter in which two POS terminals 5 are installed. At the store, at least one microphone device M1 for picking up sound around the cashier counter, which is a sound pickup area, is installed, and at least one camera device C1 for taking an image so as to include the cashier counter is included. In addition, a sensor device S1 for detecting the entrance and exit of the customer to the store is installed near the entrance of the store. In the customer service monitoring system 100, the customer service evaluation device 3 (for example, PC :) for calculating the customer service utterance evaluation value by monitoring (monitoring) the customer service situation in the backyard (for example, monitoring room) of the store. Personal Computer) is installed and operated by, for example, a store manager (for example, store manager).
 接客モニタリングシステム100は、例えばPOS端末5を操作する店員の操作履歴を示すデータ(以下、「POS操作履歴データ」という)と、後述する種々の接客イベント毎に接客イベントの有無を検出するための接客イベント判定条件を含む接客イベントデータとを基に、店員による接客イベントを検出する。なお、POS操作履歴データには、例えば接客者(店員)の接客者ID(店員の識別情報)、接客者IDをPOS端末5に入力する接客者ID入力操作履歴、来店客(顧客)の年齢入力履歴、会計完了操作履歴等が含まれる。 The customer service monitoring system 100 detects, for example, data indicating the operation history of a store clerk who operates the POS terminal 5 (hereinafter referred to as “POS operation history data”) and presence / absence of a customer service event for each of various customer service events to be described later. A customer service event by a store clerk is detected based on customer service event data including a customer service event determination condition. The POS operation history data includes, for example, a customer ID of the customer (store clerk) (a store clerk identification information), a customer ID input operation history for inputting the customer ID into the POS terminal 5, and the age of the customer (customer). Includes input history, transaction completion operation history, and the like.
 接客モニタリングシステム100は、接客イベントを検出した場合には、例えばマイク装置M1、カメラ装置C1、センサ装置S1又はこれらの組み合わせにより得られたデータを基に、POS端末5を操作時における所定の発話想定キーワード(発話キーワード)に対応する接客発話評価値(言い換えると、発話想定キーワードを発話した店員に対する接客発話評価値)を算定する。また、接客モニタリングシステム100は、算定された接客発話評価値を、店員の識別情報と店員の接客位置及び接客時刻により特定される店員の音声データに対応付けて記憶する。ここでいう「店員(従業員)の音声データ」とは、例えば店員の位置において店員が発話した声がマイク装置M1、接客者マイク装置SM1又はマイクアレイ装置AM1により収音された音声が記録されているデータである。 When the customer service monitoring system 100 detects a customer service event, for example, based on data obtained by the microphone device M1, the camera device C1, the sensor device S1, or a combination thereof, a predetermined utterance when operating the POS terminal 5 A service utterance evaluation value corresponding to the assumed keyword (utterance keyword) (in other words, a service utterance evaluation value for a clerk who utters the utterance assumed keyword) is calculated. In addition, the customer service monitoring system 100 stores the calculated customer service utterance evaluation value in association with the clerk's voice data specified by the clerk's identification information, the clerk's customer service position, and the customer service time. The “salesperson (employee) voice data” here refers to, for example, voice recorded by the microphone device M1, the customer microphone device SM1, or the microphone array device AM1 of the voice uttered by the store clerk at the position of the store clerk. Data.
 (接客モニタリングシステムの構成例)
 次に、本実施形態の接客モニタリングシステムのシステム構成例について、図2から図5を参照して説明する。図2は、本実施形態の接客モニタリングシステム100の第1のシステム構成例を示すブロック図である。図3は、本実施形態の接客モニタリングシステム100Aの第2のシステム構成例を示すブロック図である。図4は、本実施形態の接客モニタリングシステム100Bの第3のシステム構成例を示すブロック図である。図5は、本実施形態の接客モニタリングシステム100Cの第4のシステム構成例を示すブロック図である。図2から図5の説明では、図2に示す接客モニタリングシステム100を詳細に説明し、図3,図4,図5にそれぞれ示す接客モニタリングシステム100A,100B,100Cの構成については図2に示す接客モニタリングシステム100との差異について説明し、同一の構成については同一の符号を付して説明を簡略化又は省略する。
(Configuration example of customer service monitoring system)
Next, a system configuration example of the customer service monitoring system according to the present embodiment will be described with reference to FIGS. FIG. 2 is a block diagram illustrating a first system configuration example of the customer service monitoring system 100 according to the present embodiment. FIG. 3 is a block diagram illustrating a second system configuration example of the customer service monitoring system 100A according to the present embodiment. FIG. 4 is a block diagram illustrating a third system configuration example of the customer service monitoring system 100B according to the present embodiment. FIG. 5 is a block diagram illustrating a fourth system configuration example of the customer service monitoring system 100C according to the present embodiment. 2 to FIG. 5, the customer service monitoring system 100 shown in FIG. 2 will be described in detail, and the structures of the customer service monitoring systems 100A, 100B, and 100C shown in FIG. 3, FIG. 4, and FIG. Differences from the customer service monitoring system 100 will be described, and the same components will be denoted by the same reference numerals, and description thereof will be simplified or omitted.
 図2に示す接客モニタリングシステム100は、少なくとも1個のマイク装置M1,…,ML(Lは1以上の整数。以下同様)と、管理サーバ2と、接客評価装置3と、レコーダ装置4と、POS端末5とを含む構成である。図2に示す接客モニタリングシステム100では、少なくとも1個のマイク装置M1,…,MLと、管理サーバ2と、接客評価装置3と、レコーダ装置4と、POS端末5とが、ネットワークNWを介して相互に接続されている。ネットワークNWは、有線ネットワーク(例えばイントラネット、インターネット)でも良いし、無線ネットワーク(例えば無線LAN(Local Area Network))でもよい。 The customer service monitoring system 100 shown in FIG. 2 includes at least one microphone device M1,..., ML (L is an integer of 1 or more; the same applies hereinafter), a management server 2, a customer service evaluation device 3, a recorder device 4, The POS terminal 5 is included. In the customer service monitoring system 100 shown in FIG. 2, at least one microphone device M1,..., ML, management server 2, customer service evaluation device 3, recorder device 4, and POS terminal 5 are connected via a network NW. Are connected to each other. The network NW may be a wired network (for example, an intranet or the Internet), or may be a wireless network (for example, a wireless LAN (Local Area Network)).
 収音部の一例としての少なくとも1個のマイク装置M1,…,MLは、所定の収音領域(例えば店舗のレジカウンター)の所定の設置面(例えば天井面)に設置され、収音領域の従業員(店員)の音声を収音し、収音により得られた店員の音声データをレコーダ装置4に送信する。マイク装置M1,…,MLの指向性は、指向性が無い無指向性であることも含めて、それぞれ製造時から設計仕様により既に決められているものであり、変更することはできない。また、マイク装置M1,…,MLは、従業員(店員)の音声のみ収音するものではなく、例えば顧客と従業員(店員)とが会話している際に顧客が発話した声も漏れて収音することはあり得る。 At least one microphone device M1,..., ML as an example of a sound collection unit is installed on a predetermined installation surface (for example, a ceiling surface) of a predetermined sound collection region (for example, a cash register counter of a store), and The voice of the employee (store clerk) is collected, and the voice data of the clerk obtained by the sound collection is transmitted to the recorder device 4. The directivities of the microphone devices M1,..., ML are already determined by the design specifications from the time of manufacture, including the non-directivity having no directivity, and cannot be changed. In addition, the microphone devices M1,..., ML do not collect only the voice of the employee (store clerk), but for example, the voice spoken by the customer when the customer and the employee (clerk) are talking is leaked. It is possible to pick up sound.
 また、マイク装置M1,…,MLは、店員の顧客に対する接客状況を評価するための接客用途に設けられてもよいし、店舗の防犯等の監視用途に設けられてもよい(図37参照)。監視用途のマイク装置は、店舗内の人目の付きにくい場所、店舗内のレジカウンターから遠く離れた場所等に設置される。 Further, the microphone devices M1,..., ML may be provided for customer service for evaluating the customer service situation of the clerk's customer, or may be provided for monitoring such as crime prevention in the store (see FIG. 37). . The microphone device for monitoring is installed in a place where it is difficult to see in the store, a place far away from the cashier counter in the store, and the like.
 図37は、接客者マイク装置SM1,…,SML、マイク装置M1,…,ML、マイクアレイ装置AM1,…,AMLとプライバシー保護マークとの関係の一例を示す図である。プライバシー保護マークは、例えば顧客のプライバシー保護を示す所定情報の一例であり、顧客のプライバシーを保護するために、顧客の音声が店員の音声データに混じって収音される可能性がある場合にはその音声データを用いた店員の接客評価の対象としないことを予め示す情報である。 FIG. 37 is a diagram showing an example of the relationship between the customer microphone devices SM1,..., SML, microphone devices M1,... ML, microphone array devices AM1,. The privacy protection mark is an example of predetermined information indicating protection of the privacy of the customer, for example. In order to protect the privacy of the customer, the customer's voice may be mixed with the clerk's voice data and collected. This is information indicating in advance that it is not a target for customer service evaluation using the voice data.
 つまり、図37に示すように、接客用途に設けられたマイク装置は、接客用途に設けられたことを示す情報を予め保持し、監視用途に設けられたマイク装置は、監視用途に設けられたことを示す情報を予め保持している。監視用途に設けられたマイク装置の収音により得られた店員の音声データには、マイク装置の処理により、プライバシー保護マークが付与される。しかし、接客用途に設けられたマイク装置の収音により得られた店員の音声データには、プライバシー保護マークは付与されない。 That is, as shown in FIG. 37, the microphone device provided for the customer service purpose stores information indicating that the microphone device is provided for the customer service purpose in advance, and the microphone device provided for the monitor purpose is provided for the monitor purpose. Information indicating this is held in advance. The privacy protection mark is given to the clerk's voice data obtained by the sound collection of the microphone device provided for monitoring purposes by the processing of the microphone device. However, the privacy protection mark is not given to the clerk's voice data obtained by collecting sound from the microphone device provided for customer service.
 第1記憶部の一例としての管理サーバ2は、接客評価装置3において接客イベント毎の店員の接客発話評価値を算定する際、又は接客評価装置3において算定された接客イベント毎の店員の接客発話評価値を閲覧(更に、必要に応じて修正)する際に必要となる各種データを管理DB(Database)2aとして格納(保存)している。また、管理サーバ2は、接客評価装置3において算定された接客イベント毎の店員の接客発話評価値を管理DB2aにおいて格納している。 The management server 2 as an example of the first storage unit calculates the customer service utterance evaluation value of the store clerk for each customer service event in the customer service evaluation device 3, or the customer service utterance of the store clerk for each customer service event calculated by the customer service evaluation device 3. Various data necessary for browsing the evaluation value (further modified as necessary) is stored (saved) as a management DB (Database) 2a. Further, the management server 2 stores the customer service utterance evaluation value of the store clerk for each customer service event calculated by the customer service evaluation device 3 in the management DB 2a.
 管理DB2aは、接客イベント情報DBと、閲覧者DBと、接客者DBと、接客発話評価DBと、接客状況DBとを有する構成である。各々のDBの内容の詳細については後述する。なお、管理サーバ2は、接客モニタリングシステム100が設置される店舗自体に配置されなくてもよく、例えばネットワークNWを介して接続されるオンラインストレージ(例えばクラウドサービスにおいて使用されるストレージ)でもよい。 The management DB 2a includes a customer service event information DB, a viewer DB, a customer service DB, a customer service utterance evaluation DB, and a customer service situation DB. Details of the contents of each DB will be described later. Note that the management server 2 may not be disposed in the store itself where the customer service monitoring system 100 is installed, and may be, for example, online storage (for example, storage used in a cloud service) connected via the network NW.
 接客評価装置3は、所定の収音領域(例えば店舗内)における各種の接客イベントを検出し、更に、検出された接客イベント中の店員の音声データを基に、接客イベント中の店員の所定の発話想定キーワードに対応する接客発話評価値を算定する。接客評価装置3は、例えばPC(ラップトップやデスクトップも含む)、スマートフォン、タブレット端末、携帯電話機又はPDA(Personal Digital Assistant)等のデータ通信機器を用いて構成され、操作部31と、メモリ32と、接客発話評価部33と、出力部34と、表示装置35と、スピーカ装置36と、モニタリングデータ抽出部38とを有する構成である。 The customer service evaluation device 3 detects various customer service events in a predetermined sound collection area (for example, in a store), and further, based on the detected clerk's voice data during the customer service event, The customer utterance evaluation value corresponding to the utterance assumption keyword is calculated. The customer service evaluation device 3 includes, for example, a data communication device such as a PC (including a laptop and a desktop), a smartphone, a tablet terminal, a mobile phone, or a PDA (Personal Digital Assistant), and includes an operation unit 31, a memory 32, and the like. The customer utterance evaluation unit 33, the output unit 34, the display device 35, the speaker device 36, and the monitoring data extraction unit 38.
 操作部31は、ユーザ(例えば店舗の店長)の操作の内容を接客発話評価部33又は出力部34に通知するためのユーザインターフェース(UI:User Interface)であり、例えばマウス、キーボード等のポインティングデバイスである。また、操作部31は、例えば表示装置35の画面に対応して配置され、ユーザの指FG又はスタイラスペンによって操作が可能なタッチパネル又はタッチパッドを用いて構成されてもよい。 The operation unit 31 is a user interface (UI: User Interface) for notifying the customer utterance evaluation unit 33 or the output unit 34 of the operation contents of the user (for example, the store manager), for example, a pointing device such as a mouse or a keyboard. It is. Moreover, the operation part 31 may be comprised using the touch panel or touchpad which is arrange | positioned corresponding to the screen of the display apparatus 35, for example, and can be operated with a user's finger | toe FG or a stylus pen.
 メモリ32は、例えばRAM(Random Access Memory)を用いて構成され、接客評価装置3の各部の動作時のワークメモリとして機能し、更に、接客評価装置3の各部の動作時に必要なデータを記憶する。 The memory 32 is configured by using, for example, a RAM (Random Access Memory), functions as a work memory when each part of the customer service evaluation device 3 operates, and further stores data necessary when each part of the customer service evaluation device 3 operates. .
 接客発話評価部33は、例えばCPU(Central Processing Unit)、MPU(Micro Processing Unit)又はDSP(Digital Signal Processor)を用いて構成され、接客イベント検出部331と、接客発話評価値算定部332とを有する構成である。 The customer service utterance evaluation unit 33 includes, for example, a CPU (Central Processing Unit), an MPU (Micro Processing Unit), or a DSP (Digital Signal Processor), and includes a customer service event detection unit 331 and a service utterance evaluation value calculation unit 332. It is the composition which has.
 検出部の一例としての接客イベント検出部331は、管理サーバ2の管理DB2aの接客イベント情報DB(接客イベントデータ、後述参照)と、所定の業務端末の一例としてのPOS端末5に対する店員の操作履歴を示すPOS操作履歴データ(端末操作履歴データ)とを基に、店員の接客イベントを検出する。接客イベントの検出方法の詳細については後述する。 The customer service event detection unit 331 as an example of the detection unit is a customer service event information DB (customer service event data, see later) of the management DB 2a of the management server 2 and a store clerk's operation history for the POS terminal 5 as an example of a predetermined business terminal. The customer service event of the store clerk is detected based on the POS operation history data (terminal operation history data). Details of the service event detection method will be described later.
 算定部の一例としての接客発話評価値算定部332は、接客イベント検出部331により検出された接客イベントにおいて、レコーダ装置4に格納(記憶)された店員の音声データを基に、POS端末5の操作時における所定の発話想定キーワードに対応する接客発話評価値を算定する。接客発話評価値の算定方法の詳細については後述する。 The customer service utterance evaluation value calculation unit 332 as an example of a calculation unit is based on the clerk's voice data stored (stored) in the recorder device 4 in the customer service event detected by the customer service event detection unit 331. A service utterance evaluation value corresponding to a predetermined utterance assumed keyword at the time of operation is calculated. Details of the method for calculating the customer service utterance evaluation value will be described later.
 出力部34は、例えばCPU、MPU又はDSPを用いて構成され、接客発話評価値算定部332により算定された接客発話評価値を、店員の識別情報(例えば店員が装着する名札に印字された、店員の識別情報を示すバーコード)と店員の接客位置(例えば座標情報)及び接客時刻により特定される店員の音声データに対応付けてメモリ32又は管理サーバ2の管理DB2aに記憶する。 The output unit 34 is configured using, for example, a CPU, MPU, or DSP, and the customer service utterance evaluation value calculated by the customer service utterance evaluation value calculation unit 332 is printed on identification information of the store clerk (for example, on a name tag worn by the store clerk, The data is stored in the memory 32 or the management DB 2a of the management server 2 in association with the clerk's voice data specified by the clerk's customer service position (for example, coordinate information) and the customer service time.
 また、出力部34は、接客モニタリングシステム100の各装置との間のネットワークNWを介した通信機能(有線通信機能、無線通信機能)を有するとともに、表示装置35やスピーカ装置36の動作を制御し、ユーザの所定の入力操作に応じて、接客モニタリングシステム100に関する各種画面を表示装置35に表示させたり、マイク装置M1,…,MLから送信された音声パケットを受信してスピーカ装置36から出力させたりする。 The output unit 34 has a communication function (wired communication function, wireless communication function) via the network NW with each device of the customer service monitoring system 100, and controls the operation of the display device 35 and the speaker device 36. Various screens related to the customer service monitoring system 100 are displayed on the display device 35 according to a predetermined input operation by the user, or voice packets transmitted from the microphone devices M1,... ML are received and output from the speaker device 36. Or
 表示部の一例としての表示装置35は、例えばLCD(Liquid Crystal Display)又は有機EL(Electroluminescence)を用いて構成され、ユーザの入力操作に応じて、出力部34の制御の下で、例えば接客モニタリングシステム100に関する各種画面(後述参照)を表示する。 The display device 35 as an example of the display unit is configured by using, for example, an LCD (Liquid Crystal Display) or an organic EL (Electroluminescence), for example, customer monitoring under the control of the output unit 34 according to a user input operation. Various screens (see later) related to the system 100 are displayed.
 音声出力部の一例としてのスピーカ装置36は、マイク装置M1,…,MLから送信された音声パケットに含まれる音声データを出力する。なお、表示装置35及びスピーカ装置36は、接客評価装置3と一体的に構成されてもよいし、異なる装置でもよい。 The speaker device 36 as an example of the audio output unit outputs audio data included in an audio packet transmitted from the microphone devices M1, ..., ML. The display device 35 and the speaker device 36 may be configured integrally with the customer service evaluation device 3 or may be different devices.
 音声データ抽出部の一例としてのモニタリングデータ抽出部38は、レコーダ装置4に記憶されているモニタリングデータ4aから、接客発話評価部33の接客発話評価値算定部332が店員の接客発話評価値を算定するために必要な所定時間間隔(例えば10~20秒程度)毎のモニタリングデータ4ak(k:1以上の整数、例えば図38参照)に切り取って抽出する。モニタリングデータ抽出部38におけるモニタリングデータの抽出処理の詳細は図38~図41を参照して後述する。 The monitoring data extracting unit 38 as an example of the voice data extracting unit calculates the customer service utterance evaluation value of the store clerk from the monitoring data 4a stored in the recorder device 4 by the customer service utterance evaluation value calculator 332 of the customer service utterance evaluation unit 33. The monitoring data 4ak (k is an integer of 1 or more, for example, see FIG. 38) is extracted every predetermined time interval (for example, about 10 to 20 seconds) necessary for the extraction. Details of the monitoring data extraction processing in the monitoring data extraction unit 38 will be described later with reference to FIGS.
 第2記憶部の一例としてのレコーダ装置4は、例えば接客イベントの際に発話した店員の音声データとPOS端末5の操作履歴を示すPOS操作履歴データとを対応付けてモニタリングデータ4aとして記憶する。店員の音声データは、少なくとも1個のマイク装置M1,…,MLにより収音されてレコーダ装置4に送信された音声データである。POS操作履歴データは、例えばPOS端末5において取得され、更に、管理サーバ2又は接客評価装置3がPOS端末5から取得してレコーダ装置4に送信されたデータである。 The recorder device 4 as an example of the second storage unit stores, as monitoring data 4a, voice data of a clerk uttered at a customer service event and POS operation history data indicating an operation history of the POS terminal 5, for example. The clerk's voice data is voice data collected by at least one microphone device M1, ..., ML and transmitted to the recorder device 4. The POS operation history data is, for example, data acquired at the POS terminal 5 and further acquired by the management server 2 or the customer service evaluation device 3 from the POS terminal 5 and transmitted to the recorder device 4.
 業務端末の一例としてのPOS端末5は、店舗のレジカウンターに設置され、入力装置51と、表示装置52と、メモリ53とを有する構成である。POS端末5は、例えば店舗の売上情報や各商品の価格に関する情報をメモリ53に記憶する。なお、図2から図5では、POS端末5は1個のみ図示されているが、複数のPOS端末5がネットワークNWを介して接続されてもよい。 A POS terminal 5 as an example of a business terminal is installed at a cash register counter in a store and has an input device 51, a display device 52, and a memory 53. The POS terminal 5 stores, for example, store sales information and information regarding the price of each product in the memory 53. 2 to 5, only one POS terminal 5 is illustrated, but a plurality of POS terminals 5 may be connected via the network NW.
 入力装置51は、操作部31と同様に、ユーザ(例えば店員)の入力操作を受け付けてPOS端末5に通知するためのユーザインターフェース(UI)であり、例えばマウス、キーボード等のポインティングデバイスである。また、入力装置51は、例えば表示装置52の画面に対応して配置され、ユーザの指FG又はスタイラスペンによって操作が可能なタッチパネル又はタッチパッドを用いて構成されてもよい。 As with the operation unit 31, the input device 51 is a user interface (UI) for receiving an input operation of a user (for example, a store clerk) and notifying the POS terminal 5, and is, for example, a pointing device such as a mouse or a keyboard. Further, the input device 51 may be configured using a touch panel or a touch pad that is arranged corresponding to the screen of the display device 52 and can be operated by a user's finger FG or stylus pen, for example.
 表示装置52は、表示装置35と同様にLCD(Liquid Crystal Display)又は有機EL(Electroluminescence)を用いて構成され、ユーザの入力操作に応じて、店舗の売上情報、各商品の価格に関する情報や商品の精算に関する画面を表示する。 The display device 52 is configured by using an LCD (Liquid Crystal Display) or an organic EL (Electroluminescence) in the same manner as the display device 35, and according to a user input operation, store sales information, information on the price of each product, and products Displays a screen related to the settlement of
 メモリ53は、メモリ32と同様にRAM(Random Access Memory)を用いて構成され、POS端末5の各部の動作時のワークメモリとして機能し、更に、POS端末5の各部の動作時に必要なデータを記憶する。 The memory 53 is configured by using a RAM (Random Access Memory) similarly to the memory 32, functions as a work memory when each part of the POS terminal 5 is operated, and further stores data necessary when each part of the POS terminal 5 is operated. Remember.
 図3に示す接客モニタリングシステム100Aでは、図2に示す接客モニタリングシステム100の少なくとも1個のマイク装置M1,…,MLが少なくとも1個の接客者マイク装置SM1,…,SMLに置き換わった構成であり、その他の構成は図2に示す接客モニタリングシステム100と同様である。接客者マイク装置SM1,…,SMLの指向性は、それぞれ製造時から設計仕様により既に決められているものであり、変更することはできない。 In the customer service monitoring system 100A shown in FIG. 3, at least one microphone device M1,..., ML of the customer service monitoring system 100 shown in FIG. 2 is replaced with at least one customer microphone device SM1,. The other configuration is the same as that of the customer service monitoring system 100 shown in FIG. The directivities of the customer microphone devices SM1,..., SML are already determined by design specifications from the time of manufacture, and cannot be changed.
 収音部の一例としての少なくとも1個の接客者マイク装置SM1,…,SMLは、例えばピンマイクを用いて構成され、店舗内の各々の店員に個々に装着され、該当する店員の音声を収音し、収音により得られた店員の音声データをレコーダ装置4に送信する。なお、接客者マイク装置SM1,…,SMLは、従業員(店員)の音声のみ収音するものではなく、例えば顧客と従業員(店員)とが会話している際に顧客が発話した声も漏れて収音することはあり得る。 The at least one customer microphone device SM1,..., SML as an example of the sound collection unit is configured by using, for example, a pin microphone and is individually attached to each store clerk in the store to collect the sound of the corresponding store clerk. Then, the clerk's voice data obtained by sound collection is transmitted to the recorder device 4. Note that the customer microphone devices SM1,..., SML do not pick up only the voice of the employee (store clerk), for example, the voice spoken by the customer when the customer and the employee (clerk) are talking. It is possible to leak and pick up sound.
 なお、接客者マイク装置は接客用途に設けられることが多いため、本実施形態の接客者マイク装置SM1,…,SMLは接客用途に設けられたことを示す情報を予め保持している。このため、図37に示すように、接客者マイク装置SM1,…,SMLの収音により得られた店員の音声データには、プライバシー保護マークは付与されない。 Since the customer service microphone device is often provided for customer service purposes, the customer service microphone devices SM1,..., SML of this embodiment hold information indicating that they are provided for customer service purposes in advance. Therefore, as shown in FIG. 37, the privacy protection mark is not given to the clerk's voice data obtained by the sound collection of the customer microphone devices SM1,..., SML.
 なお、図3に示す接客モニタリングシステム100Aでは、レコーダ装置4に格納(記憶)されるモニタリングデータ4aは、図2に示す接客モニタリングシステム100と同様に、店員毎のPOS操作履歴データと接客者マイク装置(例えば接客者マイク装置SM1)により収音された店員の音声データである。 In the customer service monitoring system 100A shown in FIG. 3, the monitoring data 4a stored (stored) in the recorder device 4 is the same as the customer service monitoring system 100 shown in FIG. This is voice data of a clerk picked up by a device (for example, a customer microphone device SM1).
 図4に示す接客モニタリングシステム100Bでは、図2に示す接客モニタリングシステム100に対し、少なくとも1個のカメラ装置C1,…,CMと、少なくとも1個のセンサ装置S1,…,SNとが更に追加された構成であり、その他の構成は図2に示す接客モニタリングシステム100と同様である。 In the customer service monitoring system 100B shown in FIG. 4, at least one camera device C1,..., CM and at least one sensor device S1,. Other configurations are the same as those of the customer service monitoring system 100 shown in FIG.
 撮像部の一例としての少なくとも1個のカメラ装置C1,…、CM(Mは1以上の整数)は、例えば店舗の天井面に固定して設置され、監視カメラ又は防犯カメラとしての機能を有し、ネットワークNWに接続された接客評価装置3からの遠隔操作によって、ズーム機能(例えばズームイン処理、ズームアウト処理)や光軸移動機能(パン、チルト)を用いて、それぞれのカメラ装置C1,…,CMの画角内の映像を撮像する。 At least one camera device C1,..., CM (M is an integer of 1 or more) as an example of an imaging unit is installed fixedly on a ceiling surface of a store, for example, and has a function as a surveillance camera or a security camera. , By remote operation from the customer service evaluation device 3 connected to the network NW, using the zoom function (for example, zoom-in process, zoom-out process) and the optical axis movement function (pan, tilt), respectively, the camera devices C1,. Captures an image within the angle of view of the CM.
 それぞれのカメラ装置C1,…,CMの設置位置や方向は、例えば接客評価装置3のメモリ32に予め登録されており、パンチルトズームに関する制御情報は、接客評価装置3に随時送信されて、映像を構成する各画像位置と指向方向との位置関係は常に関連付けが行われている。また、各カメラ装置C1,…,CMは、例えば全方位カメラである場合には、収音領域の全方位の映像を示す映像データ(即ち、全方位映像データ)、又は全方位映像データに所定の歪み補正処理を施してパノラマ変換して生成した平面映像データを、ネットワークNWを介して接客評価装置3に送信する。なお、それぞれのカメラ装置C1,…,CMの画角及び光軸が固定でもよい。 The installation positions and directions of the respective camera devices C1,..., CM are registered in advance in, for example, the memory 32 of the customer service evaluation device 3, and control information related to pan / tilt / zoom is transmitted to the customer service evaluation device 3 as needed to display images. The positional relationship between each image position and the directing direction is always associated with each other. Further, when each of the camera devices C1,..., CM is an omnidirectional camera, for example, predetermined video data (that is, omnidirectional video data) indicating omnidirectional video in the sound collection area or omnidirectional video data is predetermined. The plane image data generated by performing the panorama conversion by performing the distortion correction process is transmitted to the customer service evaluation device 3 via the network NW. Note that the angle of view and the optical axis of each of the camera devices C1, ..., CM may be fixed.
 出力部34は、例えばユーザの入力操作に応じて、カメラ装置C1,…,CMのうちいずれかのカメラ装置から送信された映像データを表示装置35に表示させる。 The output unit 34 causes the display device 35 to display video data transmitted from any one of the camera devices C1,..., CM according to, for example, a user input operation.
 顧客検出部の一例としての少なくとも1個のセンサ装置S1,…,SNは、店舗に対する顧客の登場又は退場(言い換えると、顧客の入店又は退店)を検出し、検出結果に関する情報をセンサデータとしてレコーダ装置4に送信する。なお、センサ装置S1,…,SNは、接客モニタリングシステム100が検出可能な接客イベントの種類や数に応じて、複数設けられてもよい。 The at least one sensor device S1,..., SN as an example of the customer detection unit detects the appearance or exit of the customer from the store (in other words, the customer enters or exits the store), and information on the detection result is obtained as sensor data. As shown in FIG. A plurality of sensor devices S1,..., SN may be provided according to the type and number of customer service events that can be detected by the customer service monitoring system 100.
 なお、図4に示す接客モニタリングシステム100B,図5に示す接客モニタリングシステム100Cにおいて、店舗内の既定位置(プリセット位置)を収音するマイク装置と既定位置を撮像するカメラ装置とは予め対応付けされている。このため、プリセット位置の識別情報であるプリセットIDとそのプリセット位置を撮像するカメラ装置の識別情報であるカメラIDとは予め対応づけられる。 In the service monitoring system 100B shown in FIG. 4 and the service monitoring system 100C shown in FIG. 5, a microphone device that picks up a predetermined position (preset position) in the store and a camera device that images the predetermined position are associated in advance. ing. For this reason, the preset ID that is the identification information of the preset position and the camera ID that is the identification information of the camera device that captures the preset position are associated in advance.
 なお、図4に示す接客モニタリングシステム100Bでは、レコーダ装置4に格納(記憶)されるモニタリングデータ4bは、図2に示す接客モニタリングシステム100における店員毎のPOS操作履歴データとマイク装置(例えばマイク装置SM1)により収音された店員の音声データ以外に、店員による接客イベントの際に少なくとも1個のカメラ装置C1,…,CMから送信された映像データと少なくとも1個のセンサ装置S1,…,SNから送信されたセンサデータとが更に追加された構成である。 In the customer service monitoring system 100B shown in FIG. 4, the monitoring data 4b stored (stored) in the recorder device 4 includes the POS operation history data for each store clerk and the microphone device (for example, the microphone device) in the customer service monitoring system 100 shown in FIG. In addition to the clerk's voice data collected by SM1), video data transmitted from at least one camera device C1,..., CM and at least one sensor device S1,. The sensor data transmitted from is further added.
 図5に示す接客モニタリングシステム100Cでは、図2に示す接客モニタリングシステム100の少なくとも1個のマイク装置M1,…,MLが少なくとも1個のマイクアレイ装置AM1,…,AMLに置き換わり、接客評価装置3Cにおいて指向性制御部37が追加され、また少なくとも1個のカメラ装置C1,…,CMと、少なくとも1個のセンサ装置S1,…,SNとが更に追加された構成であり、その他の構成は図2に示す接客モニタリングシステム100と同様である。 5, the customer monitoring system 100C shown in FIG. 2 replaces at least one microphone device M1,..., ML of the customer monitoring system 100 with at least one microphone array device AM1,. , A directivity control unit 37 is added, and at least one camera device C1,..., CM and at least one sensor device S1,..., SN are further added. This is the same as the customer service monitoring system 100 shown in FIG.
 収音部の一例としての少なくとも1個のマイクアレイ装置AM1,…,AMLは、所定の収音領域(例えば店舗のレジカウンター)の所定の設置面(例えば天井面)に設置され、収音領域の店員の音声を収音する。少なくとも1個のマイクアレイ装置AM1,…,AMLは、具体的には収音素子の一例としてのマイクロホンを複数含み、複数のマイクロホンを用いて、マイクアレイ装置AM1,…,AMLの設置位置を中心として360°の方向(全方位)からの音声(例えば店員の音声)を収音する。また、マイクアレイ装置AM1,…,AMLは、従業員(店員)の音声のみ収音するものではなく、例えば顧客と従業員(店員)とが会話している際に顧客が発話した声も漏れて収音することはあり得る。 The at least one microphone array device AM1,..., AML as an example of the sound collection unit is installed on a predetermined installation surface (for example, a ceiling surface) of a predetermined sound collection region (for example, a cash register counter of a store). Pick up the clerk's voice. Specifically, at least one microphone array device AM1,..., AML includes a plurality of microphones as an example of sound collection elements, and a plurality of microphones are used to center the installation positions of the microphone array devices AM1,. As an example, a voice (for example, a clerk's voice) from a 360 ° direction (omnidirectional) is collected. In addition, the microphone array devices AM1,..., AML do not pick up only the voice of the employee (store clerk), and for example, the voice spoken by the customer when the customer and the employee (clerk) are talking is leaked. It is possible to pick up sound.
 また、マイクアレイ装置AM1,…,AMLは、店員の顧客に対する接客状況を評価するための接客用途に設けられてもよいし、店舗の防犯等の監視用途に設けられてもよい(図37参照)。監視用途のマイクアレイ装置は、店舗内の人目の付きにくい場所、店舗内のレジカウンターから遠く離れた場所等に設置される。 Further, the microphone array devices AM1,..., AML may be provided for customer service for evaluating the customer service situation of the clerk's customer, or may be provided for surveillance such as store crime prevention (see FIG. 37). ). The microphone array device for monitoring is installed in a place where it is difficult to see in the store, a place far away from the cashier counter in the store, and the like.
 つまり、図37に示すように、接客用途に設けられたマイクアレイ装置は、接客用途に設けられたことを示す情報を予め保持し、監視用途に設けられたアレイマイク装置は、監視用途に設けられたことを示す情報を予め保持している。監視用途に設けられたマイクアレイ装置の収音により得られた店員の音声データには、アレイマイク装置の処理により、プライバシー保護マークが付与される。しかし、接客用途に設けられたマイクアレイ装置の収音により得られた店員の音声データには、プライバシー保護マークは付与されない。 That is, as shown in FIG. 37, the microphone array device provided for customer service holds information indicating that it has been provided for customer service, and the array microphone device provided for monitoring is provided for monitoring purposes. Information indicating that it has been received is held in advance. A privacy protection mark is given to the sound data of the store clerk obtained by the sound collection of the microphone array device provided for monitoring purposes by the processing of the array microphone device. However, the privacy protection mark is not given to the clerk's voice data obtained by collecting sound from the microphone array device provided for customer service.
 少なくとも1個のマイクアレイ装置AM1,…,AMLは、各々のマイクロホンにより収音された音声を音声データとして含む音声パケットを、ネットワークNWを介して、レコーダ装置4に送信する。 The at least one microphone array device AM1,..., AML transmits a voice packet including the voice collected by each microphone as voice data to the recorder device 4 via the network NW.
 操作部31は、表示装置35に表示された画面上の画像(例えばカメラ装置C1,…,CMのうちいずれかのカメラ装置により撮像された画像。以下同様。)に対し、ユーザの操作によって指定された画面上の画像の位置を示す座標データを取得して接客発話評価部33又は出力部34に出力する。 The operation unit 31 designates an image on the screen displayed on the display device 35 (for example, an image captured by any one of the camera devices C1,..., CM; the same applies hereinafter) by a user operation. The coordinate data indicating the position of the image on the screen is acquired and output to the customer service utterance evaluation unit 33 or the output unit 34.
 それぞれのカメラ装置C1,…,CMは、各カメラ装置が撮像した映像データが表示装置35の画面上に表示された状態でユーザの指FG又はスタイラスペンによって任意の位置が指定されると、指定位置の座標データを接客評価装置3から受信し、各カメラ装置から、指定位置に対応する実空間上の位置(以下、単に「音声位置」と略記する)までの距離、方向(水平角及び垂直角を含む。以下同様。)のデータを算定して接客評価装置3に送信する。なお、いずれのカメラ装置における距離、方向のデータ算定処理は公知技術であるため、説明は省略する。 Each camera device C1,..., CM is designated when an arbitrary position is designated by the user's finger FG or stylus pen in a state in which the video data captured by each camera device is displayed on the screen of the display device 35. The coordinate data of the position is received from the customer service evaluation device 3, and the distance and direction (horizontal angle and vertical) from each camera device to the position in the real space corresponding to the designated position (hereinafter simply referred to as “voice position”). Data including corners, and so on) is calculated and transmitted to the customer service evaluation device 3. Note that the distance and direction data calculation processing in any camera device is a known technique, and thus description thereof is omitted.
 指向性制御部37は、表示装置35の画面上に表示された映像からユーザの位置の指定操作に応じて、その映像を撮像したカメラ装置に対応付けられたいずれかのマイクアレイ装置から、指定位置に対応する音声位置に向かう指向方向を示す座標を算定する。指向性制御部37の指向方向を示す座標の算定方法は公知技術であるため、詳細な説明を省略する。 The directivity control unit 37 designates from one of the microphone array devices associated with the camera device that captured the video in accordance with the user's position designation operation from the video displayed on the screen of the display device 35. Coordinates indicating the directivity direction toward the voice position corresponding to the position are calculated. Since the calculation method of the coordinate which shows the directivity direction of the directivity control part 37 is a well-known technique, detailed description is abbreviate | omitted.
 指向性制御部37は、例えばカメラ装置C1の設置位置から音声位置までの距離、方向のデータをカメラ装置C1から取得し、これらのデータを用いて、例えばマイクアレイ装置AM1(例えばカメラ装置C1とマイクアレイ装置AM1とが予め対応付けられているとする)の設置位置から音声位置に向かう指向方向を示す座標を算定する。例えばカメラ装置C1の筐体を囲むようにマイクアレイ装置AM1の筐体とカメラ装置C1とが一体的に取り付けられている場合には、カメラ装置C1から音声位置までの方向(水平角,垂直角)は、マイクアレイ装置AM1から音声位置までの指向方向を示す座標として用いることができる。 The directivity control unit 37 acquires, for example, data on the distance and direction from the installation position of the camera device C1 to the audio position from the camera device C1, and uses these data, for example, the microphone array device AM1 (for example, the camera device C1 and the like). The coordinates indicating the directivity direction from the installation position of the microphone array apparatus AM1 to the voice position are calculated. For example, when the housing of the microphone array device AM1 and the camera device C1 are integrally attached so as to surround the housing of the camera device C1, the direction from the camera device C1 to the sound position (horizontal angle, vertical angle) ) Can be used as coordinates indicating the directivity direction from the microphone array device AM1 to the voice position.
 なお、カメラ装置C1の筐体とマイクアレイ装置AM1の筐体とが離れて取り付けられている場合には、指向性制御部37は、事前に算定されたキャリブレーションパラメータのデータと、カメラ装置C1から音声位置までの方向(水平角,垂直角)のデータとを用いて、マイクアレイ装置AM1から音声位置までの指向方向を示す座標を算定する。なお、キャリブレーションとは、接客評価装置3Cの指向性制御部37が指向方向を示す座標を算定するために必要となる所定のキャリブレーションパラメータを算定又は取得する動作であり、公知技術により予め行われているとする。 When the housing of the camera device C1 and the housing of the microphone array device AM1 are mounted apart from each other, the directivity control unit 37 calculates calibration parameter data calculated in advance and the camera device C1. The coordinates indicating the directivity direction from the microphone array device AM1 to the voice position are calculated using the data from the direction to the voice position (horizontal angle, vertical angle). The calibration is an operation for calculating or acquiring a predetermined calibration parameter required for the directivity control unit 37 of the customer service evaluation device 3C to calculate the coordinates indicating the directivity direction. Suppose that
 指向方向を示す座標は、マイクアレイ装置AM1から音声位置に向かう指向方向の水平角と、マイクアレイ装置AM1から音声位置に向かう指向方向の垂直角とにより示される。なお、音声位置は、操作部31が表示装置35の画面上に表示された映像においてユーザの指FG又はスタイラスペンによって指定された指定位置に対応する実際の監視対象又は収音対象となる現場の位置である(図1参照)。 The coordinates indicating the directivity direction are indicated by a horizontal angle in the directivity direction from the microphone array device AM1 toward the sound position and a vertical angle in the directivity direction from the microphone array device AM1 toward the sound position. The sound position is the actual monitoring target or sound collection target corresponding to the designated position designated by the user's finger FG or stylus pen in the video displayed on the screen of the display device 35 by the operation unit 31. Position (see FIG. 1).
 また、指向性制御部37は、例えばマイクアレイ装置AM1から送信された音声パケットに含まれる店員の音声データを用いて、算定した座標が示す方向に指向性を形成することで音声データを強調処理し、強調処理後の音声データを生成して出力部34に渡す。なお、指向性制御部37における強調処理は、ユーザが選択した映像を撮像したカメラ装置に対応するいずれかのマイクアレイ装置により行われてもよい。 The directivity control unit 37 emphasizes the voice data by forming the directivity in the direction indicated by the calculated coordinates using the voice data of the store clerk included in the voice packet transmitted from the microphone array device AM1, for example. Then, the audio data after the enhancement processing is generated and passed to the output unit 34. Note that the enhancement processing in the directivity control unit 37 may be performed by any microphone array device corresponding to the camera device that captured the video selected by the user.
 なお、図5に示す接客モニタリングシステム100Cでは、レコーダ装置4に格納(記憶)されるモニタリングデータ4bは、図2に示す接客モニタリングシステム100における店員毎のPOS操作履歴データとマイク装置(例えばマイク装置M1)により収音された店員の音声データ以外に、店員による接客イベントの際に少なくとも1個のカメラ装置C1,…,CMから送信された映像データと少なくとも1個のセンサ装置S1,…,SNから送信されたセンサデータとが更に追加された構成である。 In the customer service monitoring system 100C shown in FIG. 5, the monitoring data 4b stored (stored) in the recorder device 4 includes the POS operation history data for each store clerk and the microphone device (for example, the microphone device) in the customer service monitoring system 100 shown in FIG. In addition to the clerk's voice data collected by M1), video data transmitted from at least one camera device C1,..., CM and at least one sensor device S1,. The sensor data transmitted from is further added.
 また、図5ではネットワークNWにマイクアレイ装置AM1,…,AMLが接続されているが、図2に示すマイク装置M1,…MLのうち一部又は全部が接続されてもよいし、更に、図3に示す接客者マイク装置SM1,…SMLのうち一部又は全部が接続されてもよい。 5, the microphone array devices AM1,..., AML are connected to the network NW. However, some or all of the microphone devices M1,... ML shown in FIG. A part or all of the customer microphone devices SM1,... SML shown in FIG.
 次に、本実施形態の接客モニタリングシステム100,100A,100B,100Cに共通する全体の動作手順について、図6を参照して説明する。図6は、本実施形態の接客モニタリングシステムの全体の動作手順の一例を説明するフローチャートである。以下の説明では、基本的にどの接客モニタリングシステム100,100A,100B,100Cの構成を用いても同様に動作するが、説明を簡単にするために、例えば図2に示す接客モニタリングシステム100のシステム構成を用いて説明し、必要に応じて、図3から図5に示す各接客モニタリングシステム100A,100B,100Cの構成を参照して説明する。 Next, an overall operation procedure common to the customer service monitoring systems 100, 100A, 100B, and 100C of the present embodiment will be described with reference to FIG. FIG. 6 is a flowchart for explaining an example of the overall operation procedure of the customer service monitoring system of this embodiment. In the following description, the service monitoring system 100, 100A, 100B, 100C basically operates in the same manner, but for the sake of simplicity, for example, the system of the service monitoring system 100 shown in FIG. It demonstrates using a structure, and it demonstrates with reference to the structure of each customer service monitoring system 100A, 100B, 100C shown to FIGS. 3-5 as needed.
 図6において、接客評価が終了する場合には(S1、YES)、図6に示す接客モニタリングシステム100の動作は終了する。なお、接客評価が終了する場合とは、例えば接客評価装置3にインストールされている接客評価用のアプリケーションの「終了」ボタンが押下された場合や、接客評価装置3がシャットダウンした場合が挙げられるが、これらの場合に限定されない。 In FIG. 6, when the customer service evaluation is completed (S1, YES), the operation of the customer service monitoring system 100 shown in FIG. 6 is completed. Note that the case where the customer service evaluation is completed includes, for example, a case where the “end” button of the customer service evaluation application installed in the customer service evaluation device 3 is pressed, or a case where the customer service evaluation device 3 is shut down. However, the present invention is not limited to these cases.
 一方、接客評価が終了しない場合には(S1、NO)、接客評価装置3のモニタリングデータ抽出部38は、接客発話評価部33の接客発話評価値算定部332が店員の接客発話評価値を算定するために、レコーダ装置4から取得するモニタリングデータ4aを所定時間間隔(例えば10~20秒程度)毎に切り取る。モニタリングデータ抽出部38によって切り取られるモニタリングデータの長さは、あらかじめ想定されている接客イベント(例えば会計完了操作)が起きてから又はその前後に、あらかじめ想定されている接客キーワード(例えば「ありがとうございました」)が発話されて終了するまでの全てが入りうるものとする。その際に接客評価装置3のモニタリングデータ抽出部38は、モニタリングデータ4aを切り取るための起点となる切り取り開始時刻を、その直前のモニタリングデータの切り取り開始時刻から所定時間(例えば1秒程度)シフト処理して設定する(S2、図38参照)。但し、接客評価装置3のモニタリングデータ抽出部38は、初回のモニタリングデータ4aの切り取り時にはシフト処理を行わない。図38は、モニタリングデータ4aのシフト切り出し処理の一例を示す説明図である。ステップS2の処理は、ステップS3において取得されるモニタリングデータ(例えば図38に示すモニタリングデータ4a2)の所定時間間隔の境界時点に接客イベントがあると、切り取られたモニタリングデータ4a2の中から接客キーワードの検出が困難になってしまうことを回避するために設けられる。ステップS2の処理を行うことにより生成される複数のモニタリングデータ4a1,4a2,4a3,…の中には、あらかじめ想定されている接客キーワードの音声データを、途中でとぎらせることなく、最初から最後まで全て記録されているモニタリングデータが含まれることになる。 On the other hand, when the customer service evaluation is not completed (S1, NO), the monitoring data extraction unit 38 of the customer service evaluation device 3 calculates the customer service speech evaluation value calculation unit 332 of the customer service speech evaluation unit 33 to calculate the customer service speech evaluation value of the store clerk. Therefore, the monitoring data 4a acquired from the recorder device 4 is cut out at predetermined time intervals (for example, about 10 to 20 seconds). The length of the monitoring data cut out by the monitoring data extraction unit 38 is based on a customer service keyword (for example, “Thank you very much” after or before or after the customer service event (for example, accounting completion operation) that is assumed in advance. ")" Can be entered until the end of the utterance. At that time, the monitoring data extraction unit 38 of the customer service evaluation device 3 shifts the cutting start time, which is a starting point for cutting the monitoring data 4a, for a predetermined time (for example, about 1 second) from the cutting start time of the immediately preceding monitoring data. (S2, see FIG. 38). However, the monitoring data extraction unit 38 of the customer service evaluation device 3 does not perform a shift process when the monitoring data 4a is cut for the first time. FIG. 38 is an explanatory diagram showing an example of the shift cut-out process of the monitoring data 4a. In the process of step S2, if there is a customer service event at the boundary point of the predetermined time interval of the monitoring data (for example, monitoring data 4a2 shown in FIG. 38) acquired in step S3, the service keyword of the customer service keyword is extracted from the cut monitoring data 4a2. It is provided in order to avoid that detection becomes difficult. Among the plurality of monitoring data 4a1, 4a2, 4a3,... Generated by performing the process of step S2, the voice data of the customer service keyword assumed in advance is not interrupted in the middle, but from the beginning to the end. Monitoring data that has been recorded up to all will be included.
 例えば図38に示すように、レコーダ装置4に記憶されるモニタリングデータ4aに、時刻t1に会計完了操作が発生したことを示すPOS操作履歴データと、時刻t2から時刻t3にわたって店員が「ありがとうございました」と発話した音声データとが記憶されている場合を想定する。この場合、ステップS2においてモニタリングデータ抽出部38によって1回目に切り取られたモニタリングデータ4a1には、会計完了操作に対応する店員の音声データが含まれていないため、接客イベントの検出ができず、正確な接客評価が不可である。次回のステップS2では、接客評価装置3のモニタリングデータ抽出部38は、モニタリングデータ4aを切り取るための起点となる切り取り開始時刻を、モニタリングデータ4a1の切り取り開始時刻から所定時間tsシフト処理したモニタリングデータ4a2を抽出する。切り取りにより取得された所定時間間隔分のモニタリングデータ4ak(k:1以上の整数)のデータサイズは同一である。但し、モニタリングデータ4a2には店員の「ありがとうございました」という発話の全ての音声データが記憶されていないため、接客イベントの検出ができず、同様に正確な接客評価ができない。そして、次回のステップS2では、接客評価装置3のモニタリングデータ抽出部38は、モニタリングデータ4aを切り取るための起点となる切り取り開始時刻を、モニタリングデータ4a2の切り取り開始時刻から所定時間tsシフト処理したモニタリングデータ4a3を抽出する。モニタリングデータ4a3には、店員の「ありがとうございました」という発話の全ての音声データが記憶されているので、接客評価装置3は、接客イベントの検出ができる。 For example, as shown in FIG. 38, in the monitoring data 4a stored in the recorder device 4, the POS operation history data indicating that the accounting completion operation occurred at the time t1, and the salesclerk from the time t2 to the time t3, “Thank you. Is assumed to be stored. In this case, since the monitoring data 4a1 first cut by the monitoring data extraction unit 38 in step S2 does not include the clerk's voice data corresponding to the transaction completion operation, the customer service event cannot be detected accurately. Customer service evaluation is impossible. In the next step S2, the monitoring data extraction unit 38 of the customer service evaluation device 3 monitors the cutting start time, which is the starting point for cutting out the monitoring data 4a, by shifting the monitoring data 4a1 by a predetermined time ts from the cutting start time of the monitoring data 4a1. To extract. The data size of the monitoring data 4ak (k: an integer equal to or larger than 1) for a predetermined time interval acquired by cutting is the same. However, since the monitoring data 4a2 does not store all the voice data of the clerk's “Thank you” utterance, the customer service event cannot be detected, and the customer service evaluation cannot be performed in the same manner. In the next step S2, the monitoring data extraction unit 38 of the customer service evaluation device 3 performs monitoring by shifting the cutting start time, which is a starting point for cutting the monitoring data 4a, by a predetermined time ts from the cutting start time of the monitoring data 4a2. Data 4a3 is extracted. In the monitoring data 4a3, all the voice data of the clerk's “thank you” utterance is stored, so that the customer service evaluation device 3 can detect the customer service event.
 このように、接客評価装置3は、モニタリングデータ抽出部38により抽出された常に同じ長さ(つまり、所定時間ts)であるモニタリングデータ4a1,4a2,4a3,…のそれぞれにおいて、接客キーワードが含まれているか否かを、例えば音声認識により検出を試みるだけでよい。音声の発話開始時刻及び発話終了時刻の検出や音声認識すべき範囲の設定などの処理は不要となる。従って、あらかじめ想定されている接客キーワードを確実に検出することができる。なお、モニタリングデータの切り取り開始時刻のシフト処理のバリエーションについては図39~図41を参照して後述する。また、モニタリングデータ4bの切り取りの方法についても、モニタリングデータ4aの切り取りの方法と同様であり、説明を省略する。 As described above, the customer service evaluation device 3 includes the customer service keywords in each of the monitoring data 4a1, 4a2, 4a3,... That is always the same length (that is, the predetermined time ts) extracted by the monitoring data extraction unit 38. For example, it is only necessary to try to detect whether or not it is detected by voice recognition. Processing such as detection of speech start time and speech end time and setting of a range to be recognized by speech are not required. Therefore, it is possible to reliably detect a customer service keyword assumed in advance. Note that variations of the monitoring data cut start time will be described later with reference to FIGS. The method for cutting the monitoring data 4b is the same as the method for cutting the monitoring data 4a, and the description thereof is omitted.
 ステップS2の後、接客評価装置3のモニタリングデータ抽出部38は、ステップS2において設定された開始時刻から所定時間間隔(例えば10秒程度)毎のモニタリングデータ4aをレコーダ装置4から取得し(S3)、取得したモニタリングデータ4a(具体的には、モニタリングデータ4aに含まれる店員のPOS操作履歴データ及び音声データ)とモニタリングデータ4aの開始時刻及び終了時刻とを対応付けてメモリ32に保持する(S4)。なお、終了時刻は、開始時刻から所定時間間隔の時間を加算した時刻である。 After step S2, the monitoring data extraction unit 38 of the customer service evaluation device 3 acquires monitoring data 4a from the recorder device 4 every predetermined time interval (for example, about 10 seconds) from the start time set in step S2 (S3). The acquired monitoring data 4a (specifically, the clerk's POS operation history data and voice data included in the monitoring data 4a) and the start time and end time of the monitoring data 4a are stored in the memory 32 in association with each other (S4). ). The end time is a time obtained by adding a predetermined time interval from the start time.
 ステップS4の後、接客評価装置3の接客発話評価部33において、イベント検出可否判定処理が行われ(S5)、イベント検出可否フラグが「可」に設定されていれば(S6、YES)、接客評価装置3の出力部34は、ステップS4においてメモリ32に保持した所定時間間隔毎のモニタリングデータ4a(言い換えると、レコーダ装置4から取得した所定時間間隔毎のモニタリングデータ4a)を接客発話評価部33の各部(つまり、接客イベント検出部331,接客発話評価値算定部332)に渡す(S7)。ステップS7の後、接客評価装置3の接客発話評価部33の接客イベント検出部331において、接客イベント検出処理が行われる(S8)。 After step S4, the customer service utterance evaluation unit 33 of the customer service evaluation device 3 performs an event detection availability determination process (S5), and if the event detection availability flag is set to “permitted” (S6, YES), the customer service is performed. The output unit 34 of the evaluation device 3 receives the monitoring data 4a for each predetermined time interval held in the memory 32 in step S4 (in other words, the monitoring data 4a for each predetermined time interval acquired from the recorder device 4). (I.e., customer service event detector 331, customer service utterance evaluation value calculator 332) (S7). After step S7, a service event detection process is performed in the service event detection unit 331 of the service utterance evaluation unit 33 of the service evaluation device 3 (S8).
 一方、イベント検出可否フラグが「否」に設定されている場合(S6、NO)又はステップS8の後、接客モニタリングシステム100の動作はステップS1に戻る。 On the other hand, if the event detection enable / disable flag is set to “NO” (S6, NO) or after step S8, the operation of the customer service monitoring system 100 returns to step S1.
 図7は、イベント検出可否判定処理の詳細な動作手順の一例を説明するフローチャートである。図7において、接客評価装置3の接客発話評価部33は、ステップS4においてメモリ32に保持した所定時間間隔毎のモニタリングデータ4aの所定領域(例えばヘッダ領域、ペイロード領域の一部又はその他のオプション領域)に、顧客のプライバシー保護を示す所定情報の一例としてのプライバシー保護マークが含まれていると判定すると(S5-1、YES)、接客イベントの検出処理の有無を示すイベント検出可否フラグを「否」(つまり、接客イベント処理は行わずに省略すること)に設定する(S5-2)。ステップS5-2の後、接客モニタリングシステム100の動作はステップS6に進む。 FIG. 7 is a flowchart for explaining an example of a detailed operation procedure of the event detection availability determination process. In FIG. 7, the customer service utterance evaluation unit 33 of the customer service evaluation device 3 performs a predetermined area (for example, a header area, a part of the payload area, or other optional areas) of the monitoring data 4a for each predetermined time interval held in the memory 32 in step S4. ) Includes a privacy protection mark as an example of predetermined information indicating customer privacy protection (S5-1, YES), an event detection enable / disable flag indicating presence / absence of customer service event detection processing is set to “No”. (That is, omit the service event process without performing the service event process) (S5-2). After step S5-2, the operation of the customer service monitoring system 100 proceeds to step S6.
 一方、接客発話評価部33は、ステップS4においてメモリ32に保持した所定時間間隔毎のモニタリングデータ4aの所定領域(例えばヘッダ領域、ペイロード領域の一部又はその他のオプション領域)に、顧客のプライバシー保護を示す所定情報の一例としてのプライバシー保護マークが含まれていないと判定すると(S5-1、NO)、接客評価装置3の接客発話評価部33は、店舗に入店した顧客の音声がモニタリングデータ4aに含まれていないかを判定する(S5-3)。 On the other hand, the customer service utterance evaluation unit 33 protects the privacy of the customer in a predetermined area (for example, a header area, a part of the payload area, or other optional areas) of the monitoring data 4a held in the memory 32 in step S4 at every predetermined time interval. If it is determined that the privacy protection mark as an example of the predetermined information is not included (S5-1, NO), the customer utterance evaluation unit 33 of the customer service evaluation device 3 uses the monitoring data as the voice of the customer who entered the store. It is determined whether it is not included in 4a (S5-3).
 例えば、接客発話評価部33は、モニタリングデータ4aに含まれる音声データに、顧客が店舗において発話する可能性が高いキーワードが含まれている(より具体的には、例えばモニタリングデータ4aに含まれる音声データに対して顧客が店舗において発話する可能性が高いキーワードのワードスポッティング処理結果が所定のレベル以上)と判定した場合には、顧客の音声がモニタリングデータ4aに含まれていると判定する(S5-4、YES)。 For example, the customer service utterance evaluation unit 33 includes keywords in the voice data included in the monitoring data 4a that are likely to be uttered by the customer in the store (more specifically, for example, the voice included in the monitoring data 4a). If it is determined that the word spotting processing result of a keyword that is likely to be spoken by the customer in the store with respect to the data is equal to or higher than a predetermined level), it is determined that the voice of the customer is included in the monitoring data 4a (S5). -4, YES).
 又は、接客発話評価部33は、店員に個々に装着される接客者マイク装置において、予め登録された店員以外の音声が収音されている(より具体的には、例えば収音された音声データに対して予め登録された店員の声紋認識結果が所定のレベル以下)と判定した場合には、顧客の音声がモニタリングデータ4aに含まれていると判定してもよい(S5-4、YES)。 Alternatively, the customer service utterance evaluation unit 33 collects voices other than the clerk registered in advance in the customer mic device individually attached to the clerk (more specifically, for example, collected voice data If the voice clerk recognition result of the clerk registered in advance is determined to be below a predetermined level), it may be determined that the customer's voice is included in the monitoring data 4a (S5-4, YES). .
 又は、接客評価装置3Cの接客発話評価部33は、モニタリングデータ4bに含まれる映像データに対して画像処理することで、予め登録された店員の顔画像以外の顔を検出し、更にその検出された顔の位置から発せられた音声、或いはその検出された顔の位置に指向性を形成し強調処理された音声に人の音声が含まれていると判定した場合には、顧客の音声がモニタリングデータ4aに含まれていると判定してもよい(S5-4、YES)。 Alternatively, the customer service utterance evaluation unit 33 of the customer service evaluation device 3C detects a face other than the face image of the store clerk registered in advance by performing image processing on the video data included in the monitoring data 4b, and further detects the detected face. If it is determined that the voice uttered from the position of the face or the voice of the detected face position that forms directivity and is emphasized includes human voice, the voice of the customer is monitored It may be determined that it is included in the data 4a (S5-4, YES).
 接客評価装置3(又は接客評価装置3C)の接客発話評価部33は、顧客音声の有無判定処理結果として顧客の音声が含まれていると判定した場合には(S5-4、YES)、イベント検出可否フラグを「否」に設定する(S5-5)。 When the customer service utterance evaluation unit 33 of the customer service evaluation device 3 (or the customer service evaluation device 3C) determines that the customer's voice is included as a result of the customer voice presence / absence determination process (S5-4, YES), the event The detection enable / disable flag is set to “NO” (S5-5).
 一方、接客評価装置3(又は接客評価装置3C)の接客発話評価部33は、顧客音声の有無判定処理結果として顧客の音声が含まれていないと判定した場合には(S5-4、NO)、イベント検出可否フラグを「可」に設定する(S5-6)。ステップS5-5の後、及びステップS5-6の後、接客モニタリングシステム100の動作はステップS6に進む。 On the other hand, when the customer service utterance evaluation unit 33 of the customer service evaluation device 3 (or customer service evaluation device 3C) determines that the customer's voice is not included as a result of the customer voice presence / absence determination process (S5-4, NO). Then, the event detection enable / disable flag is set to “enable” (S5-6). After step S5-5 and after step S5-6, the operation of the customer service monitoring system 100 proceeds to step S6.
 次に、図6に示すステップS8の接客イベント検出処理を説明する前に、図2から図5に示す接客モニタリングシステム100,100A,100B,100Cに対応した所定の接客イベント毎の判定条件を含む接客イベントデータの一例としての接客イベント情報DBの例について、図11から図14を参照して説明する。図11から図14に示す各接客イベント情報DBは、管理サーバ2の管理DB2aに格納されている。 Next, before describing the customer service event detection process in step S8 shown in FIG. 6, the determination conditions for each predetermined customer service event corresponding to the customer service monitoring systems 100, 100A, 100B, and 100C shown in FIGS. 2 to 5 are included. An example of a customer service event information DB as an example of customer service event data will be described with reference to FIGS. Each customer service event information DB shown in FIGS. 11 to 14 is stored in the management DB 2 a of the management server 2.
 図11は、図2に示す接客モニタリングシステム100に対応した接客イベント情報DBの一例を示す図である。図12は、図3に示す接客モニタリングシステム100Aに対応した接客イベント情報DBの一例を示す図である。図13は、図4に示す接客モニタリングシステム100Bに対応した接客イベント情報DBの一例を示す図である。図14は、図5に示す接客モニタリングシステム100Cに対応した接客イベント情報DBの一例を示す図である。図12から図14の説明では、図11の説明と重複する内容の説明は省略し、異なる内容について説明する。 FIG. 11 is a diagram showing an example of a customer service event information DB corresponding to the customer service monitoring system 100 shown in FIG. FIG. 12 is a diagram illustrating an example of a customer service event information DB corresponding to the customer service monitoring system 100A illustrated in FIG. FIG. 13 is a diagram illustrating an example of a customer service event information DB corresponding to the customer service monitoring system 100B illustrated in FIG. FIG. 14 is a diagram illustrating an example of a customer service event information DB corresponding to the customer service monitoring system 100C illustrated in FIG. In the description of FIG. 12 to FIG. 14, description of contents overlapping with those of FIG. 11 will be omitted, and different contents will be described.
 図11に示す接客イベント情報DBは、接客イベントIDと、接客イベント名と、接客イベント判定条件(つまり、モニタリングデータ4aに接客イベントが検出されたか否かを判定するための条件)と、接客イベント出力情報(つまり、接客イベントが検出されたときに出力する情報)との各項目に対応するデータの種類、種別が規定される。 The customer service event information DB shown in FIG. 11 includes a customer service event ID, a customer service event name, a customer service event determination condition (that is, a condition for determining whether or not a customer service event is detected in the monitoring data 4a), and a customer service event. The type and type of data corresponding to each item of output information (that is, information output when a customer service event is detected) are defined.
 図11に示す接客イベント情報DBの接客イベント判定条件では、接客イベントの検出トリガがPOS端末5に対する所定の動作(POS動作)が行われたことであると規定されている。 The customer service event determination condition of the customer service event information DB shown in FIG. 11 stipulates that a customer service event detection trigger is a predetermined operation (POS operation) performed on the POS terminal 5.
 図11に示す接客イベント出力情報では、プリセットIDと接客者IDと接客イベントIDとが出力されると規定されている。 The customer service event output information shown in FIG. 11 stipulates that a preset ID, customer service ID, and customer service event ID are output.
 図12に示す接客イベント情報DBの接客イベント判定条件では、接客イベントの検出トリガが接客イベント毎に異なり、具体的にはPOS端末5に対する所定の動作(POS動作)が行われたこと、音声データ中に特定のキーワードが含まれることがそれぞれ規定されている。 In the customer service event determination condition of the customer service event information DB shown in FIG. 12, the customer service event detection trigger is different for each customer service event. Specifically, a predetermined operation (POS operation) on the POS terminal 5 is performed, voice data Each of them is specified to contain a specific keyword.
 図12に示す接客イベント情報DBの接客イベント出力情報では、接客イベント毎に異なり、具体的にはプリセットID(店舗内の既定位置の識別情報。以下同様)と接客者ID(店員の識別情報。以下同様)と接客イベントIDとの組み合わせと、単に接客者ID(店員が装着している接客者マイク装置の識別情報。以下同様)のみが規定されている。 The customer service event output information in the customer service event information DB shown in FIG. 12 is different for each customer service event, and specifically, a preset ID (identification information of a predetermined position in the store; the same applies hereinafter) and a customer service ID (store employee identification information. The same applies hereinafter) and the customer service event ID, and only the customer service ID (identification information of the customer microphone device worn by the store clerk; the same applies hereinafter) is defined.
 図13に示す接客イベント情報DBは、接客イベントIDと、全接客者対象イベントであるか否かを示す項目と、接客イベント名と、接客イベント判定条件(つまり、モニタリングデータ4aに接客イベントが検出されたか否かを判定するための条件)と、接客イベント出力情報(つまり、接客イベントが検出されたときに出力する情報)との各項目に対応するデータの種類、種別が規定される。 The customer service event information DB shown in FIG. 13 includes a customer service event ID, an item indicating whether or not the event is a customer service event, a customer service event name, a customer service event determination condition (that is, a customer service event is detected in the monitoring data 4a). And the type of data corresponding to each item of customer service event output information (that is, information output when a customer service event is detected) are defined.
 図13に示す接客イベント情報DBの接客イベント判定条件では、接客イベントの検出トリガが接客イベント毎に異なり、具体的には店舗入り口付近に設置されたセンサ装置S1(例えば自動ドア)が開閉動作したことを検出したこと、所定のプリセットIDに対応する既定位置(プリセット位置)に店員が滞在して更に所定の来客者位置プリセットIDに対応する位置(即ち、接客イベント時に顧客(来客者)がいる可能性が高い位置)に顧客が所定時間(例えば5秒程度)滞在したこと、POS端末5に対する所定の動作(POS動作)が行われたことがそれぞれ規定されている。 In the customer service event determination condition of the customer service event information DB shown in FIG. 13, the customer service event detection trigger is different for each customer service event. Specifically, the sensor device S1 (for example, an automatic door) installed near the store entrance is opened and closed. When a store clerk stays at a predetermined position (preset position) corresponding to a predetermined preset ID and there is a position corresponding to the predetermined visitor position preset ID (that is, there is a customer (customer) at the customer service event) It is specified that the customer stays at a highly likely position) for a predetermined time (for example, about 5 seconds) and that a predetermined operation (POS operation) for the POS terminal 5 has been performed.
 図13に示す接客イベント情報DBの接客イベント出力情報では、接客イベント毎に異なり、具体的にはマイクID(後述参照)とカメラID(後述参照)と接客者IDと接客イベントIDとの組み合わせと、プリセットIDと接客者IDと接客イベントIDとの組み合わせとがそれぞれ規定されている。 The customer service event output information of the customer service event information DB shown in FIG. 13 is different for each customer service event, and specifically includes a combination of a microphone ID (see later), a camera ID (see later), a customer ID, and a customer service event ID. A combination of a preset ID, a customer service ID, and a customer service event ID is defined.
 図14に示す接客イベント情報DBの接客イベント出力情報では、接客イベント毎に異なり、具体的には接客者位置座標とカメラIDと接客者IDと接客イベントIDとの組み合わせと、プリセットIDと接客者IDと接客イベントIDとの組み合わせとがそれぞれ規定されている。接客者位置座標は、指向性制御部37が各店員の音声データを収音したマイクアレイ装置から各店員に向かう方向に音声の指向性を形成する際に用いられる。 The customer service event output information in the customer service event information DB shown in FIG. 14 differs for each customer service event. Specifically, the customer service location coordinates, camera ID, customer service ID, customer service event ID combination, preset ID, and customer service A combination of ID and customer service event ID is defined. The customer service position coordinates are used when the directivity control unit 37 forms the sound directivity in the direction from the microphone array device that picks up the sound data of each store clerk toward each store clerk.
 次に、図11から図14に示す接客イベント情報DBの具体例を用いた接客イベント検出処理(図6に示すステップS8参照)の詳細について、図8及び図9を参照して説明する。図8は、接客イベント検出処理の詳細な動作手順の一例を説明するフローチャートである。図9は、接客イベント情報処理の詳細な動作手順の一例を説明するフローチャートである。 Next, details of the service event detection process (see step S8 shown in FIG. 6) using the specific example of the service event information DB shown in FIGS. 11 to 14 will be described with reference to FIGS. FIG. 8 is a flowchart for explaining an example of a detailed operation procedure of the customer service event detection process. FIG. 9 is a flowchart illustrating an example of a detailed operation procedure of customer service event information processing.
 図8及び図9の説明では、説明を具体的かつ分かりやすくするために、図2から図5に示す接客モニタリングシステム100,100A,100B,100Cの各システム構成に対応する図11から図14に示す接客イベント情報DBの各レコードの内容を参照して具体的に説明する。また、図11から図14に示す接客イベント情報DBにおいて重複する接客イベントIDのレコードが規定されている場合には、重複する説明は省略し、異なる内容について説明する。 In the description of FIGS. 8 and 9, in order to make the description concrete and easy to understand, FIGS. 11 to 14 corresponding to the system configurations of the customer monitoring systems 100, 100A, 100B, and 100C shown in FIGS. A specific description will be given with reference to the contents of each record of the customer service event information DB shown. In addition, when duplicate customer service event ID records are defined in the customer service event information DB shown in FIG. 11 to FIG. 14, duplicate explanation will be omitted and different contents will be explained.
 (図2に示す接客モニタリングシステム100における接客イベント検出処理)
 先ず、接客イベント検出部331は、開始時刻及び終了時刻が定められた所定時間間隔(例えば10秒程度)毎のモニタリングデータ4aを接客発話評価部33から受け取り(S8-1)、管理サーバ2の管理DB2aに格納されている接客イベント情報DB(図11参照)を読み込む(S8-2)。
(Service Event Detection Processing in the Service Monitoring System 100 shown in FIG. 2)
First, the customer service event detection unit 331 receives the monitoring data 4a from the customer service utterance evaluation unit 33 at predetermined time intervals (for example, about 10 seconds) in which the start time and the end time are determined (S8-1). The customer service event information DB (see FIG. 11) stored in the management DB 2a is read (S8-2).
 接客イベント検出部331は、未取得である接客イベント情報DBの1行目のレコード(接客イベントID「EID1」、接客イベント名「会計完了挨拶」)を取得し(S8-3)、接客イベント情報処理を開始する。接客イベント検出部331は、モニタリングデータ4aから照合対象データとしてのPOS操作履歴データを取得し(S11)、POS操作履歴データから会計完了操作の接客イベント判定条件の検出トリガ(即ち、POS端末5に対する所定の動作(操作)があったか否か)を満たすか否かを照合する(S12)。 The customer service event detection unit 331 acquires a record (service event ID “EID1”, customer service event name “accounting completion greeting”) in the first line of the customer service event information DB that has not been acquired (S8-3), and receives customer service event information. Start processing. The customer service event detection unit 331 acquires the POS operation history data as the verification target data from the monitoring data 4a (S11), and detects a customer event determination condition detection trigger for the transaction completion operation from the POS operation history data (that is, for the POS terminal 5). It is verified whether or not a predetermined operation (operation) has been satisfied (S12).
 POS操作履歴データから会計完了操作の接客イベント判定条件の検出トリガを満たさない場合には(S13、NO)、図9に示す接客イベント情報処理は終了し、接客イベント検出部331の処理はステップS8-5に進む。 If the detection trigger of the customer service event determination condition for the transaction completion operation is not satisfied from the POS operation history data (NO in S13), the customer service event information processing shown in FIG. 9 is terminated, and the service of the customer service event detection unit 331 performs step S8. Proceed to -5.
 一方、POS操作履歴データから会計完了操作の接客イベント判定条件の検出トリガを満たす場合には(S13、YES)、接客イベント検出部331は、図11に示す接客イベント情報DBの接客イベント出力情報(具体的には、該当するプリセットID(1~PNのうち該当するもの)、接客者ID(店員の識別情報:1~ENのうち該当するもの)、接客イベントID)を接客状況DB(図10参照)に格納(保持)する(S14)。 On the other hand, when the detection trigger of the customer service event determination condition for the transaction completion operation is satisfied from the POS operation history data (S13, YES), the customer service event detection unit 331 receives the customer service event output information ( Specifically, the corresponding preset ID (corresponding one of 1 to PN), customer service ID (identification information of store clerk: corresponding to 1 to EN), customer service event ID) are stored in the customer service status DB (FIG. 10). (Refer to) is stored (held) (S14).
 なお、検出された接客イベントにおいて、該当するPOS端末5を操作した接客者(店員)の識別情報(接客者ID:1~EN)は、POS端末5の操作開始時に例えば名札に印字されたバーコードをバーコードリーダによって読み取られた接客者IDが用いられる。 In the detected customer service event, the identification information (customer ID: 1 to EN) of the customer (store clerk) who operated the corresponding POS terminal 5 is, for example, a bar printed on the name tag when the operation of the POS terminal 5 is started. The customer ID obtained by reading the code with the barcode reader is used.
 図2に示す接客モニタリングシステム100では、マイク装置M1,…,MLの指向性が製造時から予め決まっているものであり、その指向性を変更することができない。すなわち、図2に示す接客モニタリングシステム100ではマイクアレイ装置AM1,…,AMLにより音声が収音されていないため(S15、NO)、指向性の形成処理が不可であり、接客イベント検出部331は、モニタリングデータ4aから、検出された接客イベントに対応する店員の音声データを取得し(S16)、接客発話評価値算定部332に音声データと接客イベントIDとを入力する(S18)。検出された接客イベントに対応する店員の音声データは、例えば「会計完了操作」の接客イベントが検出されたPOS端末5に対応付けられるマイク装置により収音された音声データである。 In the customer service monitoring system 100 shown in FIG. 2, the directivity of the microphone devices M1,..., ML is predetermined from the time of manufacture, and the directivity cannot be changed. That is, in the customer service monitoring system 100 shown in FIG. 2, since the voice is not collected by the microphone array devices AM1,..., AML (S15, NO), directivity formation processing is impossible, and the customer service event detection unit 331 is Then, voice data of the clerk corresponding to the detected customer service event is acquired from the monitoring data 4a (S16), and the voice data and customer service event ID are input to the customer service utterance evaluation value calculation unit 332 (S18). The clerk's voice data corresponding to the detected customer service event is, for example, voice data collected by the microphone device associated with the POS terminal 5 in which the customer service event “accounting completion operation” is detected.
 接客発話評価値算定部332は、図15に示す接客発話評価処理を実行し(S19)、接客発話評価出力値を接客状況DBに格納(保持)する(S20)。これにより、図9に示す接客イベント情報処理が終了する。図8において、ステップS8-4の後、接客イベント情報DBの全てのレコードが取得されていない場合には(S8-5、NO)、接客イベント検出部331の処理はステップS8-3に戻る。一方、接客イベント情報DBの全てのレコードが取得された場合には(S8-5、YES)、接客イベント検出部331の処理は終了する。 The customer service utterance evaluation value calculation unit 332 executes the customer service utterance evaluation process shown in FIG. 15 (S19), and stores (holds) the service utterance evaluation output value in the customer service status DB (S20). Thereby, the service event information processing shown in FIG. 9 ends. In FIG. 8, after step S8-4, if not all records in the customer service event information DB have been acquired (S8-5, NO), the process of the customer service event detector 331 returns to step S8-3. On the other hand, when all the records of the customer service event information DB have been acquired (S8-5, YES), the process of the customer service event detector 331 ends.
 (図3に示す接客モニタリングシステム100Aにおける接客イベント検出処理)
 ステップS8-2の後、接客イベント検出部331は、未取得である接客イベント情報DBの2行目のレコード(接客イベントID「EID2」、接客イベント名「入店挨拶」)を取得し(S8-3)、接客イベント情報処理を開始する。接客イベント検出部331は、モニタリングデータ4aから照合対象データとしての接客者マイク装置において収音された全ての音声データを取得し(S11)、全ての音声データを音声認識処理し、その結果に特定キーワード「いらっしゃいませ」が含まれるか否かを照合する(S12)。
(Service event detection process in the service monitoring system 100A shown in FIG. 3)
After step S8-2, the customer service event detecting unit 331 acquires a record (service customer event ID “EID2”, customer service event name “greeting in store”) in the second line of the customer service event information DB that has not been acquired (S8 -3) Start customer service event information processing. The customer service event detection unit 331 acquires all voice data collected by the customer microphone device as verification target data from the monitoring data 4a (S11), performs voice recognition processing on all the voice data, and identifies the result. It is verified whether or not the keyword “I welcome you” is included (S12).
 全ての音声データの音声認識処理結果に特定キーワード「いらっしゃいませ」が含まれない場合には(S13、NO)、図9に示す接客イベント情報処理は終了し、接客イベント検出部331の処理はステップS8-5に戻る。 If the voice recognition process result of all the voice data does not include the specific keyword “Welcome” (S13, NO), the customer service event information processing shown in FIG. Return to S8-5.
 一方、音声データのいずれかの音声認識処理結果に特定キーワード「いらっしゃいませ」が含まれる場合には(S13、YES)、接客イベント検出部331は、図12に示す接客イベント情報DBの接客イベント出力情報(具体的には、接客者ID(店員の識別情報:1~ENのうち該当するもの)、接客イベントID)を接客状況DB(図10参照)に格納(保持)する(S14)。 On the other hand, when the specific keyword “welcome” is included in any voice recognition processing result of the voice data (S13, YES), the customer service event detection unit 331 outputs a customer service event in the customer service event information DB shown in FIG. Information (specifically, customer service ID (store employee identification information: one corresponding to 1 to EN), customer service event ID) is stored (held) in the customer service DB (see FIG. 10) (S14).
 図3に示す接客モニタリングシステム100Aでは、接客者マイク装置SM1,…,SMLの指向性が製造時から予め決まっているものであり、その指向性を変更することができない。すなわち、図3に示す接客モニタリングシステム100Aではマイクアレイ装置AM1,…,AMLにより音声が収音されていないため(S15、NO)、指向性の形成処理が不可であり、接客イベント検出部331は、モニタリングデータ4aから、検出された接客イベントに対応する店員の音声データを取得し(S16)、接客発話評価値算定部332に音声データと接客イベントIDとを入力する(S18)。検出された接客イベントに対応する店員の音声データとは、接客イベント検出部331における音声認識処理結果に特定キーワード「いらっしゃいませ」を含む音声データである。 In the customer service monitoring system 100A shown in FIG. 3, the directivities of the customer microphone devices SM1,..., SML are predetermined from the time of manufacture, and the directivity cannot be changed. That is, in the service monitoring system 100A shown in FIG. 3, since voice is not picked up by the microphone array devices AM1,..., AML (S15, NO), directivity formation processing is impossible, and the service event detection unit 331 Then, voice data of the clerk corresponding to the detected customer service event is acquired from the monitoring data 4a (S16), and the voice data and customer service event ID are input to the customer service utterance evaluation value calculation unit 332 (S18). The salesclerk's voice data corresponding to the detected customer service event is voice data that includes the specific keyword “welcome” in the voice recognition processing result in the customer service event detection unit 331.
 接客発話評価値算定部332は、図15に示す接客発話評価処理を実行し(S19)、接客発話評価出力値を接客状況DBに格納(保持)する(S20)。これにより、図9に示す接客イベント情報処理が終了する。図8において、ステップS8-4の後、接客イベント情報DBの全てのレコードが取得されていない場合には(S8-5、NO)、接客イベント検出部331の処理はステップS8-3に戻る。一方、接客イベント情報DBの全てのレコードが取得された場合には(S8-5、YES)、接客イベント検出部331の処理は終了する。 The customer service utterance evaluation value calculation unit 332 executes the customer service utterance evaluation process shown in FIG. 15 (S19), and stores (holds) the service utterance evaluation output value in the customer service status DB (S20). Thereby, the service event information processing shown in FIG. 9 ends. In FIG. 8, after step S8-4, if not all records in the customer service event information DB have been acquired (S8-5, NO), the process of the customer service event detector 331 returns to step S8-3. On the other hand, when all the records of the customer service event information DB have been acquired (S8-5, YES), the process of the customer service event detector 331 ends.
 (図4に示す接客モニタリングシステム100Bにおける接客イベント検出処理)
 ステップS8-2の後、接客イベント検出部331は、未取得である接客イベント情報DBの1行目のレコード(接客イベントID「EID1」、接客イベント名「入退店挨拶」)を取得し(S8-3)、接客イベント情報処理を開始する。なお、この「入退店挨拶」の接客イベントは、全接客者(全店員)を対象とする接客イベントであるため、接客イベント出力情報として、全店員を把握可能なマイクID及びカメラIDが出力され、以下同様である。接客イベント検出部331は、モニタリングデータ4bから照合対象データとしての映像データとセンサデータに含まれる検出結果(自動ドア開閉履歴データ)とを取得し(S11)、自動ドア開閉履歴データに自動ドアの開閉動作があるか否かを照合する(S12)。
(Service event detection process in the service monitoring system 100B shown in FIG. 4)
After step S8-2, the customer service event detection unit 331 acquires a record (service event ID “EID1”, customer service event name “entrance / exit greeting”) in the first line of the service event information DB that has not been acquired ( S8-3), customer service event information processing is started. In addition, since the customer service event of this “greeting entrance / exit” is a customer service event for all customers (all store clerks), a microphone ID and a camera ID capable of grasping all store clerk are output as customer service event output information. The same applies hereinafter. The customer service event detection unit 331 acquires video data as verification target data and detection results (automatic door opening / closing history data) included in the sensor data from the monitoring data 4b (S11), and the automatic door opening / closing history data includes the automatic door opening / closing history data. It is checked whether or not there is an opening / closing operation (S12).
 自動ドア開閉履歴データが自動ドアの開閉動作を含まない場合には(S13、NO)、図9に示す接客イベント情報処理は終了し、接客イベント検出部331の処理はステップS8-5に戻る。 If the automatic door opening / closing history data does not include an automatic door opening / closing operation (S13, NO), the customer service event information processing shown in FIG. 9 ends, and the process of the customer service event detection unit 331 returns to step S8-5.
 一方、自動ドア開閉履歴データが自動ドアの開閉動作を含む場合には(S13、YES)、接客イベント検出部331は、図13に示す接客イベント情報DBの接客イベント出力情報(具体的には、マイクID(マイク装置の識別情報:1~MNのうち該当するもの)、カメラID(カメラ装置の識別情報:1~CNのうち該当するもの)、接客者ID(店員の識別情報:1~ENのうち該当するもの)、接客イベントID)を接客状況DB(図10参照)に格納(保持)する(S14)。カメラIDは、接客イベント検出部331が所定の映像データを画像処理することで、自動ドアの開閉動作時に各店員が存在した位置を最も近くで撮像するカメラ装置の識別情報として出力される。 On the other hand, when the automatic door opening / closing history data includes an automatic door opening / closing operation (S13, YES), the customer service event detection unit 331 receives the service event output information (specifically, the customer service event information DB shown in FIG. 13). Microphone ID (identifying information of microphone device: 1 to MN), camera ID (identifying information of camera device: 1 to CN), customer service ID (clerk identification information: 1 to EN) And the customer service event ID) are stored (held) in the customer service status DB (see FIG. 10) (S14). The camera ID is output as identification information of a camera device that captures the position where each store clerk is closest when the automatic door opening / closing operation is performed by the customer service event detection unit 331 performing image processing on predetermined video data.
 図4に示す接客モニタリングシステム100Bでは、マイク装置M1・・・MLの指向性が製造時から予め決まっているものであり、その指向性を変更することができない。すなわち、図4に示す接客モニタリングシステム100Bではマイクアレイ装置AM1,…,AMLにより音声が収音されていないため(S15、NO)、指向性の形成処理が不可であり、接客イベント検出部331は、モニタリングデータ4bから、検出された接客イベントに対応する各店員の音声データを取得し(S16)、接客発話評価値算定部332に音声データと接客イベントIDとを入力する(S18)。 In the customer service monitoring system 100B shown in FIG. 4, the directivity of the microphone devices M1... ML is predetermined from the time of manufacture, and the directivity cannot be changed. That is, in the customer service monitoring system 100B shown in FIG. 4, since the voice is not collected by the microphone array devices AM1,..., AML (S15, NO), the directivity forming process is impossible, and the customer service event detecting unit 331 is Then, voice data of each clerk corresponding to the detected customer service event is acquired from the monitoring data 4b (S16), and the voice data and customer service event ID are input to the customer service utterance evaluation value calculation unit 332 (S18).
 検出された接客イベントに対応する各店員の音声データは、接客イベント検出部331が所定の映像データを画像処理することで、自動ドアの開閉動作時に各店員が存在した位置に最も近いと判定されたマイク装置により収音された音声データである。マイク装置からは、その識別情報であるマイクIDが出力される。 The voice data of each store clerk corresponding to the detected customer service event is determined to be closest to the position where each store clerk was present during the automatic door opening / closing operation by image processing the predetermined video data by the customer service event detection unit 331. Voice data collected by the microphone device. From the microphone device, a microphone ID as the identification information is output.
 なお、所定の映像データとは、例えば店舗内の全域が分かるための必要最低限な1個以上のカメラ装置により撮像された映像データ又は複数の映像データの組み合わせであり、該当するカメラ装置は固定でもよいし、ユーザの入力操作に応じて適宜変更されてもよく、以下同様である。 The predetermined video data is, for example, video data or a combination of a plurality of video data captured by at least one camera device necessary for understanding the entire area in the store, and the corresponding camera device is fixed. However, it may be changed as appropriate according to the user's input operation, and so on.
 接客発話評価値算定部332は、図15に示す接客発話評価処理を実行し(S19)、接客発話評価出力値を接客状況DBに格納(保持)する(S20)。これにより、図9に示す接客イベント情報処理が終了する。図8において、ステップS8-4の後、接客イベント情報DBの全てのレコードが取得されていない場合には(S8-5、NO)、接客イベント検出部331の処理はステップS8-3に戻る。 The customer service utterance evaluation value calculation unit 332 executes the customer service utterance evaluation process shown in FIG. 15 (S19), and stores (holds) the service utterance evaluation output value in the customer service status DB (S20). Thereby, the service event information processing shown in FIG. 9 ends. In FIG. 8, after step S8-4, if not all records in the customer service event information DB have been acquired (S8-5, NO), the process of the customer service event detector 331 returns to step S8-3.
 そこで、接客イベント検出部331は、未取得である接客イベント情報DBの2行目のレコード(接客イベントID「EID2」、接客イベント名「会計開始挨拶」)を取得し(S8-3)、接客イベント情報処理を開始する。接客イベント検出部331は、モニタリングデータ4bから照合対象データとしての上述した所定の映像データを取得し(S11)、映像データを画像処理することで、店員が「会計開始挨拶」を行うための既定位置(例えばレジカウンターの操作立ち位置)に存在し、かつ顧客(来店者)が所定の位置(例えばレジカウンターの前又は店舗内に敷かれた所定の待機位置)に所定期間(例えば5秒程度)以上滞在したか否かを照合する(S12)。顧客が所定期間以上滞在した否かを判定することで、顧客がレジカウンターを素通りした場合を除外することが可能となる。 Accordingly, the customer service event detection unit 331 acquires the second line record (customer service event ID “EID2”, customer service event name “accounting start greeting”) of the customer service event information DB that has not been acquired (S8-3). Start event information processing. The customer service event detection unit 331 acquires the predetermined video data as the verification target data from the monitoring data 4b (S11), and performs image processing on the video data so that the store clerk performs a “accounting start greeting”. It exists in a position (for example, the standing position of the cashier counter), and the customer (visitor) is in a predetermined position (for example, a predetermined standby position in front of the cashier counter or in the store) for a predetermined period (for example, about 5 seconds) ) It is verified whether or not the stay has been made (S12). By determining whether or not the customer has stayed for a predetermined period or longer, it is possible to exclude the case where the customer passes through the cashier counter.
 映像データの画像処理結果として、店員が「会計開始挨拶」を行うための既定位置(例えばレジカウンターの操作立ち位置)に存在しない、又は顧客(来店者)が所定の位置(例えばレジカウンターの前又は店舗内に敷かれた所定の待機位置)に所定期間(例えば5秒程度)以上滞在していないと判定された場合には(S13、NO)、図9に示す接客イベント情報処理は終了し、接客イベント検出部331の処理はステップS8-5に戻る。 As a result of the image processing of the video data, the store clerk does not exist at a predetermined position (for example, the operation standing position of the cashier counter) for performing “accounting start greeting”, or the customer (visitor) has a predetermined position (for example, before the cashier counter) Alternatively, if it is determined that the user has not stayed for a predetermined period (for example, about 5 seconds) at a predetermined standby position in the store (S13, NO), the customer service event information processing shown in FIG. 9 ends. Then, the process of the customer service event detection unit 331 returns to step S8-5.
 一方、映像データの画像処理結果として、店員が「会計開始挨拶」を行うための既定位置(例えばレジカウンターの操作立ち位置)に存在し、かつ顧客(来店者)が所定の位置(例えばレジカウンターの前又は店舗内に敷かれた所定の待機位置)に所定期間(例えば5秒程度)以上滞在したと判定された場合には(S13、YES)、接客イベント検出部331は、図13に示す接客イベント情報DBの接客イベント出力情報(具体的には、プリセットID(既定位置の識別情報:1~PNのうち該当するもの)、接客者ID(店員の識別情報:1~ENのうち該当するもの)、接客イベントID)を接客状況DB(図10参照)に格納(保持)する(S14)。 On the other hand, as a result of image processing of the video data, the store clerk is present at a predetermined position (for example, an operation standing position of the cashier counter) for performing “accounting start greeting”, and a customer (visitor) has a predetermined position (for example, cashier counter). If it is determined that the user has stayed for a predetermined period of time (for example, about 5 seconds) at a predetermined standby position in front of or in the store) (S13, YES), the customer service event detection unit 331 is shown in FIG. Customer service event output information in the customer service event information DB (specifically, a preset ID (identifying information of a predetermined position: applicable to 1 to PN) and a customer service ID (identifying information of a clerk: 1 to EN) Stuff) and customer service event ID) are stored (held) in the customer service situation DB (see FIG. 10) (S14).
 図4に示す接客モニタリングシステム100Bでは、マイク装置M1・・・MLの指向性が製造時から予め決まっているものであり、その指向性を変更することができない。すなわち、図4に示す接客モニタリングシステム100Bではマイクアレイ装置AM1,…,AMLにより音声が収音されていないため(S15、NO)、指向性の形成処理が不可であり、接客イベント検出部331は、モニタリングデータ4bから、検出された接客イベントに対応する店員の音声データを取得し(S16)、接客発話評価値算定部332に音声データと接客イベントIDとを入力する(S18)。検出された接客イベントに対応する店員の音声データとは、既定位置(プリセット位置)と対応付けられたマイク装置により収音された店員の音声データである。 In the customer service monitoring system 100B shown in FIG. 4, the directivity of the microphone devices M1... ML is predetermined from the time of manufacture, and the directivity cannot be changed. That is, in the customer service monitoring system 100B shown in FIG. 4, since the voice is not collected by the microphone array devices AM1,..., AML (S15, NO), the directivity forming process is impossible, and the customer service event detecting unit 331 is Then, voice data of the store clerk corresponding to the detected customer service event is acquired from the monitoring data 4b (S16), and the voice data and customer service event ID are input to the customer service utterance evaluation value calculation unit 332 (S18). The clerk's voice data corresponding to the detected customer service event is the clerk's voice data collected by the microphone device associated with the predetermined position (preset position).
 接客発話評価値算定部332は、図15に示す接客発話評価処理を実行し(S19)、接客発話評価出力値を接客状況DBに格納(保持)する(S20)。これにより、図9に示す接客イベント情報処理が終了する。図8において、ステップS8-4の後、接客イベント情報DBの全てのレコードが取得されていない場合には(S8-5、NO)、接客イベント検出部331の処理はステップS8-3に戻る。一方、接客イベント情報DBの全てのレコードが取得された場合には(S8-5、YES)、接客イベント検出部331の処理は終了する。 The customer service utterance evaluation value calculation unit 332 executes the customer service utterance evaluation process shown in FIG. 15 (S19), and stores (holds) the service utterance evaluation output value in the customer service status DB (S20). Thereby, the service event information processing shown in FIG. 9 ends. In FIG. 8, after step S8-4, if not all records in the customer service event information DB have been acquired (S8-5, NO), the process of the customer service event detector 331 returns to step S8-3. On the other hand, when all the records of the customer service event information DB have been acquired (S8-5, YES), the process of the customer service event detector 331 ends.
 (図5に示す接客モニタリングシステム100Cにおける接客イベント検出処理)
 ステップS8-2の後、接客イベント検出部331は、未取得である接客イベント情報DBの1行目のレコード(接客イベントID「EID1」、接客イベント名「入退店挨拶」)を取得し(S8-3)、接客イベント情報処理を開始する。接客イベント検出部331は、モニタリングデータ4bから照合対象データとしての映像データとセンサデータに含まれる検出結果(自動ドア開閉履歴データ)とを取得し(S11)、自動ドア開閉履歴データに自動ドアの開閉動作があるか否かを照合する(S12)。
(Service Event Detection Processing in the Service Monitoring System 100C shown in FIG. 5)
After step S8-2, the customer service event detection unit 331 acquires a record (service event ID “EID1”, customer service event name “entrance / exit greeting”) in the first line of the service event information DB that has not been acquired ( S8-3), customer service event information processing is started. The customer service event detection unit 331 acquires video data as verification target data and detection results (automatic door opening / closing history data) included in the sensor data from the monitoring data 4b (S11), and the automatic door opening / closing history data includes the automatic door opening / closing history data. It is checked whether or not there is an opening / closing operation (S12).
 自動ドア開閉履歴データが自動ドアの開閉動作を含まない場合には(S13、NO)、図9に示す接客イベント情報処理は終了し、接客イベント検出部331の処理はステップS8-5に戻る。 If the automatic door opening / closing history data does not include an automatic door opening / closing operation (S13, NO), the customer service event information processing shown in FIG. 9 ends, and the process of the customer service event detection unit 331 returns to step S8-5.
 一方、自動ドア開閉履歴データが自動ドアの開閉動作を含む場合には(S13、YES)、接客イベント検出部331は、図13に示す接客イベント情報DBの接客イベント出力情報(具体的には、各店員の接客者位置座標、カメラID(カメラ装置の識別情報:1~CNのうち該当するもの)、接客者ID(店員の識別情報:1~ENのうち該当するもの)、接客イベントID)を接客状況DB(図10参照)に格納(保持)する(S14)。 On the other hand, when the automatic door opening / closing history data includes an automatic door opening / closing operation (S13, YES), the customer service event detection unit 331 receives the service event output information (specifically, the customer service event information DB shown in FIG. 13). Each store employee's customer service location coordinates, camera ID (camera device identification information: 1 to CN), customer service ID (store employee identification: 1 to EN), customer service event ID) Is stored (held) in the customer service situation DB (see FIG. 10) (S14).
 接客者位置座標は、接客イベント検出部331が所定の映像データを画像処理することで得られ、表示装置35の画面上に表示された映像データにおいて各店員が存在する位置の座標として出力される。また、カメラIDは、接客イベント検出部331が所定の映像データを画像処理することで、自動ドアの開閉動作時に各店員が存在した位置を最も近くで撮像するカメラ装置の識別情報として出力される。マイクIDは、カメラIDと予め対応付けられているので、カメラIDが選択された時点で選択されることになって出力される。 The customer service position coordinates are obtained by image processing of predetermined video data by the customer service event detection unit 331, and are output as the coordinates of the position where each store clerk exists in the video data displayed on the screen of the display device 35. . The camera ID is output as identification information of a camera device that images the position where each store clerk was closest when the automatic door opening / closing operation is performed by the customer service event detection unit 331 performing image processing on predetermined video data. . Since the microphone ID is associated with the camera ID in advance, the microphone ID is selected and output when the camera ID is selected.
 図5に示す接客モニタリングシステム100Cではマイクアレイ装置AM1,…,AMLにより収音され(S15、YES)、その収音により得られた音声データによる指向性の形成処理が可能のため、接客イベント検出部331は、検出された接客イベントに対応する各店員の接客者位置座標のデータと、モニタリングデータ4bに含まれる接客イベントに対応する各店員の音声データとを指向性制御部37に入力し、指向性制御部37において各店員の音声データに対して各店員に最も近いマイクアレイ装置から各店員に向かう方向に指向性が形成された後の音声データを取得する(S17)。接客イベント検出部331は、接客発話評価値算定部332に、各店員の接客者位置座標(例えば表示装置35の画面上に表示された店員の存在する位置の座標)のデータとステップS17において取得した音声データと接客イベントIDとを入力する(S18)。 In the customer service monitoring system 100C shown in FIG. 5, sound is collected by the microphone array devices AM1,..., AML (S15, YES). The unit 331 inputs, to the directivity control unit 37, the data on the customer position coordinates of each store clerk corresponding to the detected customer service event and the voice data of each store clerk corresponding to the customer service event included in the monitoring data 4b. The directivity control unit 37 acquires voice data after directivity is formed in the direction from the microphone array device closest to each store clerk toward each store clerk with respect to the voice data of each store clerk (S17). The customer service event detection unit 331 obtains in the service customer utterance evaluation value calculation unit 332 the data on the customer service staff position coordinates (for example, the coordinates of the position of the sales staff displayed on the screen of the display device 35) and the data in step S17. The received voice data and customer service event ID are input (S18).
 接客発話評価値算定部332は、図15に示す接客発話評価処理を実行し(S19)、接客発話評価出力値を接客状況DBに格納(保持)する(S20)。これにより、図9に示す接客イベント情報処理が終了する。図8において、ステップS8-4の後、接客イベント情報DBの全てのレコードが取得されていない場合には(S8-5、NO)、接客イベント検出部331の処理はステップS8-3に戻る。 The customer service utterance evaluation value calculation unit 332 executes the customer service utterance evaluation process shown in FIG. 15 (S19), and stores (holds) the service utterance evaluation output value in the customer service status DB (S20). Thereby, the service event information processing shown in FIG. 9 ends. In FIG. 8, after step S8-4, if not all records in the customer service event information DB have been acquired (S8-5, NO), the process of the customer service event detector 331 returns to step S8-3.
 次に、図8に示す接客イベント検出処理の結果として出力される接客イベント出力情報(図11から図14参照)が含まれる接客状況DBの一例について、図10を参照して説明する。図10は、接客状況DBの一例を示す図である。 Next, an example of a customer service situation DB including customer service event output information (see FIGS. 11 to 14) output as a result of the customer service event detection process shown in FIG. 8 will be described with reference to FIG. FIG. 10 is a diagram illustrating an example of the customer service situation DB.
 図10に示す接客状況DBは、接客状況データIDと、接客発話評価値と、イベント開始時刻と、イベント終了時刻と、接客者IDと、接客イベントIDと、接客者位置(プリセット)と、接客者位置(プリセット外)との各項目に対応するデータが規定される。 The customer service status DB shown in FIG. 10 includes a customer service status data ID, a customer service utterance evaluation value, an event start time, an event end time, a customer service ID, a customer service event ID, a customer service location (preset), and a customer service. Data corresponding to each item of the person position (outside the preset) is defined.
 接客状況データID「ID1」では、接客発話評価値はV11…V1nであり、接客者位置は接客者(店員)が既定位置(プリセット位置)ではない位置に存在することが検出されたためにその店員を撮像したカメラ装置のカメラIDと、表示装置35の画面上に表示された映像データ上の店員の位置を示す座標(画面上座標位置)とにより構成される。なお、カメラID「C1」のカメラ装置は全方位カメラ装置でもよいし、固定の画角を有するカメラ装置、又はパンチルトズーム機能を有するPTZカメラ装置でもよい。 In the customer service status data ID “ID1”, the customer service utterance evaluation value is V11... V1n, and the customer service position is detected because the customer service (store clerk) is not at the default position (preset position). And the coordinates indicating the position of the sales clerk on the video data displayed on the screen of the display device 35 (coordinate position on the screen). The camera device with the camera ID “C1” may be an omnidirectional camera device, a camera device having a fixed angle of view, or a PTZ camera device having a pan / tilt / zoom function.
 また、接客発話評価値V11の左側の添え字である「1」は接客イベントID「EID1」に対応しており、接客発話評価値V11の右側の添え字である「1」は所定時間間隔毎に切り取られるモニタリングデータ4a,4bの中で同一の接客イベントIDの接客イベントが検出された場合の接客イベントの識別情報を示す。nは1以上の整数であり、例えば10秒程度のモニタリングデータ4a,4bの中で同一の接客イベントIDの接客イベントが複数検出された場合には、nは2以上の整数となる。 Further, “1” as the subscript on the left side of the customer service utterance evaluation value V11 corresponds to the customer service event ID “EID1”, and “1” as the subscript on the right side of the customer service utterance evaluation value V11 is set at every predetermined time interval. The customer service event identification information when a customer service event with the same customer service event ID is detected in the monitoring data 4a and 4b cut out at the same time. n is an integer of 1 or more. For example, when a plurality of customer service events having the same customer service ID are detected in the monitoring data 4a and 4b of about 10 seconds, n is an integer of 2 or more.
 接客状況データID「ID2」では、接客発話評価値はV21…V2mであり、上述したnと同様にmは1以上の整数である。接客者位置は接客者(店員)が既定位置(プリセット位置)に存在することが検出されたためにその既定位置(プリセット位置)を示すプリセットIDにより構成される。 In the customer service status data ID “ID2”, the customer service utterance evaluation value is V21... V2m, and m is an integer of 1 or more as in the case of n described above. The customer service position is configured by a preset ID indicating the default position (preset position) since it is detected that the customer (store clerk) is present at the default position (preset position).
 従って、図2から図5に示す接客モニタリングシステム100,100A,100B,100Cは、このように図10に示す接客状況DBを有することで、該当する接客イベントの開始時刻から終了時刻までの音声データ及び映像データを接客評価装置3において出力(再生)することができ、店舗の責任者(例えば店長)が接客イベント時の店員の接客状況を音声及び映像で確認しながら注意深く観察、レビューすることができる。なお、音声データはレコーダ装置4に格納(記憶)されているので、接客評価装置3は、接客イベントIDに対応する接客イベントの検出時に収音された音声データをレコーダ装置4から取得した上で出力(再生)する。 Accordingly, the customer service monitoring systems 100, 100A, 100B, and 100C shown in FIGS. 2 to 5 have the customer service situation DB shown in FIG. 10 as described above, so that the voice data from the start time to the end time of the corresponding customer service event. And video data can be output (reproduced) by the customer service evaluation device 3, and a store manager (for example, a store manager) can carefully observe and review the customer service status of the store clerk at the customer service event with sound and video. it can. Since the voice data is stored (stored) in the recorder device 4, the customer service evaluation device 3 obtains the voice data collected when the customer service event corresponding to the customer service event ID is detected from the recorder device 4. Output (play).
 次に、図9に示す接客発話評価処理(ステップS19参照)の動作手順の詳細について、図15を参照して説明する。図15は、接客発話評価処理の動作手順の一例を説明するフローチャートである。 Next, details of the operation procedure of the customer service utterance evaluation process (see step S19) shown in FIG. 9 will be described with reference to FIG. FIG. 15 is a flowchart for explaining an example of the operation procedure of the customer service utterance evaluation process.
 図15において、接客発話評価値算定部332は、ステップS18において接客イベント検出部331から渡された音声データと接客イベントIDとを取得し(S21)、騒音レベル判定処理を行い(S22)、接客キーワードの発話判定処理を行う(S23)。ステップS23の後、接客発話評価値算定部332は、検出状態(後述参照)のフラグが「1」であるか否かを判定する(S24)。検出状態のフラグが「1」である場合には(S24、YES)、接客発話評価値算定部332は、スコアリング処理を行う(S25)。一方、検出状態のフラグが「1」ではない場合には(S24、NO)、接客発話評価値算定部332は、接客発話評価値をゼロ点に設定するか所定点数を減点する(S26)。 In FIG. 15, the customer service utterance evaluation value calculation unit 332 acquires the voice data and customer service event ID passed from the customer service event detection unit 331 in step S18 (S21), performs a noise level determination process (S22), and receives the customer service. A keyword utterance determination process is performed (S23). After step S23, the customer service utterance evaluation value calculation unit 332 determines whether or not the flag of the detection state (see later) is “1” (S24). When the flag of the detection state is “1” (S24, YES), the customer service utterance evaluation value calculation unit 332 performs a scoring process (S25). On the other hand, when the flag of the detection state is not “1” (S24, NO), the service utterance evaluation value calculation unit 332 sets the service utterance evaluation value to a zero point or deducts a predetermined score (S26).
 ステップS25又はステップS26の後、接客発話評価値算定部332は、スコアリングデータとして、検出したキーワードID(後述参照)と接客発話評価値とを接客発話評価部33に出力する(S27)。 After step S25 or step S26, the service utterance evaluation value calculation unit 332 outputs the detected keyword ID (described later) and the service utterance evaluation value as scoring data to the service utterance evaluation unit 33 (S27).
 次に、図15に示す騒音レベル判定処理(ステップS22参照)の動作手順の詳細について、図16を参照して説明する。図16は、騒音レベル判定処理の動作手順の一例を説明するフローチャートである。 Next, details of the operation procedure of the noise level determination process (see step S22) shown in FIG. 15 will be described with reference to FIG. FIG. 16 is a flowchart for explaining an example of the operation procedure of the noise level determination process.
 図16において、接客発話評価値算定部332は、接客評価装置3が取得した収音領域(例えば店舗)の周囲の騒音レベルが所定値x[dB]以下であるか否かを判定する(S22-1)。騒音レベルは、例えばマイク装置、接客者マイク装置又はマイクアレイ装置のいずれかにより収音されて接客評価装置3に送信される。接客発話評価値算定部332は、騒音レベルが所定値x[dB]以下であると判定した場合には(S22-1、YES)、発話判定閾値(後述参照)をα1に決定する(S22-2)。 In FIG. 16, the customer service utterance evaluation value calculation unit 332 determines whether or not the noise level around the sound collection area (for example, a store) acquired by the customer service evaluation device 3 is equal to or less than a predetermined value x [dB] (S22). -1). The noise level is collected by, for example, any one of a microphone device, a customer service microphone device, and a microphone array device, and is transmitted to the customer service evaluation device 3. When the customer service utterance evaluation value calculation unit 332 determines that the noise level is equal to or less than the predetermined value x [dB] (S22-1, YES), the utterance determination threshold (see later) is determined to be α1 (S22-). 2).
 一方、接客発話評価値算定部332は、騒音レベルが所定値x[dB]を超える場合には(S22-1、NO)、騒音レベルが所定値y(>x)[dB]以下であるか否かを判定する(S22-3)。接客発話評価値算定部332は、騒音レベルが所定値y[dB]以下であると判定した場合には(S22-3、YES)、発話判定閾値(後述参照)をα2に決定する(S22-4)。一方、接客発話評価値算定部332は、騒音レベルが所定値y[dB]を超えると判定した場合には(S22-3、NO)、発話判定閾値(後述参照)をα3に決定する(S22-5)。 On the other hand, if the noise level exceeds the predetermined value x [dB] (S22-1: NO), the customer service utterance evaluation value calculation unit 332 determines whether the noise level is equal to or lower than the predetermined value y (> x) [dB]. It is determined whether or not (S22-3). If the customer service utterance evaluation value calculation unit 332 determines that the noise level is equal to or lower than the predetermined value y [dB] (S22-3, YES), it determines the utterance determination threshold (see later) to α2 (S22-). 4). On the other hand, when it is determined that the noise level exceeds the predetermined value y [dB] (S22-3, NO), the customer service utterance evaluation value calculation unit 332 determines α3 as the utterance determination threshold (see below) (S22). -5).
 次に、図15に示す接客キーワードの発話判定処理(ステップS23参照)の動作手順の詳細について、図17を参照して説明する。図17は、接客キーワードの発話判定処理の動作手順の一例を説明するフローチャートである。 Next, details of the operation procedure of the service keyword utterance determination process (see step S23) shown in FIG. 15 will be described with reference to FIG. FIG. 17 is a flowchart for explaining an example of the operation procedure of the service keyword utterance determination process.
 図17において、接客発話評価値算定部332は、検出状態のフラグを「0」に設定する(S23-1)。検出状態のフラグとは、接客イベントにおいて店員が発話する可能性が高い又は発話するべき模範的な発話想定キーワード(図22A参照)が発話された状態を示す情報である。 In FIG. 17, the customer service utterance evaluation value calculation unit 332 sets the detection state flag to “0” (S23-1). The flag of the detection state is information indicating a state in which an exemplary utterance assumption keyword (see FIG. 22A) that is likely to be uttered by a store clerk in a customer service event or to be uttered is uttered.
 接客発話評価値算定部332は、ステップS21において取得した音声データを接客発話評価値算定部332の音声認識エンジンに入力し(S23-2)、更に、接客イベントIDに対応する全ての発話想定キーワードとそれぞれの発話想定キーワードを識別するキーワードIDとの組を管理サーバ2の管理DB2aの接客発話評価DBから取得する(S23-3)。 The service utterance evaluation value calculation unit 332 inputs the voice data acquired in step S21 to the speech recognition engine of the service utterance evaluation value calculation unit 332 (S23-2), and further, all utterance assumption keywords corresponding to the service event ID And a keyword ID for identifying each utterance assumed keyword is acquired from the customer service utterance evaluation DB of the management DB 2a of the management server 2 (S23-3).
 接客発話評価値算定部332は、音声認識エンジンによる音声認識結果に、ステップS23-3で取得された発話想定キーワードが含まれるか否かを判定する(S23-4)。音声認識エンジンによる音声認識結果に発話想定キーワードが含まれないと判定された場合には(S23-4、NO)、図17に示す接客発話評価値算定部332の処理は終了する。 The customer service utterance evaluation value calculation unit 332 determines whether the speech utterance keyword acquired in step S23-3 is included in the speech recognition result by the speech recognition engine (S23-4). If it is determined that the speech recognition keyword is not included in the speech recognition result by the speech recognition engine (S23-4, NO), the processing of the customer utterance evaluation value calculation unit 332 shown in FIG. 17 ends.
 一方、接客発話評価値算定部332は、音声認識エンジンによる音声認識結果に、ステップS23-3で取得された発話想定キーワードが含まれると判定した場合には(S23-4、YES)、音声認識処理結果の評価値がステップS22-2、ステップS22-4又はステップS22-5において決定された発話判定閾値(α1,α2,α3のうちいずれか)以上であるか否かを判定する(S23-5)。音声認識処理結果の評価値が発話判定閾値(α1,α2,α3のうちいずれか)未満であると判定された場合には(S23-5、NO)、図17に示す接客発話評価値算定部332の処理は終了する。 On the other hand, if the customer service utterance evaluation value calculation unit 332 determines that the speech recognition result obtained by the speech recognition engine includes the speech utterance keyword acquired in step S23-3 (S23-4, YES), the speech recognition It is determined whether or not the evaluation value of the processing result is greater than or equal to the utterance determination threshold (any one of α1, α2, and α3) determined in step S22-2, step S22-4, or step S22-5 (S23-). 5). When it is determined that the evaluation value of the speech recognition processing result is less than the utterance determination threshold value (any of α1, α2, and α3) (S23-5, NO), the customer utterance evaluation value calculation unit shown in FIG. The processing at 332 ends.
 一方、接客発話評価値算定部332は、音声認識処理結果の評価値が発話判定閾値(α1,α2,α3のうちいずれか)以上であると判定した場合には(S23-5、YES)、検出状態のフラグを「1」に設定変更し(S23-6)、ステップS21において取得した音声データを、発話想定キーワードに対応するキーワードの発話部分のみに切り取って更新して上書き保存する(S23-7)。発話部分の前後に余計なノイズ音が入っている場合であっても、キーワードの発話部分のみを切り取ることにより、その前後に入っているノイズ音はカットされるので、音声認識の精度が向上し、後続のステップS25におけるスコアリング処理の精度も確保される。 On the other hand, when the customer service utterance evaluation value calculation unit 332 determines that the evaluation value of the speech recognition processing result is equal to or greater than the utterance determination threshold (any one of α1, α2, and α3) (S23-5, YES), The detection state flag is changed to “1” (S23-6), and the voice data acquired in step S21 is cut only to the utterance portion of the keyword corresponding to the utterance assumed keyword, and is updated and saved (S23-). 7). Even if there are extra noises before and after the utterance part, by cutting out only the utterance part of the keyword, the noise sound before and after it will be cut, which improves the accuracy of voice recognition. The accuracy of the scoring process in the subsequent step S25 is also ensured.
 次に、図15に示すスコアリング処理(ステップS25参照)の動作手順の詳細について、図18を参照して説明する。図18は、スコアリング処理の動作手順の一例を説明するフローチャートである。 Next, details of the operation procedure of the scoring process (see step S25) shown in FIG. 15 will be described with reference to FIG. FIG. 18 is a flowchart for explaining an example of the operation procedure of the scoring process.
 図18において、接客発話評価値算定部332は、ステップS23-7において更新された音声データを用いて、発話長さ判定処理を行い(S25-1)、更に、周波数特性判定処理を行う(S25-2)。更に、接客発話評価値算定部332は、発話長さ判定処理及び周波数特性判定処理の結果としてのスコアリングデータ(具体的には、ステップS23-7において更新された音声データから検出されたキーワードと同じ発話想定キーワードを識別するキーワードIDと接客発話評価値との組)をメモリ32に保持する(S25-3)。 In FIG. 18, the customer service utterance evaluation value calculation unit 332 performs utterance length determination processing using the voice data updated in step S23-7 (S25-1), and further performs frequency characteristic determination processing (S25). -2). Further, the customer service utterance evaluation value calculation unit 332 obtains scoring data (specifically, the keyword detected from the voice data updated in step S23-7) as a result of the utterance length determination process and the frequency characteristic determination process. A set of a keyword ID for identifying the same utterance assumed keyword and a customer service utterance evaluation value) is stored in the memory 32 (S25-3).
 次に、図18に示す発話長さ判定処理(ステップS25-1参照)及び周波数特性判定処理(ステップS25-2参照)の動作手順の詳細について、図19A及び図19Bを参照して説明する。図19Aは、発話長さの判定処理の動作手順の一例を説明するフローチャートである。図19Bは、周波数特性の判定処理の動作手順の一例を説明するフローチャートである。 Next, details of the operation procedure of the speech length determination process (see step S25-1) and the frequency characteristic determination process (see step S25-2) shown in FIG. 18 will be described with reference to FIGS. 19A and 19B. FIG. 19A is a flowchart for explaining an example of an operation procedure of speech length determination processing. FIG. 19B is a flowchart illustrating an example of an operation procedure of a frequency characteristic determination process.
 図19Aにおいて、接客発話評価値算定部332は、管理サーバ2の管理DB2aの接客発話評価DBを参照し、ステップS21において取得した接客イベントIDに対応する接客発話モデルIDにより特定されるモデル音声データを管理サーバ2の管理DB2aから取得する(S31)。モデル音声データは、所定の接客イベント毎の発話想定キーワードの音声データを含むキーワード音声データの一例である。接客発話評価値算定部332は、ステップS23-7において更新された音声データ(例えば店員の発話部分)の長さが模範的な所定範囲内であるか否かを判定する(S32)。 In FIG. 19A, the service utterance evaluation value calculation unit 332 refers to the service utterance evaluation DB of the management DB 2a of the management server 2, and model voice data specified by the service utterance model ID corresponding to the service event ID acquired in step S21. Is acquired from the management DB 2a of the management server 2 (S31). The model voice data is an example of keyword voice data including voice data of an utterance assumed keyword for each predetermined customer service event. The customer service utterance evaluation value calculation unit 332 determines whether or not the length of the voice data (for example, the clerk's utterance portion) updated in step S23-7 is within a typical predetermined range (S32).
 図20は、モデル音声データを用いた発話長さの判定処理の具体例を示す図である。図20では、横軸は時間を示し、例えば「入店挨拶」の接客イベントにおいて発話される「いらっしゃいませ」の模範的な所定範囲の音声の発話長さl0の「いらっしゃいませ」と、所定範囲を超える音声の発話長さl1の「いらっしゃいませ」(図20に示すNo.1参照)及び発話長さl2の「いらっしゃいませ」(図20に示すNo.2参照)とが示されている。 FIG. 20 is a diagram showing a specific example of the utterance length determination process using the model voice data. In FIG. 20, the horizontal axis indicates time. For example, “Welcome” with an utterance length l0 of an exemplary predetermined range of “Welcome” uttered in a customer service event of “greeting in store”, and a predetermined range “I welcome you” (refer to No. 1 shown in FIG. 20) and “I welcome you” (refer to No. 2 shown in FIG. 20) having an utterance length l2.
 接客発話評価値算定部332は、ステップS23-7において更新された音声データの発話長さがモデル音声データの発話長さ(発話長さl0)より所定範囲(例えば10%)内を超える場合(例えば図20に示すNo.1及びNo.2参照)には(S32、NO)、接客発話評価値から所定点数を減点する(S34)。 The service utterance evaluation value calculation unit 332 determines that the utterance length of the voice data updated in step S23-7 exceeds a predetermined range (for example, 10%) from the utterance length of the model voice data (utterance length l0) ( For example, in No. 1 and No. 2 shown in FIG. 20 (S32, NO), a predetermined score is subtracted from the customer service utterance evaluation value (S34).
 例えば図20に示すNo.1のケースでは、発話された「いらっしゃいませ」の発話長さがモデル音声データの「いらっしゃいませ」の発話長さより所定範囲ほど短く、この場合には、接客発話評価値算定部332は、所定点数として、「100×(0.9l0-l1)/l0」を減点する。l1は、図20に示すNo.1のケースで発話された「いらっしゃいませ」の発話長さを示す。より具体的には、モデル音声データの「いらっしゃいませ」の発話長さが1秒で、所定範囲がモデル音声データの「いらっしゃいませ」の発話長さの±10%であると、図20に示すNo.1のケースで発話された「いらっしゃいませ」の発話長さが0.9秒~1.1秒の間であれば減点はされないが、例えば0.7秒であった場合には、20点(=100×(0.9×1秒-0.7秒))が減点される。 For example, No. shown in FIG. In the first case, the utterance length of the uttered “Ira-saisei” is shorter than the utterance length of “Ira-saisei” in the model voice data by a predetermined range. In this case, the service utterance evaluation value calculation unit 332 has a predetermined number of points. As a result, “100 × (0.910−11) / 10” is deducted. 11 corresponds to No. 11 shown in FIG. The utterance length of “I welcome you” uttered in case 1 is shown. More specifically, FIG. 20 shows that the utterance length of the model voice data “Irasementation” is 1 second and the predetermined range is ± 10% of the utterance length of the model voice data “Irasementation”. No. If the utterance length of “I welcome you” uttered in case 1 is between 0.9 seconds and 1.1 seconds, no deduction will be given, but if it is 0.7 seconds, for example, 20 points ( = 100 x (0.9 x 1 sec-0.7 sec)) is deducted.
 また、例えば図20に示すNo.2のケースでは、発話された「いらっしゃいませ」の発話長さがモデル音声データの「いらっしゃいませ」の発話長さより所定範囲ほど長く、この場合には、接客発話評価値算定部332は、所定点数として、「100×(l2-1.1l0)/l0」を減点する。l2は、図20に示すNo.2のケースで発話された「いらっしゃいませ」の発話長さを示す。より具体的には、モデル音声データの「いらっしゃいませ」の発話長さが1秒で、所定範囲がモデル音声データの「いらっしゃいませ」の発話長さの±10%であると、図20に示すNo.2のケースで発話された「いらっしゃいませ」の発話長さが0.9秒~1.1秒の間であれば減点はされないが、例えば1.3秒であった場合には、20点(=100×(1.3秒-1.1×1秒))が減点される。 Also, for example, as shown in FIG. In the case of 2, the utterance length of the uttered “Islaimase” is longer than the utterance length of the “speaking” of the model voice data by a predetermined range. In this case, the customer utterance evaluation value calculation unit 332 As a result, “100 × (l2−1.1l0) / l0” is deducted. l2 is No. 12 shown in FIG. The utterance length of “I welcome you” uttered in case 2 is shown. More specifically, FIG. 20 shows that the utterance length of the model voice data “Irasementation” is 1 second and the predetermined range is ± 10% of the utterance length of the model voice data “Irasementation”. No. If the utterance length of “I welcome you” uttered in case 2 is between 0.9 seconds and 1.1 seconds, no deduction will be given, but if it is 1.3 seconds, for example, 20 points ( = 100 × (1.3 seconds−1.1 × 1 seconds)) is deducted.
 一方、接客発話評価値算定部332は、ステップS23-7において更新された音声データの発話長さがモデル音声データの発話長さ(発話長さl0)より所定範囲(例えば10%)内を超えない場合(S32、YES)又はステップS34の後、スコアリングデータ(具体的には、ステップS23-7において更新された音声データの中で検出したキーワードを識別するキーワードIDとステップS34で減点された後若しくはステップS34で減点されなかった初期値(例えば100点)である接客発話評価値との組)をメモリ32に保持する(S33)。 On the other hand, the service utterance evaluation value calculation unit 332 determines that the utterance length of the voice data updated in step S23-7 exceeds a predetermined range (for example, 10%) from the utterance length of the model voice data (utterance length l0). If not (S32, YES) or after step S34, the scoring data (specifically, the keyword ID for identifying the keyword detected in the voice data updated in step S23-7 and the points deducted in step S34) The initial value (for example, a set with the customer utterance evaluation value that is 100 points) that was not deducted after or in step S34 is held in the memory 32 (S33).
 図19Bにおいて、接客発話評価値算定部332は、管理サーバ2の管理DB2aの接客発話評価DBを参照し、ステップS21において取得した接客イベントIDに対応する接客発話モデルIDにより特定されるモデル音声データを管理サーバ2の管理DB2aから取得する(S41)。接客発話評価値算定部332は、ステップS23-7において更新された音声データの音素(一つ一つの言葉の音)毎の周波数特性(例えば周波数)が模範的な音素毎の基本周波数から所定範囲内であるか否かを判定する(S42)。 In FIG. 19B, the service utterance evaluation value calculation unit 332 refers to the service utterance evaluation DB of the management DB 2a of the management server 2, and model voice data specified by the service utterance model ID corresponding to the service event ID acquired in step S21. Is acquired from the management DB 2a of the management server 2 (S41). The service utterance evaluation value calculation unit 332 has a frequency characteristic (for example, frequency) for each phoneme (sound of each word) of the voice data updated in step S23-7 within a predetermined range from a basic frequency for each phoneme. It is determined whether it is within (S42).
 図21は、モデル音声データの各音素の基本周波数を用いた周波数特性の判定処理の具体例を示す図である。図21では、横軸は時間を示し、点線の丸はモデル音声データの各音素の基本周波数f1~f7を示し、実線の丸はステップS23-7において更新された音声データの各音素の周波数f’1~f’7を示し、例えば「入店挨拶」の接客イベントにおいて発話される「いらっしゃいませ」の模範的な音素の基本周波数f1~f7毎の所定範囲(図21に示す直線型の各実線矢印参照)とが示されている。 FIG. 21 is a diagram showing a specific example of frequency characteristic determination processing using the fundamental frequency of each phoneme of the model voice data. In FIG. 21, the horizontal axis indicates time, the dotted circle indicates the fundamental frequencies f1 to f7 of each phoneme of the model speech data, and the solid circle indicates the frequency f of each phoneme of the speech data updated in step S23-7. '1 to f'7, for example, a predetermined range for each of the fundamental frequencies f1 to f7 of the phoneme of “I welcome you” uttered at the customer service event “greeting at the store” (each of the linear types shown in FIG. 21) (See solid arrows).
 接客発話評価値算定部332は、ステップS23-7において更新された音声データの音素毎に、各音素の周波数特性(例えば周波数)がモデル音声データの各音素の周波数特性毎の所定範囲(例えば60[Hz])内を超えると判定した場合には(S42、NO)、所定範囲を超えた音素の数に応じて、接客発話評価値から所定点数を減点する(S44)。 For each phoneme of the voice data updated in step S23-7, the service level evaluation value calculation unit 332 sets the frequency characteristics (for example, frequency) of each phoneme to a predetermined range (for example, 60) for each frequency characteristic of each phoneme of the model voice data. [Hz]) If determined to exceed (NO in S42), the predetermined score is subtracted from the customer utterance evaluation value according to the number of phonemes exceeding the predetermined range (S44).
 例えば図21に示すケースでは、音素「ら」と音素「ま」の周波数f’1,f’6が対応する各基本周波数f1,f6毎の所定範囲を超えているので、接客発話評価値算定部332は、周波数差分(例えば|f1-f’1|)が60[Hz]から120[Hz]であれば5点を該当する音素毎に減点し、周波数差分(例えば|f1-f’1|)が120[Hz]を超えるなら10点を該当する音素毎に減点する。なお、地域や業界によっては発話想定キーワードの語尾が上げ調子で発話されることが好ましい場合もあるので(図21に示す一点鎖線参照)、例えば語尾又は語尾を含む所定数個の音素に対しては、基本周波数の値が増加されて使用されてもよい。 For example, in the case shown in FIG. 21, since the frequencies f′1 and f′6 of the phoneme “ra” and the phoneme “ma” exceed the predetermined ranges for the corresponding fundamental frequencies f1 and f6, the customer utterance evaluation value calculation is performed. If the frequency difference (for example, | f1-f′1 |) is 60 [Hz] to 120 [Hz], the unit 332 deducts five points for each corresponding phoneme, and the frequency difference (for example, | f1-f′1). If |) exceeds 120 [Hz], 10 points are deducted for each corresponding phoneme. Depending on the region and the industry, it may be preferable that the utterance of the utterance assumed keyword is uttered in an upward tone (refer to the alternate long and short dash line shown in FIG. 21). For example, for a predetermined number of phonemes including the ending or the ending May be used by increasing the value of the fundamental frequency.
 一方、接客発話評価値算定部332は、ステップS23-7において更新された音声データの音素毎に、各音素の周波数特性(例えば周波数)がモデル音声データの各音素の周波数特性毎の所定範囲(例えば60[Hz])内を超えないと判定した場合には(S42、YES)又はステップS44の後、スコアリングデータ(具体的には、ステップS23-7において更新された音声データの中で検出したキーワードを識別するキーワードIDとステップS44で減点された後若しくはステップS44で減点されなかった初期値(例えば100点)である接客発話評価値との組)をメモリ32に保持する(S43)。 On the other hand, the service utterance evaluation value calculation unit 332 sets the frequency characteristics (for example, frequency) of each phoneme to a predetermined range (for each frequency characteristic of each phoneme of the model voice data) for each phoneme of the voice data updated in step S23-7. For example, when it is determined that the frequency does not exceed 60 [Hz] (S42, YES) or after step S44, it is detected in the scoring data (specifically, the audio data updated in step S23-7) The memory ID is stored in the memory 32 with the keyword ID for identifying the selected keyword and the customer utterance evaluation value, which is the initial value (for example, 100 points) that has been deducted in step S44 or not deducted in step S44 (S43).
 図22Aは、接客発話評価DBの一部を構成する発話想定キーワード表の一例を示す図である。図22Bは、接客発話評価DBの一部を構成する接客発話モデルの一覧表の一例を示す図である。 FIG. 22A is a diagram showing an example of an utterance assumption keyword table constituting a part of the customer service utterance evaluation DB. FIG. 22B is a diagram illustrating an example of a list of service utterance models that constitute a part of the service utterance evaluation DB.
 図22Aに示す発話想定キーワード表は、接客イベントIDと、接客イベント名と、キーワードIDと、発話想定キーワードと、接客発話モデルIDとの各項目に対応するデータが規定される。キーワードIDは、発話想定キーワードを識別する。接客発話モデルIDは、図22Bに示すように、モデル音声データと対応付けられる。図22Aに示すように、1つの接客発話モデルIDに対応して1つ以上の発話想定キーワードが規定されてもよい(図22Aに示す接客イベントID「EID2」のレコード参照)。 In the utterance assumption keyword table shown in FIG. 22A, data corresponding to each item of the service event ID, the service event name, the keyword ID, the utterance assumption keyword, and the service utterance model ID are defined. The keyword ID identifies an utterance assumption keyword. The customer service utterance model ID is associated with model voice data as shown in FIG. 22B. As shown in FIG. 22A, one or more utterance assumption keywords may be defined corresponding to one customer utterance model ID (see the record of customer service event ID “EID2” shown in FIG. 22A).
 次に、本実施形態の接客モニタリングシステム100,100A,100B,100Cにおける接客状況DBの閲覧処理又は修正処理の動作手順について、図23を参照して説明する。図23は、限定された閲覧者による閲覧処理又は接客発話評価値の修正処理の動作手順の一例を示すフローチャートである。 Next, the operation procedure of the browsing process or the correction process of the customer service status DB in the customer service monitoring system 100, 100A, 100B, 100C of this embodiment will be described with reference to FIG. FIG. 23 is a flowchart illustrating an example of an operation procedure of a browsing process by a limited viewer or a service utterance evaluation value correction process.
 図23において、例えば表示装置35に表示された接客状況閲覧画面へのログイン画面WD1(図27参照)において、接客状況DBの閲覧要求者の入力操作(例えば指FGによるタッチ操作)により、ログインID及びパスワードが入力され、ログインボタンLGIが押下される(S51)。また、接客評価装置3は、管理サーバ2の管理DB2aの閲覧者DB(図26A参照)を参照し、閲覧要求者により入力された情報(具体的には図27のログイン画面WD1において入力されたパスワード)から、アクセス権、権限レベル、パスワードを取得する(S51)。図26Aは、閲覧者DBの一例を示す図である。図26Bは、接客者DBの一例を示す図である。 In FIG. 23, for example, on the login screen WD1 (see FIG. 27) for the customer service status browsing screen displayed on the display device 35, the login ID is input by the input operation (for example, touch operation with the finger FG) of the customer who requests the customer service status DB. And the password are input, and the login button LGI is pressed (S51). Further, the customer service evaluation device 3 refers to the viewer DB (see FIG. 26A) of the management DB 2a of the management server 2 and inputs information (specifically, entered on the login screen WD1 of FIG. 27) by the browse requester. The access right, authority level, and password are acquired from (password) (S51). FIG. 26A is a diagram illustrating an example of a viewer DB. FIG. 26B is a diagram illustrating an example of a customer service DB.
 図26Aに示す閲覧者DBは、閲覧者IDと、パスワードと、閲覧者権限と、権限レベルの各項目に対するデータの種類、種別がそれぞれ規定されている。閲覧者権限には、閲覧操作及び修正操作の両方が可能な権限と、閲覧操作のみ可能な権限との2種類が規定されている。パスワードは、例えば実際に入力されるパスワードでもよいし、入力されるパスワードのハッシュ値(ダイジェスト)でもよい。 In the viewer DB shown in FIG. 26A, the type and type of data for each item of the viewer ID, password, browser authority, and authority level are defined. There are two types of viewer authority defined: authority that allows both browsing and correction operations, and authority that allows only browsing operations. The password may be, for example, an actually input password or a hash value (digest) of the input password.
 図26Bに示す接客者DBは、店員の識別情報を示す接客者IDと、店舗の識別情報を示す店舗IDと、店員の氏名の各項目に対するデータの種類、種別がそれぞれ規定されている。 In the customer service DB shown in FIG. 26B, the data type and type for each item of the customer ID indicating the store clerk's identification information, the store ID indicating the store identification information, and the name of the store clerk are defined.
 図27は、接客モニタリングシステムにおける閲覧対象となる接客状況DBへのログイン画面の一例を示す図である。接客評価装置3は、ステップS51においてログイン画面に入力した閲覧要求者が閲覧者DBに規定される閲覧操作のアクセス権があって、かつパスワードが一致したか否かを判定する(S52)。閲覧操作のアクセス権が無い又はパスワードが一致しないと判定された場合には(S52、NO)、図23に示す接客評価装置3の処理は終了する。 FIG. 27 is a diagram showing an example of a login screen to the customer service status DB to be browsed in the customer service monitoring system. The customer service evaluation device 3 determines whether or not the browsing requester input on the login screen in step S51 has the access right of the browsing operation defined in the browser DB and the passwords match (S52). If it is determined that there is no access right for the browsing operation or that the passwords do not match (S52, NO), the processing of the customer service evaluation device 3 shown in FIG. 23 ends.
 一方、接客評価装置3は、閲覧者DBに規定される閲覧操作のアクセス権があって、かつパスワードが一致したと判定した場合には(S52、YES)、管理サーバ2の管理DB2aの接客状況DBにアクセスし、例えば1日単位の顧客(来店客)に対する全店員の接客発話評価値の集計結果を接客状況表示画面WD2(図28参照)として表示装置35に表示させる(S53)。 On the other hand, when the customer service evaluation device 3 determines that the user has access rights for the browsing operation defined in the viewer DB and the passwords match (S52, YES), the customer service status of the management DB 2a of the management server 2 is determined. The DB is accessed, and for example, the total result of the customer service utterance evaluation values of all the store clerks for the customer (customer) per day is displayed on the display device 35 as a customer service status display screen WD2 (see FIG. 28) (S53).
 図28は、1日単位の来店客に対する全接客者の接客発話評価値の集計結果の一例を接客状況表示画面WD2として示す図である。図28では、来店した顧客に対して検出された接客イベントの件数が255であり、そのうち入店挨拶の接客イベントの中で「いらっしゃいませ」の発話想定キーワードの発話が実行された件数が230で全件数255に比べた割合が90%であり、退店挨拶の接客イベントの中で「ありがとうございました」の発話想定キーワードの発話が実行された件数が195で全件数255に比べた割合が76%であることが示されている。 FIG. 28 is a diagram illustrating an example of a totaling result of customer service utterance evaluation values of all customer service customers per day as a customer service status display screen WD2. In FIG. 28, the number of customer service events detected for a customer who visited the store is 255, and among the customer service events of the store entrance greeting, the number of utterances of the utterance assumed keyword of “I welcome you” is 230. The ratio compared to the total number of 255 is 90%, and the number of utterances of the “Thank you” utterance assumed keyword was executed in the customer service event of closing greetings, and the ratio compared to the total number of 255 was 76 %.
 また、図28では、来店した顧客に対してレジ応対に関する接客イベントが検出された件数が180であり、そのうちポイントカードの確認漏れ(つまり、ポイントカードの提示を促すための接客イベントが検出されなかったこと)の件数が8で、全件数180に比べた割合が4%であり、店員のポイントカード確認によって顧客のポイントカードが提示された提示率が10%であることが示され、更に、温め確認漏れ(つまり、お弁当を電子レンジにおいて温めることを確認するための接客イベントが検出されなかったこと)の件数が3で、全件数180に比べた割合が2%であることが示されている。 Also, in FIG. 28, the number of customer service events related to cashier reception detected for customers who visited the store is 180, of which the point card confirmation omission (that is, the customer service event for prompting the presentation of the point card is not detected). 8), the ratio compared to the total number 180 is 4%, and the point card confirmation of the store clerk confirms that the presentation rate of the customer's point card is 10%. It is shown that the number of missed warming confirmations (that is, a customer service event for confirming that the lunch box is warmed in the microwave oven was not detected) was 3, and 2% of the total number was 180%. ing.
 なお、接客評価装置3の出力部34は、所定の入力操作に応じて、1日単位ではなく、1週間単位や1カ月単位に切り替えて図28に示す各項目のデータを集計して再表示してもよい。 It should be noted that the output unit 34 of the customer service evaluation apparatus 3 aggregates and redisplays the data of each item shown in FIG. 28 by switching to a week or month instead of a day according to a predetermined input operation. May be.
 図28に示す接客状況表示画面WD2において、ログアウトボタンLGOがユーザ(閲覧操作の権限レベルを有する者)の指FGによりタッチ操作で選択された場合には(S54、YES)、接客評価装置3の出力部34は、表示装置35において表示されている全ての閲覧画面を閉じる(クローズする)(S55)。一方、ログアウトボタンLGOが選択されない場合には(S54、NO)、図28に示す接客状況表示画面WD2の詳細表示ボタンIDTが選択され、かつ該当するアクセス権(修正操作が可能な権限レベルL1)があると判定された場合には(S56、YES)、接客評価装置3の出力部34は、図28に示す接客状況表示画面WD2から図32に示す詳細表示画面WD7に切り替えて表示装置35に表示させる(S57)。ステップS57の後、接客評価装置3において、接客発話評価値修正処理が行われる(S58)。図32は、接客状況DBの詳細表示画面WD7に表示された各レコードの具体例を示す図である。 In the customer service status display screen WD2 shown in FIG. 28, when the logout button LGO is selected by the touch operation with the finger FG of the user (who has the authority level of the browsing operation) (S54, YES), the customer service evaluation device 3 The output unit 34 closes (closes) all the browsing screens displayed on the display device 35 (S55). On the other hand, if the logout button LGO is not selected (S54, NO), the detail display button IDT on the customer service status display screen WD2 shown in FIG. 28 is selected, and the corresponding access right (authority level L1 that allows a correction operation). When it is determined that there is a service (S56, YES), the output unit 34 of the customer service evaluation device 3 switches from the customer service status display screen WD2 shown in FIG. 28 to the detailed display screen WD7 shown in FIG. It is displayed (S57). After step S57, the customer service evaluation device 3 performs a service utterance evaluation value correction process (S58). FIG. 32 is a diagram illustrating a specific example of each record displayed on the detail display screen WD7 of the customer service situation DB.
 一方、接客状況表示画面WD2の詳細表示ボタンIDTが選択されず、該当するアクセス権が無い、又はステップS58の後、図23に示す接客評価装置3の処理はステップS54に戻る。 On the other hand, the detail display button IDT on the customer service status display screen WD2 is not selected and there is no corresponding access right, or after step S58, the process of the customer service evaluation device 3 shown in FIG. 23 returns to step S54.
 次に、図23に示す接客発話評価値修正処理(ステップS58参照)の詳細について、図24及び図25を参照して説明する。図24は、接客発話評価値の修正処理の詳細な動作手順の一例を説明するフローチャートである。図25は、図24に示す接客発話評価値の修正処理の詳細な動作手順の続きを説明するフローチャートである。 Next, details of the customer service utterance evaluation value correction process (see step S58) shown in FIG. 23 will be described with reference to FIGS. FIG. 24 is a flowchart for explaining an example of a detailed operation procedure of the service utterance evaluation value correction process. FIG. 25 is a flowchart for explaining the continuation of the detailed operation procedure of the modification process of the customer service utterance evaluation value shown in FIG.
 図24において、図32に示す詳細表示画面WD7が表示装置35に表示された状態で、例えばユーザ(修正要求者)が修正したい接客状況データID(図33参照)のレコードRC1がユーザの指FGにより指定される(S58-1)。図33は、接客状況DBの詳細表示画面WD7に表示された特定のレコードRC1の接客発話評価値の修正操作の一例を示す図である。 In FIG. 24, in the state where the detailed display screen WD7 shown in FIG. 32 is displayed on the display device 35, for example, the record RC1 of the customer service situation data ID (see FIG. 33) that the user (correction requester) wants to correct is the user's finger FG. (S58-1). FIG. 33 is a diagram illustrating an example of a modification operation for the customer utterance evaluation value of the specific record RC1 displayed on the detail display screen WD7 of the customer service situation DB.
 接客評価装置3の接客発話評価部33は、出力部34を介して管理サーバ2の管理DB2aの接客状況DBにアクセスし、ステップS58-1において指定された接客イベントIDに対応するイベント開始時刻及びイベント終了時刻を抽出し、更に、イベント開始時刻及びイベント終了時刻に対応する映像データ及び音声データをレコーダ装置4より取得して出力部34に渡す(S58-2)。 The customer service utterance evaluation unit 33 of the customer service evaluation device 3 accesses the customer service status DB of the management DB 2a of the management server 2 via the output unit 34, and the event start time corresponding to the customer service event ID specified in step S58-1. The event end time is extracted, and video data and audio data corresponding to the event start time and event end time are acquired from the recorder device 4 and passed to the output unit 34 (S58-2).
 また、出力部34は、詳細表示画面WD7から図35に示す接客状況プレビュー画面WD8に切り替えて表示装置35に表示させ、取得した映像データを表示装置35において出力(再生)しながら、音声データをスピーカ装置36から出力させる(S58-2)。図35は、接客状況プレビュー画面における接客位置の修正操作の一例を示す図である。 Further, the output unit 34 switches from the detailed display screen WD7 to the customer service situation preview screen WD8 shown in FIG. 35 to display on the display device 35, and outputs (reproduces) the acquired video data on the display device 35, while outputting the audio data. Output from the speaker device 36 (S58-2). FIG. 35 is a diagram illustrating an example of a customer service position correcting operation on the customer service status preview screen.
 なお、図5に示す接客モニタリングシステム100Cであれば、接客発話評価部33は、接客状況データIDに対応する接客者位置のデータを接客状況DBから取得し、音声データとともに指向性制御部37に渡す。指向性制御部37は、音声データと接客者位置のデータとを用いて、接客状況データIDに対応する接客者IDの店員を最も近くで収音するマイクアレイ装置から店員に向かう方向に音声の指向性を形成して出力部34に渡す(S58-2)。 If the customer service monitoring system 100C shown in FIG. 5 is used, the customer service utterance evaluation unit 33 acquires the customer service location data corresponding to the customer service status data ID from the customer service status DB, and sends it to the directivity control unit 37 together with the voice data. hand over. The directivity control unit 37 uses the voice data and the customer service location data to transmit the voice in the direction from the microphone array device that picks up the customer ID of the customer service ID corresponding to the customer service status data ID to the store employee closest. Directivity is formed and passed to the output unit 34 (S58-2).
 ステップS58-2の後、接客発話評価部33は、図35に示す接客状況プレビュー画面WD8の位置変更ボタンが選択可能にアクティブ状態であって、ユーザ(修正要求者)の指FGにより接客状況プレビュー画面WD8の停止ボタンが選択され、かつユーザ(修正要求者)に該当するアクセス権(修正操作が可能な権限レベルL1)があるか否かを判定する(S58-3)。なお、図5に示す接客モニタリングシステム100Cを除く図2から図4に示す接客モニタリングシステム100,100A,100Bでは、マイクアレイ装置が使用されないので、図35に示す接客状況プレビュー画面WD8の位置変更ボタンは選択不可状態(非アクティブ状態)である。 After step S58-2, the customer service utterance evaluation unit 33 is in an active state so that the position change button of the customer service status preview screen WD8 shown in FIG. 35 can be selected, and the customer service status preview is performed by the user's (correction requester) finger FG. It is determined whether or not the stop button on the screen WD8 is selected and there is an access right (authority level L1 at which correction operation is possible) corresponding to the user (correction requester) (S58-3). In the customer service monitoring systems 100, 100A and 100B shown in FIGS. 2 to 4 except for the customer service monitoring system 100C shown in FIG. 5, the microphone array device is not used, so the position change button on the customer service situation preview screen WD8 shown in FIG. Is in a non-selectable state (inactive state).
 接客状況プレビュー画面WD8の位置変更ボタンが非アクティブ状態である場合、ユーザ(修正要求者)の指FGにより接客状況プレビュー画面WD8の停止ボタンが選択されない場合又はユーザ(修正要求者)に該当するアクセス権が無い場合には、いずれもステップS58-8(図25参照)に進む。 When the position change button of the customer service situation preview screen WD8 is in an inactive state, when the stop button of the customer service situation preview screen WD8 is not selected by the finger FG of the user (correction requester) or access corresponding to the user (correction requester) If there is no right, the process proceeds to step S58-8 (see FIG. 25).
 一方、接客状況プレビュー画面WD8の位置変更ボタンが選択可能なアクティブ状態であって、ユーザ(修正要求者)の指FGにより接客状況プレビュー画面WD8の停止ボタンが選択され、かつ該当するアクセス権(修正操作が可能な権限レベルL1)があると判定された場合には(S58-3、YES)、接客状況プレビュー画面WD8上でユーザ(修正要求者)の指FGにより指向方向が指定されると(S58-4)、指向性制御部37は、ステップS58-4において指定された方向に音声の指向性を形成するための指向方向を変更する(S58-5)。出力部34は、指向方向を変更した指向性形成後の音声データをスピーカ装置36から出力させる。この音声データはユーザ(修正要求者)によって確認される(S58-5)。 On the other hand, the position change button of the customer service status preview screen WD8 is in an active state that can be selected, and the stop button of the customer service status preview screen WD8 is selected by the user's (correction requester) finger FG, and the corresponding access right (correction) If it is determined that there is an authority level L1 that can be operated (S58-3, YES), the pointing direction is designated by the user (correction requester) finger FG on the customer service status preview screen WD8 ( In step S58-4, the directivity control unit 37 changes the directivity direction for forming the sound directivity in the direction specified in step S58-4 (S58-5). The output unit 34 causes the speaker device 36 to output the voice data after the directivity formation with the directivity direction changed. This voice data is confirmed by the user (correction requester) (S58-5).
 接客状況プレビュー画面WD8の位置変更ボタンがユーザ(修正要求者)の指FGにより選択されると(S58-6、YES)、接客発話評価部33は、図32に示す詳細表示画面WD7の接客者情報を、ステップS58-4において指定された指向方向の位置を示す座標(つまり、表示装置35に表示された画面上の位置を示す座標)に変更して表示装置35に表示させ、更に、接客状況DBの該当するレコードの接客者位置を変更(修正)して上書きして保存する(S58-7、図36参照)。図36は、接客状況DBの詳細表示画面に表示された特定のレコードの修正後の接客位置の座標の一例を示す図である。図36では、例えば接客状況データID「4」のレコードの接客者位置(プリセット外)の任意座標(つまり、指向方向が変更された表示装置35の画面上の指定位置を示す座標)のセルCL2が変更されていることが示されている。 When the position change button on the customer service status preview screen WD8 is selected by the user's (correction requester) finger FG (S58-6, YES), the customer service utterance evaluation unit 33 displays the customer on the detail display screen WD7 shown in FIG. The information is changed to the coordinates indicating the position in the pointing direction designated in step S58-4 (that is, the coordinates indicating the position on the screen displayed on the display device 35) and displayed on the display device 35. The customer position of the corresponding record in the situation DB is changed (corrected) and overwritten and stored (S58-7, see FIG. 36). FIG. 36 is a diagram illustrating an example of coordinates of the customer service position after correction of a specific record displayed on the detail display screen of the customer service DB. In FIG. 36, for example, a cell CL2 of an arbitrary coordinate (that is, a coordinate indicating a designated position on the screen of the display device 35 in which the directing direction is changed) of the customer service position (outside the preset) of the record of the customer service status data ID “4” Is shown to have changed.
 一方、接客状況プレビュー画面WD8の位置変更ボタンがユーザ(修正要求者)の指FGにより選択されない場合には(S58-6、NO)、図24に示す接客発話評価値修正処理はステップS58-4に戻る。 On the other hand, when the position change button on the customer service status preview screen WD8 is not selected with the finger FG of the user (correction requester) (S58-6, NO), the customer service utterance evaluation value correction process shown in FIG. Return to.
 ステップS58-7の後、図25において、接客発話評価部33は、表示装置35により表示された詳細表示画面WD7(図32参照)上でユーザ(修正要求者)の指FGにより選択された接客状況データID(ステップS58-1参照)のレコードの接客発話評価値のセルCL1(図34参照)がダブルタップされ、かつ該当するアクセス権(修正操作が可能な権限レベルL1)があるか否かを判定する(S58-8)。図34は、接客状況DBの詳細表示画面WD7に表示された特定のレコードの修正後の接客発話評価値の一例を示す図である。 After step S58-7, in FIG. 25, the customer service utterance evaluation unit 33 receives the customer selected by the finger FG of the user (correction requester) on the detail display screen WD7 (see FIG. 32) displayed by the display device 35. Whether or not the service utterance evaluation value cell CL1 (see FIG. 34) of the record of the situation data ID (see step S58-1) is double-tapped and there is a corresponding access right (authorization level L1 that can be modified). Is determined (S58-8). FIG. 34 is a diagram illustrating an example of a customer utterance evaluation value after correction of a specific record displayed on the detail display screen WD7 of the customer service situation DB.
 詳細表示画面WD7上でユーザ(修正要求者)の指FGにより選択された接客状況データIDのレコードの接客発話評価値のセルCL1がダブルタップされない場合、又は該当するアクセス権が無い場合には(S58-8、NO)、図25に示す接客発話評価値修正処理は終了し、接客評価装置3の処理はステップS54に戻る。 When the cell CL1 of the customer utterance evaluation value of the record of the customer service status data ID selected by the finger FG of the user (correction requester) on the detailed display screen WD7 is not double-tapped, or when there is no corresponding access right ( S58-8, NO), the service utterance evaluation value correction process shown in FIG. 25 is terminated, and the process of the service evaluation apparatus 3 returns to step S54.
 一方、詳細表示画面WD7(図32参照)上でユーザ(修正要求者)の指FGにより選択された接客状況データID(ステップS58-1参照)のレコードの接客発話評価値のセルCL1がダブルタップされ、かつ該当するアクセス権(修正操作が可能な権限レベルL1)があると判定された場合には(S58-8、YES)、例えば修正要求者の指FGを用いた操作により、ダブルタップされたセルCL1の接客発話評価値が修正(変更)された場合には(S58-9、YES)、接客発話評価部33は、修正(変更)された接客発話評価値(図34参照)を接客状況DBに上書きして保存(格納)する(S58-10)。 On the other hand, the cell CL1 of the customer utterance evaluation value of the record of the customer service status data ID (see step S58-1) selected by the user (correction requester) finger FG on the detailed display screen WD7 (see FIG. 32) is double-tapped. And when it is determined that there is a corresponding access right (authority level L1 at which a correction operation can be performed) (YES in S58-8), for example, the operation is performed by using the finger FG of the correction requester to double tap. If the service utterance evaluation value of the cell CL1 has been corrected (changed) (S58-9, YES), the service utterance evaluation unit 33 receives the corrected (changed) service utterance evaluation value (see FIG. 34). The situation DB is overwritten and saved (stored) (S58-10).
 また、修正要求者の指FGを用いた操作により、ダブルタップされたセルCL1の接客発話評価値が修正(変更)されない場合には(S58-9、NO)、図25に示す接客発話評価値修正処理は終了し、接客評価装置3の処理はステップS54に戻る。 If the service utterance evaluation value of the double-tapped cell CL1 is not corrected (changed) by an operation using the finger FG of the correction requester (S58-9, NO), the service utterance evaluation value shown in FIG. The correction process ends, and the process of the customer service evaluation device 3 returns to step S54.
 図29は、1日の時間帯における来店客に対する全接客者の接客発話評価値の集計結果の一例を示す図である。図30Aは、1日の時間帯毎の1人の接客者の接客発話評価値の集計結果の一例を示す図である。図30Bは、1日の接客者毎の接客発話評価値の集計結果の一例を示す図である。図31は、1日の店舗毎の接客発話評価値の集計結果の一例を示す図である。 FIG. 29 is a diagram showing an example of a totaling result of customer utterance evaluation values of all customers with respect to customers in a day time zone. FIG. 30A is a diagram illustrating an example of a totaling result of customer utterance evaluation values of one customer for each time slot of one day. FIG. 30B is a diagram illustrating an example of a total result of customer service utterance evaluation values for each customer service per day. FIG. 31 is a diagram illustrating an example of a totaling result of customer service utterance evaluation values for each store on a day.
 また、接客評価装置3の出力部34は、ユーザ(つまり、少なくとも図26Aに示す権限レベルL1を有する者)の所定の入力操作により、図28に示す接客状況表示画面WD2から、図29に示す接客状況表示画面WD3に切り替えて表示装置35に表示させてもよい。図29では、横軸には1日の時刻が示され、縦軸には来店者(顧客)の数(黒棒参照)とその顧客に対して店員が挨拶することができた件数(言い換えると、接客イベント名「入店挨拶」が店員により適正に実行された件数、白棒参照)が示されている。 Also, the output unit 34 of the customer service evaluation device 3 is shown in FIG. 29 from the customer service status display screen WD2 shown in FIG. 28 by a predetermined input operation of the user (that is, a person having at least the authority level L1 shown in FIG. 26A). It may be switched to the customer service status display screen WD3 and displayed on the display device 35. In FIG. 29, the horizontal axis shows the time of the day, and the vertical axis shows the number of customers (customers) (see black bars) and the number of salesclerks who could greet the customers (in other words, , The number of customer service event names “sales greetings” properly executed by store clerk, see white bars).
 また、接客評価装置3の出力部34は、ユーザ(つまり、少なくとも図26Aに示す権限レベルL1を有する者)の所定の入力操作により、図29に示す接客状況表示画面WD3から、図30Aに示すように、特定の時間帯に限定した接客状況表示画面WD4に切り替えて表示装置35に表示させてもよい。図30Aでは、横軸には1日の時刻が示され、縦軸には来店者(顧客)の数(黒棒参照)とその顧客に対して店員が挨拶することができた件数(言い換えると、接客イベント名「入店挨拶」が店員により適正に実行された件数、白棒参照)が示されている。 Further, the output unit 34 of the customer service evaluation device 3 is shown in FIG. 30A from the customer service status display screen WD3 shown in FIG. 29 by a predetermined input operation of the user (that is, a person having at least the authority level L1 shown in FIG. 26A). In this way, the service status display screen WD4 limited to a specific time zone may be switched to display on the display device 35. In FIG. 30A, the horizontal axis indicates the time of the day, and the vertical axis indicates the number of customers (see black bars) and the number of cases that the store clerk could greet the customer (in other words, , The number of customer service event names “sales greetings” properly executed by store clerk, see white bars).
 また、接客評価装置3の出力部34は、ユーザ(つまり、少なくとも図26Aに示す権限レベルL1を有する者)の所定の入力操作により、図28に示す接客状況表示画面WD2から、図30Bに示す接客状況表示画面WD5に切り替えて表示装置35に表示させてもよい。図30Bでは、1日単位で、店員(例えば4人)毎の挨拶率と平均スコアとレジ対応人数とが対比的に示されている。なお、接客評価装置3の出力部34は、所定の入力操作に応じて、1日単位ではなく、1週間単位や1カ月単位に切り替えて図30Bに示す各項目のデータを集計して再表示してもよい。 Further, the output unit 34 of the customer service evaluation device 3 is shown in FIG. 30B from the customer service status display screen WD2 shown in FIG. 28 by a predetermined input operation of the user (that is, a person having at least the authority level L1 shown in FIG. 26A). It may be switched to the customer service status display screen WD5 and displayed on the display device 35. In FIG. 30B, the greeting rate, the average score, and the number of cashiers corresponding to each store clerk (for example, four people) are shown in comparison on a daily basis. It should be noted that the output unit 34 of the customer service evaluation device 3 aggregates and redisplays the data of each item shown in FIG. 30B by switching to a weekly or monthly unit instead of a daily unit according to a predetermined input operation. May be.
 また、接客評価装置3の出力部34は、ユーザ(つまり、少なくとも図26Aに示す権限レベルL1を有する者)の所定の入力操作により、図28に示す接客状況表示画面WD2から、図31に示す接客状況表示画面WD6に切り替えて表示装置35に表示させてもよい。図31では、1日単位で、店舗(例えば4個)毎の顧客の来店数と挨拶率と平均スコアとレジ対応人数とが対比的に示されている。なお、接客評価装置3の出力部34は、所定の入力操作に応じて、1日単位ではなく、1週間単位や1カ月単位に切り替えて図31に示す各項目のデータを集計して再表示してもよい。 Further, the output unit 34 of the customer service evaluation device 3 is shown in FIG. 31 from the customer service status display screen WD2 shown in FIG. 28 by a predetermined input operation of a user (that is, a person having at least the authority level L1 shown in FIG. 26A). It may be switched to the customer service status display screen WD6 and displayed on the display device 35. In FIG. 31, the number of customers visiting the store, the greeting rate, the average score, and the number of cashiers corresponding to each store (for example, four) are shown in comparison on a daily basis. Note that the output unit 34 of the customer service evaluation device 3 switches the data for each item shown in FIG. 31 and redisplays the data by switching to a week or month instead of a day according to a predetermined input operation. May be.
 以上により、本実施形態の接客モニタリングシステム100,100A,100B,100Cは、所定の接客イベント毎の接客イベント判定条件を含む接客イベント情報DB(接客イベントデータ)とPOS端末5(所定の業務端末)に対する店員(従業員)のPOS端末5の操作履歴を示すPOS操作履歴データとを基に、店員の接客イベントを検出し、モニタリングデータ4a又はモニタリングデータ4bに含まれる店員の音声データを基に、POS端末5の操作時における所定の発話キーワードに対応する接客発話評価値を算定する。また、接客モニタリングシステム100,100A,100B,100Cは、算定された接客発話評価値を、接客者ID(店員の識別情報)と接客者位置(店員の接客位置)及び接客時刻により特定される店員の音声データに対応付けて記憶する。 As described above, the customer service monitoring systems 100, 100A, 100B, and 100C of the present embodiment have the customer service event information DB (customer service event data) including customer service event determination conditions for each predetermined customer service event and the POS terminal 5 (predetermined business terminal). Based on the POS operation history data indicating the operation history of the POS terminal 5 of the clerk (employee) for the clerk, the customer service event of the clerk is detected, and based on the voice data of the clerk included in the monitoring data 4a or the monitoring data 4b, A service utterance evaluation value corresponding to a predetermined utterance keyword when the POS terminal 5 is operated is calculated. In addition, the customer service monitoring systems 100, 100A, 100B, and 100C determine the calculated customer service utterance evaluation values based on the customer service ID (sales employee identification information), the customer service position (sales staff service location), and the customer service time. Are stored in association with the audio data.
 これにより、接客モニタリングシステム100,100A,100B,100Cは、従来技術のように調査員のような人的リソースを用いること無く、接客される顧客のプライバシーを広範に保護した上で、所定の収音領域(例えば店舗)内の顧客に対する各種の接客イベント時における該当接客者(従業員)の接客発話の監視により、該当接客者の接客発話内容の客観性を接客発話評価値として得ることができ、従業員の顧客に対する接客状況を的確かつ客観的に評価することができる。 As a result, the customer service monitoring systems 100, 100A, 100B, and 100C can protect the customer's customer's privacy extensively without using human resources like an investigator, as in the prior art. By monitoring the customer utterances of the corresponding customer (employee) during various customer service events for customers in the sound domain (for example, stores), the objectivity of the customer utterance content of the customer can be obtained as a customer utterance evaluation value. It is possible to accurately and objectively evaluate the customer service situation of employees.
 また、接客モニタリングシステム100,100A,100B,100Cは、第2記憶部であるレコーダ装置4に記憶される従業員の音声データに顧客のプライバシー保護を示す所定情報を示すデータ(つまり、プライバシー保護マーク)が付与されている場合には従業員の接客イベントの検出を省略するので、顧客が介在するような接客イベントを除外することで、接客イベントの検出の際に顧客のプライバシーを一層明確に保護することができる。 In addition, the customer service monitoring systems 100, 100A, 100B, and 100C have data (that is, privacy protection marks) indicating predetermined information indicating customer privacy protection in employee voice data stored in the recorder device 4 as the second storage unit. ) Is omitted, the detection of employee service events is omitted, so that customer privacy is more clearly protected when customer service events are detected by excluding customer service events that involve customers. can do.
 また、接客モニタリングシステム100,100A,100B,100Cは、所定の接客イベント毎の発話想定キーワードを含む接客発話評価DBの発話想定キーワード表(キーワードデータ)を管理サーバ2の管理DB2aにおいて更に記憶し、店員の音声データに接客イベントに対応する発話想定キーワードが含まれていない場合には接客発話評価値をゼロに設定し又は接客発話評価値から所定点数を減点するので、接客イベント時に発話想定キーワードを発話しない店員に対する接客状況を正確に評価することができる。 In addition, the customer service monitoring systems 100, 100A, 100B, and 100C further store in the management DB 2a of the management server 2 the utterance assumption keyword table (keyword data) of the customer utterance evaluation DB including the utterance assumption keywords for each predetermined customer service event. If the speech utterance keyword corresponding to the customer service event is not included in the clerk's voice data, the customer service utterance evaluation value is set to zero or a predetermined score is deducted from the customer service utterance evaluation value. It is possible to accurately evaluate the customer service situation for a store clerk who does not speak.
 また、接客モニタリングシステム100,100A,100B,100Cは、所定の接客イベント毎の発話想定キーワードを含む接客発話評価DBの発話想定キーワード表(キーワードデータ)を管理サーバ2の管理DB2aにおいて更に記憶し、店員の音声データに接客イベントに対応する発話想定キーワードが含まれている場合には店員の音声データを発話想定キーワードに対応するキーワードの発話部分のみに切り取って更新し上書き保存するので、不要なノイズ音をカットしてスコアリング処理精度を向上できるとともに、店員の音声データの容量を低減することができる。さらに、正確な接客発話評価値の算定に資することができる。 In addition, the customer service monitoring systems 100, 100A, 100B, and 100C further store in the management DB 2a of the management server 2 the utterance assumption keyword table (keyword data) of the customer utterance evaluation DB including the utterance assumption keywords for each predetermined customer service event. If the store clerk's voice data includes an utterance assumption keyword corresponding to a customer service event, the clerk's voice data is cut out only to the utterance part of the keyword corresponding to the utterance assumption keyword, and is overwritten and saved, so unnecessary noise The scoring process accuracy can be improved by cutting the sound, and the volume of the clerk's voice data can be reduced. Furthermore, it can contribute to the calculation of an accurate customer utterance evaluation value.
 また、接客モニタリングシステム100,100A,100B,100Cは、所定の接客イベント毎の発話想定キーワードの音声データを含む接客発話評価DBの接客発話モデル一覧表(キーワード音声データ)を管理サーバ2の管理DB2aにおいて更に記憶し、更新された店員の音声データの発話想定キーワードの発話長さがキーワード音声データの発話想定キーワードの発話長さから所定範囲を超える場合に、接客発話評価値から所定点数を減点するので、接客イベント時に発話想定キーワードを模範的な発話長さから逸脱して発話した店員に対する接客状況を正確に評価することができる。 Further, the customer service monitoring systems 100, 100A, 100B, and 100C use the customer service utterance model list (keyword speech data) of the customer service utterance evaluation DB including the speech data of the utterance assumed keywords for each predetermined customer service event as a management DB 2a of the management server 2. In the case where the utterance length of the utterance assumption keyword of the clerk's voice data that has been stored and updated exceeds the utterance length of the utterance assumption keyword of the keyword voice data, the predetermined score is subtracted from the customer utterance evaluation value Therefore, it is possible to accurately evaluate the customer service situation for a store clerk who utters an utterance assumption keyword that deviates from an exemplary utterance length at the customer service event.
 また、接客モニタリングシステム100,100A,100B,100Cは、所定の接客イベント毎の発話想定キーワードの音声データを含む接客発話評価DBの接客発話モデル一覧表(キーワード音声データ)を管理サーバ2の管理DB2aにおいて更に記憶し、更新された店員の音声データの発話想定キーワードの音素毎の周波数がキーワード音声データの発話想定キーワードの音素毎の基本周波数から所定範囲を超える場合に接客発話評価値から所定点数を減点するので、接客イベント時に発話想定キーワードを模範的な基本周波数から逸脱して発話した店員に対する接客状況を正確に評価することができる。 Further, the customer service monitoring systems 100, 100A, 100B, and 100C use the customer service utterance model list (keyword speech data) of the customer service utterance evaluation DB including the speech data of the utterance assumed keywords for each predetermined customer service event as a management DB 2a of the management server 2. In the case where the frequency for each phoneme of the utterance assumption keyword of the clerk's speech data updated and exceeds the predetermined range from the basic frequency for each phoneme of the utterance assumption keyword of the keyword speech data, a predetermined score is obtained from the customer utterance evaluation value. Since the points are deducted, it is possible to accurately evaluate the customer service situation with respect to the store clerk who utters by deviating from the typical basic frequency at the customer service event.
 また、接客モニタリングシステム100Aは、店員に個々に装着される接客者マイク装置SM1,…,SMLにより収音される各店員の音声データと接客イベントデータとを基に店員の接客イベントを検出するので、接客者マイク装置以外のマイク装置(例えば天井面に設置されたマイクロホン)と店員とが離れている場合に比べて、店員の音声を明瞭に収音することができ、入店挨拶のような接客イベントを正確に検出することができる。 Further, the customer service monitoring system 100A detects the customer service event of the store clerk based on the voice data and customer service event data of each store clerk picked up by the customer microphone devices SM1,..., SML individually attached to the store clerk. Compared with the case where the store clerk is far away from the microphone device other than the customer microphone device (for example, a microphone installed on the ceiling surface), the store clerk's voice can be clearly picked up. A customer service event can be accurately detected.
 また、接客モニタリングシステム100B,100Cは、カメラ装置C1,…,CMの撮像により得られた所定の収音領域(例えば店舗)の既定位置の映像データをモニタリングデータ4bとしてレコーダ装置4において更に記憶し、この映像データを基に店員の接客イベントを検出するので、所定の収音領域(例えば店舗)の会計開始挨拶の接客イベントが行われるPOS端末5の付近を既定位置とする映像データを画像処理することで、会計開始挨拶の接客イベントが適正に行われているか否かを正確に評価することができる。 In addition, the customer service monitoring systems 100B and 100C further store in the recorder device 4 video data at a predetermined position in a predetermined sound collection area (for example, a store) obtained by imaging of the camera devices C1, ..., CM as monitoring data 4b. Since the customer service event of the store clerk is detected based on this video data, the video data having a predetermined position near the POS terminal 5 where the customer service event of the accounting start greeting of a predetermined sound collection area (for example, store) is performed is processed. By doing so, it is possible to accurately evaluate whether or not the customer service event of the accounting start greeting is properly performed.
 また、接客モニタリングシステム100B,100Cは、センサ装置S1,…,SNにより得られた所定の収音領域における顧客の登場又は退場(例えば店舗への顧客の入店又は退店)の検出結果をモニタリングデータ4bとしてレコーダ装置4において更に記憶し、この検出結果を基に店員の接客イベントを検出するので、所定の収音領域(例えば店舗)の入退店挨拶の接客イベントが行われる動機を与えるセンサ装置(例えば開閉する自動ドア)の検出結果に応じて、入退店挨拶の接客イベントが適正に行われているか否かを正確に評価することができる。 Moreover, the customer service monitoring systems 100B and 100C monitor the detection results of the appearance or exit of a customer (for example, the customer entering or leaving the store) in a predetermined sound collection area obtained by the sensor devices S1,. Further, the data is recorded in the recorder device 4 as data 4b, and a customer service event of a store clerk is detected based on the detection result. Therefore, a sensor that gives a motivation for a customer service event for entering and leaving a store in a predetermined sound collection area (for example, a store) Depending on the detection result of the device (for example, an automatic door that opens and closes), it is possible to accurately evaluate whether or not the customer service event for entering and leaving the store is properly performed.
 また、接客モニタリングシステム100Cは、マイクアレイ装置AM1,…,AMLのうちいずれかから所定の指向方向(例えば店員(従業員)の接客時の定位置(例えばレジカウンター)など)に音声の指向性を形成した店員の音声データを基に接客発話評価値を算定する時に、店員(従業員)の音声が強調されるので、指向性が形成されない場合に比べて、接客発話評価値の算定精度を向上することができ、店員に対する接客発話評価値を正確に算定することができる。 Further, the customer service monitoring system 100C has a voice directivity from any one of the microphone array devices AM1,..., AML to a predetermined directivity direction (for example, a fixed position (for example, a cashier counter) at the time of customer service). When calculating the customer service utterance evaluation value based on the voice data of the store clerk who formed the clerk, the voice of the store clerk (employee) is emphasized, so the calculation accuracy of the service utterance evaluation value is higher than when the directivity is not formed It is possible to improve, and the customer service utterance evaluation value for the store clerk can be accurately calculated.
 また、接客モニタリングシステム100,100A,100B,100Cは、接客イベント毎の接客発話評価値の表示装置35における接客状況DBの詳細表示画面WD7(接客発話評価値表示画面)の閲覧操作の権限情報を含む権限レベル(権限データ)を管理サーバ2の管理DB2aの閲覧者DBにおいて更に記憶し、接客発話評価値表示画面の閲覧要求者(ユーザ)の権限情報が権限データに含まれる閲覧操作の権限情報を満たす場合に、接客発話評価値表示画面を表示装置35に表示させることができる。 In addition, the customer service monitoring systems 100, 100A, 100B, and 100C display authority information for the browsing operation on the detailed display screen WD7 (service customer utterance evaluation value display screen) of the customer service situation DB in the customer utterance evaluation value display device 35 for each customer service event. The authority level (authority data) that is included is further stored in the viewer DB of the management DB 2a of the management server 2, and the authority information of the browsing requester (user) on the service utterance evaluation value display screen is included in the authority data. If the condition is satisfied, the customer service utterance evaluation value display screen can be displayed on the display device 35.
 また、接客モニタリングシステム100,100A,100B,100Cは、接客状況DBの詳細表示画面WD7(接客発話評価値表示画面)に対する接客発話評価値の修正操作の権限情報を権限データとして更に含み、接客発話評価値の修正要求者(ユーザ)の権限情報が権限データに含まれる修正操作の権限情報を満たす場合に、接客発話評価値の修正操作に応じて、接客発話評価値表示画面の接客発話評価値を更新(修正)することができる。 In addition, the customer service monitoring systems 100, 100A, 100B, and 100C further include, as authority data, authority information for correcting the customer service utterance evaluation value for the service status DB detail display screen WD7 (customer service utterance evaluation value display screen). If the authority information of the requester for correction of the evaluation value (user) satisfies the authority information of the correction operation included in the authority data, the customer utterance evaluation value on the customer service utterance evaluation value display screen according to the service utterance evaluation value correction operation Can be updated (corrected).
 また、接客モニタリングシステム100B,100Cは、カメラ装置C1,…,CMの撮像により得られた所定の収音領域の既定位置の映像データをモニタリングデータ4bとしてレコーダ装置4において更に記憶し、閲覧操作及び修正操作の両方の権限情報を有するユーザの操作によって映像データを出力している間の店員の接客イベントの接客位置の修正操作に応じて、接客発話評価値表示画面の接客発話評価値を更新(修正)することができる。 In addition, the customer service monitoring systems 100B and 100C further store video data at a predetermined position in a predetermined sound collection area obtained by imaging of the camera devices C1,..., CM as monitoring data 4b in the recorder device 4 for browsing operation and The customer service utterance evaluation value on the customer service utterance evaluation value display screen is updated in accordance with the operation of correcting the customer service position of the customer service event of the store clerk while the video data is output by the operation of the user who has the authority information for both corrective operations ( Correction).
 また、接客モニタリングシステム100,100A,100B,100Cは、少なくとも閲覧操作の権限情報を有するユーザによる所定の入力操作に応じて、接客発話評価値表示画面の接客イベント毎の接客発話評価値を所定項目毎に対比的に表示装置35に表示させることができ、所定項目毎の比較を簡易に行うことができる。 In addition, the customer service monitoring systems 100, 100A, 100B, and 100C set the service utterance evaluation value for each customer service event on the service utterance evaluation value display screen according to a predetermined input operation by a user who has at least browsing authority information. Each display can be displayed on the display device 35 in a comparative manner, and comparison for each predetermined item can be easily performed.
 なお、本実施形態において、図6に示すステップS2のシフト処理は、モニタリングデータ4aを切り取るための起点となる切り取り開始時刻を所定時間tsシフトする方法に限定されない。シフト処理のバリエーションについて、図39~図41を参照して説明する。図39は、モニタリングデータ4aの切り出し処理のバリエーションの一例を示す説明図である。図40は、接客イベント情報処理の詳細な動作手順の他の一例を説明するフローチャートである。図41は、接客発話評価DBの一部を構成する接客イベントごとの発話想定区間表の一例を示す図である。なお、図40の説明では、図9に示すそれぞれの処理と同一の処理については同一のステップ番号を付与して説明を簡略化又は省略する。 In the present embodiment, the shift process of step S2 shown in FIG. 6 is not limited to the method of shifting the cut start time that is the starting point for cutting the monitoring data 4a by a predetermined time ts. Variations of the shift process will be described with reference to FIGS. FIG. 39 is an explanatory diagram showing an example of a variation of the cutting process of the monitoring data 4a. FIG. 40 is a flowchart for explaining another example of the detailed operation procedure of customer service event information processing. FIG. 41 is a diagram illustrating an example of an utterance assumption section table for each service event that constitutes a part of the service utterance evaluation DB. In the description of FIG. 40, the same processes as those shown in FIG. 9 are assigned the same step numbers, and the description is simplified or omitted.
 例えば図38と同様に、図39に示すモニタリングデータ4aに、時刻t1に会計完了操作が発生したことを示すPOS操作履歴データと、時刻t2から時刻t3にわたって店員が「ありがとうございました」と発話した音声データとが記憶されている場合を想定する。この場合、図38を参照して説明したシフト処理を用いると、接客評価装置3のモニタリングデータ抽出部38は、モニタリングデータ4aの先頭から所定時間間隔分(つまり、切り取り範囲RG0)の音声データとPOS操作履歴データとを切り取ったモニタリングデータ4a1を抽出する。但し、図38と同様にモニタリングデータ4a1には、会計完了操作に対応する店員の音声データが含まれていないため、接客イベントの検出ができず、正確な接客評価が不可である。 For example, as in FIG. 38, in the monitoring data 4a shown in FIG. 39, the POS operation history data indicating that the accounting completion operation occurred at time t1, and the salesclerk uttered “Thank you” from time t2 to time t3. Assume that audio data is stored. In this case, when the shift process described with reference to FIG. 38 is used, the monitoring data extraction unit 38 of the customer service evaluation device 3 receives the audio data for a predetermined time interval (that is, the cutout range RG0) from the head of the monitoring data 4a. The monitoring data 4a1 cut out from the POS operation history data is extracted. However, as in FIG. 38, the monitoring data 4a1 does not include the clerk's voice data corresponding to the transaction completion operation. Therefore, the customer service event cannot be detected, and accurate customer service evaluation is impossible.
 そこで、接客評価装置3のモニタリングデータ抽出部38は、図40に示すステップS14の後、図41に示す発話想定区間表を参照し、接客イベントIDに対応した発話想定区間に関する情報を取得し、モニタリングデータ4a1から、切り取るべき音声データの範囲を変更してモニタリングデータ4a2’を再取得する(図40に示すS14-1)。ステップS14-1以降の処理は、図9に示すステップS15以降の処理と同一であるため、説明を省略する。 Therefore, the monitoring data extraction unit 38 of the customer service evaluation device 3 acquires information related to the utterance assumption section corresponding to the customer service event ID with reference to the utterance assumption section table shown in FIG. 41 after step S14 shown in FIG. The monitoring data 4a2 ′ is reacquired from the monitoring data 4a1 by changing the range of the audio data to be cut (S14-1 shown in FIG. 40). The processing after step S14-1 is the same as the processing after step S15 shown in FIG.
 ここで、ステップS14-1の詳細について、図41を参照して説明する。図41に示す発話想定区間表には、接客イベントIDと、接客イベント名と、発話想定区間とが対応付けられている。例えば接客イベントIDが「EID1」の「会計完了挨拶」である場合には、店員が発話すると想定される区間は、その接客イベント(つまり、会計完了挨拶)が検出された時点から10秒間である。従って、接客評価装置3のモニタリングデータ抽出部38は、接客イベントIDを基に、モニタリングデータ4aの音声データを切り取るための起点となる切り取り開始時刻を、接客イベント(例えば会計完了挨拶)の検出時である時刻t1に変更し、時刻t1から10秒間の音声データを切り取り範囲RG1の音声データとして再取得する。これにより、接客評価装置3のモニタリングデータ抽出部38は、1回目に取得したモニタリングデータ4a1を、モニタリングデータ4a2’にアップデートするので、接客イベント時に店員が発話する音声の区間に相当する音声データが得られ、接客イベントの検出精度が向上する。 Here, the details of step S14-1 will be described with reference to FIG. In the utterance assumption section table shown in FIG. 41, a service event ID, a service event name, and an utterance section are associated. For example, when the customer service event ID is “accounting completion greeting” with “EID1”, the section assumed to be uttered by the store clerk is 10 seconds from the time when the customer service event (that is, the accounting completion greeting) is detected. . Therefore, the monitoring data extraction unit 38 of the customer service evaluation device 3 uses the customer service event ID as a starting point for cutting the audio data of the monitoring data 4a when the customer service event (for example, accounting completion greeting) is detected. Is changed to the time t1, and the audio data for 10 seconds from the time t1 is re-acquired as the audio data in the cutout range RG1. As a result, the monitoring data extraction unit 38 of the customer service evaluation device 3 updates the monitoring data 4a1 acquired at the first time to the monitoring data 4a2 ′, so that the audio data corresponding to the voice section spoken by the store clerk at the customer service event is obtained. As a result, the customer service event detection accuracy is improved.
 また、図41に示すように、音声データの切り取り範囲は、接客イベント毎に異なり、接客イベントIDが「EID2」の「会計開始挨拶」では、接客イベント(つまり、会計開始挨拶)が検出された時点を起点とした前後5秒間にわたる計10秒間である(例えば図39に示す切り取り範囲RG3参照)。 In addition, as shown in FIG. 41, the cut-out range of the voice data is different for each customer service event, and a customer service event (that is, a customer service start greeting) is detected in the customer service event ID “EID2” “account start greeting”. The total time is 10 seconds over 5 seconds before and after the starting point (see, for example, the cutout range RG3 shown in FIG. 39).
 また、接客イベントIDが「EID3」の「入店挨拶」では、音声データの切り取り範囲は、接客イベントIDが「EID1」の「会計完了挨拶」と同様に、接客イベント(つまり、入店挨拶)が検出された時点から10秒間である。更に、接客イベントIDが「EID4」の「退店挨拶」では、音声データの切り取り範囲は、接客イベント(つまり、退店挨拶)が検出された時点から10秒前の期間である(例えば図39に示す切り取り範囲RG2参照)。 In addition, in the “entry greeting” with the customer service event ID “EID3”, the cut-out range of the audio data is the customer service event (that is, the customer greeting) similar to the “accounting completion greeting” with the customer service event ID “EID1”. It is 10 seconds from the time when is detected. Further, in the “sales greeting” with the customer service event ID “EID4”, the cut-out range of the audio data is a period of 10 seconds before the customer service event (that is, the salutation greeting) is detected (for example, FIG. 39). The cutting range RG2 shown in FIG.
 これにより、接客評価装置3のモニタリングデータ抽出部38は、検出された接客イベントIDに応じた接客イベント時の店員の発話音声が含まれると想定される音声区間に関する情報を発話想定区間表として保持でき、これを利用することで、接客イベント毎に店員が発話した最適な音声区間の音声データをモニタリングデータとして切り取って抽出することができ、正確な接客評価を行うことができる。 As a result, the monitoring data extraction unit 38 of the customer service evaluation device 3 holds, as an utterance assumption interval table, information related to a speech segment that is assumed to include the speech of the clerk at the customer service event corresponding to the detected customer service event ID. By using this, it is possible to cut out and extract the voice data of the optimum voice section spoken by the store clerk for each customer service event as monitoring data, and to perform accurate customer service evaluation.
 以上、図面を参照しながら各種の実施形態について説明したが、本開示はかかる例に限定されないことは言うまでもない。当業者であれば、特許請求の範囲に記載された範疇内において、各種の変更例または修正例に想到し得ることは明らかであり、それらについても当然に本開示の技術的範囲に属するものと了解される。 Although various embodiments have been described above with reference to the drawings, it goes without saying that the present disclosure is not limited to such examples. It will be apparent to those skilled in the art that various changes and modifications can be made within the scope of the claims, and these are naturally within the technical scope of the present disclosure. Understood.
 本開示は、調査員のような人的リソースを用いず、顧客のプライバシーを広範に保護した上で、店舗内の顧客に対する各種の接客イベント時における該当接客者の発話内容を監視することで接客状況を的確かつ客観的に評価する利便性を向上させる接客モニタリングシステム及び接客モニタリング方法として有用である。 This disclosure does not use human resources like investigators, and protects customer privacy extensively, and monitors customer utterances during various customer service events for customers in the store. It is useful as a service monitoring system and service monitoring method that improves the convenience of accurately and objectively evaluating the situation.
2 管理サーバ
2a 管理DB
3,3C 接客評価装置
4 レコーダ装置
4a,4b モニタリングデータ
5 POS端末
31 操作部
32,53 メモリ
33 接客発話評価部
34 出力部
35,52 表示装置
36 スピーカ装置
37 指向性制御部
51 入力装置
331 接客イベント検出部
332 接客発話評価値算定部
100,100A,100B,100C 接客モニタリングシステム
AM1,AML マイクアレイ装置
C1,CM カメラ装置
M1,ML マイク装置
S1,SN センサ装置
SM1,SML 接客者マイク装置
2 Management server 2a Management DB
3, 3C Customer service evaluation device 4 Recorder device 4a, 4b Monitoring data 5 POS terminal 31 Operation unit 32, 53 Memory 33 Service utterance evaluation unit 34 Output unit 35, 52 Display device 36 Speaker device 37 Directivity control unit 51 Input device 331 Customer service Event detection unit 332 Service utterance evaluation value calculation unit 100, 100A, 100B, 100C Service monitoring system AM1, AML Microphone array device C1, CM Camera device M1, ML Microphone device S1, SN Sensor device SM1, SML Customer service microphone device

Claims (16)

  1.  所定の収音領域における従業員の音声を収音する収音部と、
     所定の接客イベント毎の判定条件を含む接客イベントデータを記憶する第1記憶部と、
     所定の業務端末に対する従業員の操作履歴を示す端末操作履歴データと、前記収音部により収音された前記従業員の音声データとを対応付けて記憶する第2記憶部と、
     前記第1記憶部に記憶される前記接客イベントデータと前記第2記憶部に記憶される前記端末操作履歴データとを基に、前記従業員の前記接客イベントを検出する検出部と、
     前記検出部により検出された前記接客イベントにおいて、前記第2記憶部に記憶される前記従業員の音声データを基に、前記業務端末の操作時における所定の発話キーワードに対応する接客発話評価値を算定する算定部と、
     前記算定部により算定された前記接客発話評価値を、前記従業員の識別情報と前記従業員の接客位置及び接客時刻により特定される前記従業員の音声データに対応付けて記憶する出力部と、を備える、
     接客モニタリングシステム。
    A sound collection unit for collecting employee's voice in a predetermined sound collection area;
    A first storage unit for storing customer service event data including determination conditions for each predetermined customer service event;
    A second storage unit that stores terminal operation history data indicating an operation history of an employee for a predetermined business terminal and voice data of the employee collected by the sound collection unit in association with each other;
    A detection unit for detecting the customer service event of the employee based on the customer service event data stored in the first storage unit and the terminal operation history data stored in the second storage unit;
    In the customer service event detected by the detection unit, based on the employee voice data stored in the second storage unit, a customer service speech evaluation value corresponding to a predetermined speech keyword at the time of operation of the business terminal is calculated. A calculation unit to calculate,
    An output unit that stores the customer utterance evaluation value calculated by the calculation unit in association with the employee identification data, the employee voice data specified by the employee reception position and the customer reception time, and Comprising
    Service monitoring system.
  2.  請求項1に記載の接客モニタリングシステムであって、
     前記検出部は、前記第2記憶部に記憶される前記従業員の音声データに顧客のプライバシー保護を示す情報が含まれる場合に、前記従業員の前記接客イベントの検出を省略する、
     接客モニタリングシステム。
    The customer service monitoring system according to claim 1,
    The detection unit omits detection of the customer service event of the employee when the employee's voice data stored in the second storage unit includes information indicating customer privacy protection.
    Service monitoring system.
  3.  請求項1に記載の接客モニタリングシステムであって、
     前記第1記憶部は、前記所定の接客イベント毎の発話想定キーワードを含むキーワードデータを更に記憶し、
     前記算定部は、前記第2記憶部に記憶される前記従業員の音声データに、前記検出部により検出された前記接客イベントに対応する前記発話想定キーワードが含まれていない場合に、前記接客発話評価値をゼロに設定し又は前記接客発話評価値から所定点数を減点する、
     接客モニタリングシステム。
    The customer service monitoring system according to claim 1,
    The first storage unit further stores keyword data including an utterance assumption keyword for each predetermined customer service event,
    The calculation unit, when the employee speech data stored in the second storage unit does not include the speech utterance keyword corresponding to the customer service event detected by the detection unit, Set the evaluation value to zero or deduct a predetermined point from the customer utterance evaluation value,
    Service monitoring system.
  4.  請求項1に記載の接客モニタリングシステムであって、
     前記第1記憶部は、前記所定の接客イベント毎の発話想定キーワードを含むキーワードデータを更に記憶し、
     前記算定部は、前記第2記憶部に記憶される前記従業員の音声データに、前記検出部により検出された前記接客イベントに対応する前記発話想定キーワードが含まれている場合に、前記第2記憶部に記憶される前記従業員の音声データを前記発話想定キーワードの音声データに更新する、
     接客モニタリングシステム。
    The customer service monitoring system according to claim 1,
    The first storage unit further stores keyword data including an utterance assumption keyword for each predetermined customer service event,
    The calculation unit may include the second utterance keyword corresponding to the customer service event detected by the detection unit in the employee voice data stored in the second storage unit. Updating the employee's voice data stored in the storage unit to the voice data of the utterance assumption keyword;
    Service monitoring system.
  5.  請求項4に記載の接客モニタリングシステムであって、
     前記第1記憶部は、前記所定の接客イベント毎の発話想定キーワードの音声データを含むキーワード音声データを更に記憶し、
     前記算定部は、更新された前記従業員の音声データの前記発話想定キーワードの発話長さが前記第1記憶部に記憶される前記キーワード音声データの前記発話想定キーワードの発話長さから第1所定範囲を超える場合に、前記接客発話評価値から所定点数を減点する、
     接客モニタリングシステム。
    The customer service monitoring system according to claim 4,
    The first storage unit further stores keyword voice data including voice data of an assumed utterance keyword for each predetermined customer service event,
    The calculation unit calculates a first predetermined value from the utterance length of the utterance assumption keyword of the keyword voice data stored in the first storage unit as the utterance length of the utterance assumption keyword of the updated voice data of the employee. When exceeding the range, a predetermined score is deducted from the customer service utterance evaluation value,
    Service monitoring system.
  6.  請求項4に記載の接客モニタリングシステムであって、
     前記第1記憶部は、前記所定の接客イベント毎の発話想定キーワードの音声データを含むキーワード音声データを更に記憶し、
     前記算定部は、更新された前記従業員の音声データの前記発話想定キーワードの音素毎の周波数が前記第1記憶部に記憶される前記キーワード音声データの前記発話想定キーワードの音素毎の基本周波数から第2所定範囲を超える場合に、前記接客発話評価値から所定点数を減点する、
     接客モニタリングシステム。
    The customer service monitoring system according to claim 4,
    The first storage unit further stores keyword voice data including voice data of an assumed utterance keyword for each predetermined customer service event,
    The calculation unit calculates the frequency for each phoneme of the utterance assumed keyword of the updated voice data of the employee from the fundamental frequency for each phoneme of the utterance assumed keyword of the keyword speech data stored in the first storage unit. When exceeding the second predetermined range, a predetermined score is subtracted from the customer utterance evaluation value,
    Service monitoring system.
  7.  請求項1に記載の接客モニタリングシステムであって、
     前記収音部は、前記従業員に個々に装着されるマイク装置であり、
     前記検出部は、前記第1記憶部に記憶される前記接客イベントデータと前記マイク装置により収音された前記従業員の音声データとを基に、前記従業員の前記接客イベントを更に検出する、
     接客モニタリングシステム。
    The customer service monitoring system according to claim 1,
    The sound collection unit is a microphone device that is individually attached to the employee,
    The detection unit further detects the customer service event of the employee based on the customer service event data stored in the first storage unit and the employee's voice data collected by the microphone device;
    Service monitoring system.
  8.  請求項1に記載の接客モニタリングシステムであって、
     前記所定の収音領域の既定位置を撮像する撮像部、を更に備え、
     前記第2記憶部は、前記撮像部により撮像された前記所定の収音領域の既定位置における映像データを更に記憶し、
     前記検出部は、前記第2記憶部に記憶される前記所定の収音領域の既定位置における映像データを基に、前記従業員の前記接客イベントを検出する、
     接客モニタリングシステム。
    The customer service monitoring system according to claim 1,
    An imaging unit that images a predetermined position of the predetermined sound collection area;
    The second storage unit further stores video data at a predetermined position of the predetermined sound collection area captured by the imaging unit,
    The detection unit detects the customer service event of the employee based on video data at a predetermined position of the predetermined sound collection area stored in the second storage unit;
    Service monitoring system.
  9.  請求項8に記載の接客モニタリングシステムであって、
     前記所定の収音領域に対する顧客の登場又は退場を検出する顧客検出部と、を更に備え、
     前記第2記憶部は、前記顧客検出部による前記顧客の登場又は退場の検出結果を更に記憶し、
     前記検出部は、前記顧客検出部による前記顧客の登場又は退場の検出結果を基に、前記従業員の前記接客イベントを検出する、
     接客モニタリングシステム。
    The customer service monitoring system according to claim 8,
    A customer detection unit for detecting the appearance or exit of the customer with respect to the predetermined sound collection area,
    The second storage unit further stores a detection result of the appearance or exit of the customer by the customer detection unit,
    The detection unit detects the customer service event of the employee based on the detection result of the appearance or leaving of the customer by the customer detection unit,
    Service monitoring system.
  10.  請求項8又は9に記載の接客モニタリングシステムであって、
     前記第2記憶部に記憶される前記従業員の音声データを基に、前記収音部から所定の指向方向に音声の指向性を形成する指向性制御部、を更に備え、
     前記算定部は、前記指向性制御部により前記音声の指向性が形成された前記従業員の音声データを基に、前記接客発話評価値を算定する、
     接客モニタリングシステム。
    The customer service monitoring system according to claim 8 or 9,
    A directivity control unit that forms voice directivity from the sound collection unit in a predetermined directivity direction based on the employee voice data stored in the second storage unit;
    The calculation unit calculates the customer service utterance evaluation value based on the voice data of the employee in which the voice directivity is formed by the directivity control unit.
    Service monitoring system.
  11.  請求項1に記載の接客モニタリングシステムであって、
     前記第1記憶部は、前記接客イベント毎に前記算定部により算定された前記接客発話評価値の表示部における接客発話評価値表示画面の閲覧操作の権限情報を含む権限データを更に記憶し、
     前記出力部は、前記接客発話評価値表示画面の閲覧要求者の権限情報が前記第1記憶部に記憶される前記権限データに含まれる前記閲覧操作の権限情報を満たす場合に、前記接客発話評価値表示画面を前記表示部に表示させる、
     接客モニタリングシステム。
    The customer service monitoring system according to claim 1,
    The first storage unit further stores authority data including authority information for browsing operation of the customer service utterance evaluation value display screen in the display unit of the customer service utterance evaluation value calculated by the calculator for each customer service event,
    The output unit evaluates the service utterance when the authority information of the requester of browsing on the service utterance evaluation value display screen satisfies the authority information of the browsing operation included in the authority data stored in the first storage unit. Display a value display screen on the display unit;
    Service monitoring system.
  12.  請求項11に記載の接客モニタリングシステムであって、
     前記権限データは、前記接客発話評価値表示画面に対する前記接客発話評価値の修正操作の権限情報を含み、
     前記出力部は、前記接客発話評価値の修正要求者の権限情報が前記第1記憶部に記憶される前記権限データに含まれる前記修正操作の権限情報を満たす場合に、前記接客発話評価値の修正操作に応じて、前記接客発話評価値表示画面の前記接客発話評価値を更新する、
     接客モニタリングシステム。
    The customer service monitoring system according to claim 11,
    The authority data includes authority information for correcting the service utterance evaluation value for the service utterance evaluation value display screen,
    The output unit, when the authority information of the requester for correcting the customer service utterance evaluation value satisfies the authority information of the correction operation included in the authority data stored in the first storage unit, In response to a correction operation, the customer service utterance evaluation value on the customer service utterance evaluation value display screen is updated.
    Service monitoring system.
  13.  請求項12に記載の接客モニタリングシステムであって、
     前記所定の収音領域の既定位置を撮像する撮像部、を更に備え、
     前記第2記憶部は、前記撮像部により撮像された前記所定の収音領域の既定位置における映像データを更に記憶し、
     前記出力部は、前記第2記憶部に記憶される前記所定の収音領域の既定位置における映像データを出力している間に、前記検出部により検出された前記従業員の接客イベントの接客位置の修正操作に応じて、前記接客発話評価値表示画面の前記接客発話評価値を更新する、
     接客モニタリングシステム。
    The customer service monitoring system according to claim 12,
    An imaging unit that images a predetermined position of the predetermined sound collection area;
    The second storage unit further stores video data at a predetermined position of the predetermined sound collection area captured by the imaging unit,
    The output unit is a customer service event detected by the detection unit while the video data at the predetermined position of the predetermined sound collection area stored in the second storage unit is being output. Updating the customer service utterance evaluation value on the customer service utterance evaluation value display screen according to the correction operation of
    Service monitoring system.
  14.  請求項11に記載の接客モニタリングシステムであって、
     前記出力部は、所定の入力操作に応じて、前記接客発話評価値表示画面の前記接客イベント毎の前記接客発話評価値を所定項目毎に対比的に前記表示部に表示させる、
     接客モニタリングシステム。
    The customer service monitoring system according to claim 11,
    The output unit causes the display unit to display the service utterance evaluation value for each service event on the service utterance evaluation value display screen for each predetermined item in response to a predetermined input operation.
    Service monitoring system.
  15.  請求項1に記載の接客モニタリングシステムであって、
     前記第2記憶部に記憶される前記従業員の音声データを抽出する音声データ抽出部を更に備え、
     前記第1記憶部には、前記接客イベント毎に前記従業員が発話すると想定される発話想定区間に関する情報が記憶され、
     前記音声データ抽出部は、前記発話想定区間に関する情報を用いて、前記検出部により検出された前記接客イベントに応じた前記発話想定区間に対応する前記従業員の音声データを抽出する、
     接客モニタリングシステム。
    The customer service monitoring system according to claim 1,
    A voice data extraction unit for extracting the employee's voice data stored in the second storage unit;
    The first storage unit stores information related to an utterance assumption section in which the employee is assumed to utter for each customer service event,
    The voice data extraction unit extracts voice data of the employee corresponding to the speech utterance section corresponding to the customer service event detected by the detection unit, using information on the speech utterance section.
    Service monitoring system.
  16.  所定の収音領域における従業員の音声を収音する収音部を含む接客モニタリングシステムにおける接客モニタリング方法であって、
     所定の接客イベント毎の判定条件を含む接客イベントデータを第1記憶部に記憶し、
     所定の業務端末に対する従業員の操作履歴を示す端末操作履歴データと、前記収音部により収音された前記従業員の音声データとを対応付けて第2記憶部に記憶し、
     前記第1記憶部に記憶される前記接客イベントデータと前記第2記憶部に記憶される前記端末操作履歴データとを基に、前記従業員の前記接客イベントを検出し、
     検出された前記接客イベントにおいて、前記第2記憶部に記憶される前記従業員の音声データを基に、前記業務端末の操作時における所定の発話キーワードに対応する接客発話評価値を算定し、
     算定された前記接客発話評価値を、前記従業員の識別情報と前記従業員の接客位置及び接客時刻により特定される前記従業員の音声データに対応付けて記憶する、
     接客モニタリング方法。
    A customer service monitoring method in a customer service monitoring system including a sound collection unit that collects an employee's voice in a predetermined sound collection region,
    Storing customer service event data including determination conditions for each predetermined customer service event in the first storage unit;
    Terminal operation history data indicating an employee's operation history with respect to a predetermined business terminal and voice data of the employee collected by the sound collection unit are associated with each other and stored in the second storage unit;
    Detecting the customer service event of the employee based on the customer service event data stored in the first storage unit and the terminal operation history data stored in the second storage unit;
    In the detected customer service event, based on the employee's voice data stored in the second storage unit, calculate a customer service utterance evaluation value corresponding to a predetermined utterance keyword when operating the business terminal,
    The calculated customer service utterance evaluation value is stored in association with the employee's voice data specified by the employee identification information, the customer service position and the customer service time,
    Customer service monitoring method.
PCT/JP2015/004661 2014-09-30 2015-09-14 Service monitoring system and service monitoring method WO2016051693A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/513,622 US10706448B2 (en) 2014-09-30 2015-09-14 Service monitoring system and service monitoring method

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP2014201881 2014-09-30
JP2014-201881 2014-09-30
JP2015171556A JP6624368B2 (en) 2014-09-30 2015-08-31 Customer service monitoring system and customer service monitoring method
JP2015-171556 2015-08-31

Publications (1)

Publication Number Publication Date
WO2016051693A1 true WO2016051693A1 (en) 2016-04-07

Family

ID=55629770

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2015/004661 WO2016051693A1 (en) 2014-09-30 2015-09-14 Service monitoring system and service monitoring method

Country Status (1)

Country Link
WO (1) WO2016051693A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106960290A (en) * 2017-04-11 2017-07-18 西华大学 A kind of 4 S auto shop team sales service quality evaluation system and evaluation method
JP6300130B1 (en) * 2017-01-24 2018-03-28 パナソニックIpマネジメント株式会社 Service situation analysis device and service situation analysis system
JP6300131B1 (en) * 2017-01-24 2018-03-28 パナソニックIpマネジメント株式会社 Service situation analysis device and service situation analysis system
CN112562550A (en) * 2020-11-12 2021-03-26 国网山东省电力公司潍坊供电公司 Electronic table board
US11144955B2 (en) * 2016-01-25 2021-10-12 Sony Group Corporation Communication system and communication control method
CN114999103A (en) * 2022-05-12 2022-09-02 刘帅 Intelligent early warning system and method for highway road-related operation safety

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0737170A (en) * 1993-07-22 1995-02-07 Io Planning:Kk Sales management system provided with utterance recognition function
JPH07129848A (en) * 1993-11-02 1995-05-19 Casio Comput Co Ltd Sales data processor with voice recording function
JP2004192545A (en) * 2002-12-13 2004-07-08 Toshiba Tec Corp Customer service management device
JP2006267465A (en) * 2005-03-23 2006-10-05 Tokyo Electric Power Co Inc:The Uttering condition evaluating device, uttering condition evaluating program, and program storage medium
JP2009012329A (en) * 2007-07-05 2009-01-22 Canon Inc Manufacturing method of liquid discharge device and liquid discharge device
JP2011086087A (en) * 2009-10-15 2011-04-28 Seiko Epson Corp Information processing apparatus, control method for the same, and program
JP2011237966A (en) * 2010-05-10 2011-11-24 Seiko Epson Corp Customer service support device, customer service support method and program

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0737170A (en) * 1993-07-22 1995-02-07 Io Planning:Kk Sales management system provided with utterance recognition function
JPH07129848A (en) * 1993-11-02 1995-05-19 Casio Comput Co Ltd Sales data processor with voice recording function
JP2004192545A (en) * 2002-12-13 2004-07-08 Toshiba Tec Corp Customer service management device
JP2006267465A (en) * 2005-03-23 2006-10-05 Tokyo Electric Power Co Inc:The Uttering condition evaluating device, uttering condition evaluating program, and program storage medium
JP2009012329A (en) * 2007-07-05 2009-01-22 Canon Inc Manufacturing method of liquid discharge device and liquid discharge device
JP2011086087A (en) * 2009-10-15 2011-04-28 Seiko Epson Corp Information processing apparatus, control method for the same, and program
JP2011237966A (en) * 2010-05-10 2011-11-24 Seiko Epson Corp Customer service support device, customer service support method and program

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11144955B2 (en) * 2016-01-25 2021-10-12 Sony Group Corporation Communication system and communication control method
US20210406956A1 (en) * 2016-01-25 2021-12-30 Sony Group Corporation Communication system and communication control method
JP6300130B1 (en) * 2017-01-24 2018-03-28 パナソニックIpマネジメント株式会社 Service situation analysis device and service situation analysis system
JP6300131B1 (en) * 2017-01-24 2018-03-28 パナソニックIpマネジメント株式会社 Service situation analysis device and service situation analysis system
JP2018120345A (en) * 2017-01-24 2018-08-02 パナソニックIpマネジメント株式会社 Customer service status analyzing apparatus and customer service status analyzing system
JP2018120344A (en) * 2017-01-24 2018-08-02 パナソニックIpマネジメント株式会社 Customer service status analyzing apparatus and customer service status analyzing system
CN106960290A (en) * 2017-04-11 2017-07-18 西华大学 A kind of 4 S auto shop team sales service quality evaluation system and evaluation method
CN106960290B (en) * 2017-04-11 2023-12-22 西华大学 System and method for evaluating sales service quality of automobile 4S shop team
CN112562550A (en) * 2020-11-12 2021-03-26 国网山东省电力公司潍坊供电公司 Electronic table board
CN114999103A (en) * 2022-05-12 2022-09-02 刘帅 Intelligent early warning system and method for highway road-related operation safety
CN114999103B (en) * 2022-05-12 2023-07-25 刘帅 Intelligent early warning system and method for expressway road-related operation safety

Similar Documents

Publication Publication Date Title
JP6624368B2 (en) Customer service monitoring system and customer service monitoring method
WO2016051693A1 (en) Service monitoring system and service monitoring method
US11763280B2 (en) Transitioning of devices from their primary function to providing security system functionality
US9734831B2 (en) Utilizing voice biometrics
US20230351397A1 (en) Systems for detecting biometric response to attempts at coercion
US20160112569A1 (en) Utilizing Voice Biometrics
US9070233B2 (en) Automated banking machine system and monitoring
JP5950484B2 (en) Crime prevention system
JP5962916B2 (en) Video surveillance system
WO2019071903A1 (en) Auxiliary method, device and storage medium for micro-expression face examination
US20140379525A1 (en) Utilizing voice biometrics
JP5939493B1 (en) Service evaluation device, service evaluation system and service evaluation method provided with the same
KR101983558B1 (en) Method and apparatus for claiming insurance benefit
US20180174146A1 (en) Situational access override
JP2016021184A (en) Face identification system and program
US11615177B2 (en) Information processing system, information processing device, control method, and storage medium
JP5260233B2 (en) TRACKING DEVICE, TRACKING METHOD, AND TRACKING PROGRAM
JP6610992B2 (en) Service attitude evaluation system and service attitude evaluation method
Carter Confirm not command: examining fraudsters’ use of language to compel victim compliance in their own exploitation
JP2016165157A (en) Video monitoring system
JP2016045555A (en) Customer service management system, customer terminal, server device, customer service management method, and customer service method
US20190005500A1 (en) Automatic payment determination
KR102431697B1 (en) Server performing real estate risk management using cltv and operating method thereof
US11093943B1 (en) Account security system
KR20230085281A (en) Image analysis system and method based on image and lidar sensor

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15846727

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 15513622

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 15846727

Country of ref document: EP

Kind code of ref document: A1