WO2019097804A1 - Image recording device and image recording method - Google Patents

Image recording device and image recording method Download PDF

Info

Publication number
WO2019097804A1
WO2019097804A1 PCT/JP2018/031838 JP2018031838W WO2019097804A1 WO 2019097804 A1 WO2019097804 A1 WO 2019097804A1 JP 2018031838 W JP2018031838 W JP 2018031838W WO 2019097804 A1 WO2019097804 A1 WO 2019097804A1
Authority
WO
WIPO (PCT)
Prior art keywords
phase
unit
recording
procedure
image
Prior art date
Application number
PCT/JP2018/031838
Other languages
French (fr)
Japanese (ja)
Inventor
裕亮 櫻田
Original Assignee
オリンパス株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by オリンパス株式会社 filed Critical オリンパス株式会社
Publication of WO2019097804A1 publication Critical patent/WO2019097804A1/en

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/04Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
    • A61B1/045Control thereof
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H10/00ICT specially adapted for the handling or processing of patient-related medical or healthcare data
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast

Definitions

  • the present invention relates to an image recording apparatus and an image recording method for recording an image obtained by a medical device such as an endoscope.
  • endoscopes are widely adopted in the medical field and the like. Medical images obtained by the endoscope are recorded in various media for diagnosis and case recording. In recent years, with the increase in capacity of recording media, recording of moving images from endoscopes has also been performed.
  • various images such as an endoscopic image, an ultrasound image, an X-ray image during the procedure or examination, an image of an operator's hand, and an image of an indoor condition Images (hereinafter referred to as medical images) may be recorded as moving images.
  • image recording apparatuses not only the image recording apparatus main body but also those which can perform a recording operation by a scope switch or the like provided in the endoscope.
  • a diagnostic support system is also developed that records not only the image recording but also the sound related to the examination object to be photographed together with the information of the recording time.
  • a medical image as a backup for an evidence image or the like, or to use it as an educational material.
  • recorded images can be shared at academic meetings or in-hospital conferences and used for education of young doctors.
  • the technology certification system it is also possible to record the procedure of the endoscope and the like, and perform the procedure recognition from the recorded image.
  • An object of the present invention is to provide an image recording apparatus and an image recording method capable of adding index information at a desired timing by estimating a procedure phase based on a conversation such as a doctor or a nurse during a procedure. Do.
  • An image recording apparatus includes a video input unit for acquiring a medical image, an audio acquisition unit for acquiring audio generated by a medical worker, and an audio analysis unit for analyzing audio acquired by the audio acquisition unit.
  • a procedure phase estimation unit that estimates a procedure phase based on words and phrases analyzed by the voice analysis unit; a change of the phase estimated by the procedure phase estimation unit; and a time series of phases set in advance Based on the comparison of the first phase, index data for correlating the relationship between the medical image and the phase to the medical image according to the judgment result of the judgment of the estimated phase and the judgment result of the judgment portion, and recording data
  • a recording data generation unit for recording the information in the recording unit.
  • An image recording method includes a video input step of acquiring a medical image, an audio acquisition step of acquiring audio generated by a medical worker, and an audio analysis step of analyzing audio acquired in the audio acquisition step.
  • a procedure phase estimation step of estimating a procedure phase based on words and phrases obtained by analysis in the voice analysis step, a change of the phase estimated in the procedure phase estimation step, and a time series of predetermined phases Based on the comparison of the second phase, index information for correlating the relationship between the medical image and the phase according to the determination step of determining the estimated phase, and the determination result in the determination step, Recording data generation step of generating the .
  • FIG. 1 is a block diagram showing an image recording apparatus according to a first embodiment of the present invention.
  • Explanatory drawing which shows the mode of the operating room in which the image recording apparatus of FIG. 1 is arrange
  • FIG. 6 is a flowchart for explaining the operation of the first embodiment.
  • FIG. 6 is an explanatory view showing an example of an image file generated by a recording unit 55.
  • FIG. 6 is an explanatory view showing an example of an image file generated by a recording unit 55.
  • the flowchart which shows the operation
  • FIG. 1 is a block diagram showing an image recording apparatus according to a first embodiment of the present invention.
  • FIG. 2 is an explanatory view showing the operating room where the image recording apparatus of FIG. 1 is disposed.
  • the medical system 3 disposed in the operating room 2 is provided with a system controller 41 that controls medical equipment such as the operating table 10 on which the patient 48 lies and the electric scalpel device 13.
  • a first cart 11 and a second cart 12 are provided in the operating room 2, and a system controller 41 is placed on the first cart 11.
  • the first cart 11 devices such as an electric scalpel device 13, an insufflation device 14, a video processor 15, and a light source device 16 as medical devices which are controlled devices, and a gas cylinder 18 filled with carbon dioxide are mounted. It is placed.
  • the video processor 15 is connected to the first endoscope 31 via a camera cable 31a.
  • the light source device 16 is connected to the first endoscope 31 via the light guide cable 31 b. Further, on the first cart 11, the display device 19, the first central display panel 20, the operation panel 49 and the like are placed.
  • the display device 19 is, for example, a TV monitor that displays an endoscope image or the like from the video processor 15, for example.
  • the central display panel 20 is a display means capable of selectively displaying any data during surgery.
  • the operation panel 49 includes, for example, a display screen such as a liquid crystal display and a touch sensor integrally provided on the display screen, and is a centralized operation device operated by a nurse or the like who is in a non-sterile area. .
  • the operating table 10, the shadowless lamp 6, the electric scalpel device 13, the insufflation device 14, the video processor 15, and the light source device 16 are connected to the system controller 41 which is a central control device via a communication line (not shown) There is.
  • the first cart 11 can read / write the individual ID information of an object wirelessly by an ID tag embedded in the first endoscope 31 or the treatment tool of the electric scalpel device 13 or the like.
  • a Radio Frequency Identification) terminal 35 is provided.
  • a video processor 23 which is a controlled device, a light source device 24, an image processing device 25, a display device 26, a second centralized display panel 27, and an image recording device 50 are placed.
  • the video processor 23 is connected to the second endoscope 32 via a camera cable 32a.
  • the light source device 24 is connected to the second endoscope 32 via the light guide cable 32b.
  • the display device 26 displays an endoscopic image or the like captured by the video processor 23.
  • the second central display panel 27 can selectively display any data during the operation.
  • the video processor 23, the light source device 24, the image processing device 25 and the image recording device 50 are connected to the relay unit 28 placed on the second cart 12 via a communication line (not shown).
  • the relay unit 28 is connected to the system controller 41 mounted on the first cart 11 by a relay cable 29.
  • the system controller 41 includes the video processor 23 mounted on the second cart 12, the light source device 24, the image processing device 25 and the image recording device 50, and the electric scalpel device mounted on the first cart 11. 13, the insufflation device 14, the video processor 15, the light source device 16, and the operating table 10 can be centrally controlled.
  • the system controller 41 displays setting states of connected devices, setting screens such as operation switches, etc. on the display screen of the operation panel 49. It can be done.
  • the system controller 41 can perform operation input such as change of setting values by touching a desired operation switch and operating the touch panel in a predetermined area.
  • the remote controller 30 is a second centralized control device operated by a surgeon or the like who is in a sterile area, and can operate other devices with established communication via the system controller 41. .
  • an infrared communication port (not shown) which is a communication means is attached to the system controller 41.
  • the infrared communication port is provided at a position where infrared light can be easily emitted, such as in the vicinity of the display device 19, and the system controller 41 is connected with a cable.
  • the system controller 41 is connected to the patient monitoring system 4 by the cable 9.
  • the patient monitoring system 4 can analyze biological information and can display the analysis result on a required display device.
  • a camera 37 for imaging a medical device such as the operating table 10 is also provided. It is possible to determine the operation state by imaging a medical device such as the operating table 10 with the camera 37 and analyzing the captured image. The determination result and the image captured by the camera 37 are supplied to the system controller 41.
  • the video processors 15 and 23 can generate endoscopic images based on the outputs of the endoscopes 31 and 32, respectively.
  • the endoscopic images from the video processors 15 and 23 are supplied to the image recording device 50.
  • the video processors 15 and 23 and the image recording device 50 are connected via a network (not shown), and various information including inspection information is supplied from the video processors 15 and 23 to the image recording device 50. (Not shown in Figure 1).
  • a network for connecting the video processors 15 and 23 and the image recording apparatus 50 communication lines of various communication standards can be adopted.
  • a headset type microphone 33 can be connected to the image recording apparatus 50.
  • the microphone 33 picks up the voice emitted by the wearer and outputs a voice signal to the image recording device 50.
  • a plurality of microphones 33 can be connected to the image recording device 50, and the image recording device 50 acquires audio from the plurality of microphones 33.
  • the microphone 33 is shown as being wired connected by a cable, transmitting an audio signal to the image recording apparatus 50 via a wireless transmission path such as Wi-Fi (registered trademark) or Bluetooth (registered trademark). You may be able to
  • a microphone 34 is also provided, supported by a support member (not shown).
  • the microphone 34 is connected to the image recording apparatus 50 via a cable (not shown) or a wireless transmission path such as Wi-Fi or Bluetooth (registered trademark), and picks up the sound in the operating room 2 to image the sound signal.
  • the recording apparatus 50 can be supplied.
  • FIG. 1 shows an example of a specific configuration of the image recording apparatus 50 in FIG.
  • the image recording device 50 is provided with a control unit 51.
  • the control unit 51 can control each part of the image recording apparatus 50.
  • the control unit 51 may be configured by a processor such as a CPU (not shown) and operated according to a program stored in a memory (not shown) to control each unit, and may be configured by a field programmable gate array (FPGA) or the like. It may be done.
  • a processor such as a CPU (not shown) and operated according to a program stored in a memory (not shown) to control each unit, and may be configured by a field programmable gate array (FPGA) or the like. It may be done.
  • FPGA field programmable gate array
  • the video input unit 52 of the image recording device 50 is an interface suitable for image transmission, and takes in various medical images.
  • the video input unit 52 can adopt various terminals such as a DVI (Digital Visual Interface) terminal, an SDI (Serial Digital Interface) terminal, an RGB terminal, a Y / C terminal, and a VIDEO terminal.
  • the image input unit 52 is, for example, an endoscopic image from the video processor 15 or 23, an ultrasound apparatus, an operation field camera, an X-ray observation apparatus, an endoscope processor (not shown) other than the video processor 15 or 23 Can capture various medical images from
  • the medical image captured by the image input unit 52 is given to the record data generation unit 54.
  • the recording data generation unit 54 converts the input medical image into a video signal of a predetermined image format by performing a predetermined encoding process on the input medical image.
  • the recording data generation unit 54 may input the medical image into the MPEG2 format or the MPEG-4 AVC / H. It is possible to convert it into a video signal of the H.264 format or the like and output it as recording data.
  • the recording data generation unit 54 is controlled by the control unit 51 and can add index information to the recording data.
  • the recording data generation unit 54 may include index information as meta information in an image file which is recording data, or may generate a file including index information different from the image file. Two types of index information may be recorded.
  • the procedure includes phases such as anesthesia, ablation, excision, dissection, hemostasis, and bypass phase.
  • the index information corresponds to the time information of the medical image included in the recording data, and is information corresponding to the recording time of each stage (phase) of the procedure among the scenes of the medical image as described later. By using index information at the time of reproduction, it is possible to make the reproduction position jump to a predetermined position of each phase in the medical image, for example, the head position.
  • the recording data generation unit 54 is configured to output the generated recording data to the recording unit 55.
  • the recording unit 55 is controlled by the control unit 51 to record the record data of the medical image as an image file.
  • a hard disk drive or the like can be employed as the recording unit 55.
  • the recording unit 55 can be controlled by the control unit 51 to read out the recording data and output it to the external media recording and reproducing unit 56.
  • the external media recording and reproducing unit 56 can apply recording data from the recording unit 55 to an external medium 65, which is an external recording medium (not shown), and can record the data.
  • an external medium 65 not only BD (Blu-ray Disc), DVD, USB, but also a server or the like on a network may be adopted, and other recording media may be adopted.
  • the record data generation unit 54 can also output the input medical image to the video output unit 57.
  • the video output unit 57 outputs the input medical image to the external monitor 66. Thereby, the external monitor 66 can display a medical image.
  • the image recording apparatus 50 is also provided with a sound acquisition unit 53.
  • the voice acquisition unit 53 receives voice signals from the microphones 33 and 34.
  • the audio acquisition unit 53 is configured by a connector or the like to which a cable is connected when the transmission path of the audio signal from the microphones 33 and 34 is wired, and the transmission path of the audio signal from the microphones 33 and 34 is wireless. Is constituted by an antenna, a receiver or the like for receiving a radio signal, and takes in the input voice signal and outputs it to the voice analysis unit 58.
  • the sound acquisition unit 53 can acquire a plurality of sound signals.
  • the voice acquisition unit 53 can detect which microphone each voice signal is picked up by. For example, when transmission of an audio signal is performed by wire, the audio acquisition unit 53 can detect which microphone the audio signal is picked up by which connector is connected by the connector connected to each cable. In the case where signal transmission is performed wirelessly, it is possible to detect which microphone an audio signal is picked up by, for example, a device ID acquired at the time of establishment of wireless communication.
  • the voice analysis unit 58 performs analysis processing on the input voice to obtain a voice analysis result.
  • the speech analysis unit 58 performs speech recognition using an acoustic model, a word dictionary, a language model and the like prepared in advance in the system.
  • the speech analysis unit 58 analyzes the speech contents of the doctor or nurse collected by the microphones 33 and 34 by speech recognition processing, acquires the speech recognition result of the word or speech spoken by the doctor or nurse, and analyzes the result Output to the determination unit 59 as
  • the determination unit 59 determines the phase of the procedure corresponding to the voice collected by the microphones 33 and 34 by referring to the lookup table stored in the storage unit 60.
  • the determination unit 59 may be configured by a processor using a CPU or the like, and may operate in accordance with a program stored in a memory (not shown), or a part or all of the functions of the electronic circuit of hardware. It may be realized.
  • FIG. 3 is an explanatory diagram for explaining an example of the contents of the lookup table stored in the storage unit 60.
  • the look-up table of FIG. 3 registers the relationship between each phase (Phase 1, Phase 2,..., Phase n) of the procedure and the corresponding word.
  • the phrases “start” and “start” are registered corresponding to Phase 1
  • the phrases “separation complete”, “separation complete” and “end complete” are registered corresponding to Phase 5 It is done.
  • FIG. 3 shows an example in which each phase is associated with one or a relatively small number of words or phrases
  • each phase (Phase1, Phase2, ...) of FIG. 3 generate
  • FIG. 3 corresponds to a predetermined procedure, and if the procedure is different, each phase corresponding to the procedure and a lookup table in which words and phrases corresponding to each phase are registered are adopted.
  • the determination unit 59 includes a procedure phase estimation unit 59a.
  • the procedure phase estimation unit 59a estimates the phase corresponding to the content of the phrase obtained from the collected voice by referring to the lookup table using the phrase obtained by the analysis result of the voice analysis unit 58, and the estimation result Are output to the control unit 51.
  • the procedure phase estimation unit 59a is acquired by the microphones 33 and 34 by referring to the look-up table stored in the storage unit 60 when “peeling” is obtained as the analysis result of the voice. It is estimated that the procedure shifts to the phase (Phase 7) at the timing of the voice.
  • FIG. 1 shows an example in which the procedure phase estimation unit 59a is provided in the determination unit 59
  • the procedure phase estimation unit 59a may be provided separately from the determination unit 59.
  • the image recording device 50 is provided with an operation unit 61.
  • the operation unit 61 includes, for example, switches, buttons, keys, and a touch panel (not shown), and is configured to receive a user operation and output an operation signal based on the user operation to the control unit 51.
  • the control unit 51 is configured to control each unit based on a user operation on the operation unit 61. For example, the control unit 51 can edit the look-up table stored in the storage unit 60 based on a user operation on the operation unit 61.
  • control unit 51 can also provide various types of information about the patient to the recording data generation unit 54 and record them as metadata of an image file to be recorded It is supposed to be.
  • the control unit 51 may be configured to take in various information related to the patient from an external database server (not shown) via a communication circuit (not shown).
  • control unit 51 may read the look-up table stored in the external medium 65 via the external medium recording and reproducing unit 56 and store the lookup table in the storage unit 60.
  • the determination unit 59 determines whether or not the estimation result of the phase based on the speech recognition indicates a phase corresponding to the actual order of the phases in consideration of the order of each phase arranged in time series. If so, it is determined that the estimation result of the phase based on speech recognition is correct, otherwise it is determined that the estimation result of the phase based on speech recognition is incorrect. That is, the determination unit 59 determines the estimated phase based on the comparison between the change of the phase, which is the estimation result of the procedure phase, and the time series of the phase set in advance.
  • the determination unit 59 determines that the current phase is correct in the phase (Phase 5) and then the estimation result of the phase (Phase 4) is obtained based on speech recognition, the estimation result of the phase (Phase 4) If the estimation result of the phase (Phase 6) is obtained based on speech recognition, it is determined that the estimation result of the phase (Phase 6) is correct.
  • the determination unit 59 stores the determination result including the determination of the phase transition based on speech recognition and the information on the determined phase only when the estimation result of the phase based on speech recognition is correct in time series. , And output to the control unit 51.
  • the determination unit 59 may store the determination result in a memory (not shown) in the determination unit 59.
  • control unit 51 when control unit 51 receives the determination result of the phase from determination unit 59, control unit 51 adds index information corresponding to the determination result to the recording data at the timing when the determination result is input. And generates a recording request to the recording data generation unit 54. Thereby, the recording data generation unit 54 performs medical treatment on the index information associated with the timing when the phase of the procedure shifts to the next phase, that is, the recording time (reproduction time) at which the scene of the medical image to be recorded shifts to the next phase. Add to recorded data of image.
  • the index information is the value of the time code of the timing to shift to each phase. It may be information including In addition, the index information may include not only information on recording time but also information on the type of the determined phase.
  • the recording data generation unit 54 may generate the file as another file independent of the image file for recording the index information. For example, each time the recording data generation unit 54 determines a phase based on voice recognition, even if it generates a file in which information of determination time (recording time, reproduction time) and text information of voice recognition results are listed. Good.
  • the control unit 51 is configured to receive, from the voice acquisition unit 53, information indicating which microphone the voice subjected to voice analysis in the voice analysis unit 58 is picked up by.
  • the control unit 51 refers to the table using the information from the voice acquisition unit 53, It is possible to specify a speaker such as a word and the like used in the phase determination, and control the recording data generation unit 54 so that the information of the speaker is included in the index information.
  • FIG. 4 is a flowchart for explaining the operation of the first embodiment.
  • the video input unit 52 takes in a medical image from the video processor 15 and outputs the medical image to the recording data generation unit 54.
  • the record data generation unit 54 is controlled by the control unit 51, converts the medical image into a moving image of a predetermined format, and starts recording in the recording unit 55 in step S1 of FIG. 4. Further, the medical image from the record data generation unit 54 is displayed on the external monitor 66 by the video output unit 57.
  • the voice acquisition unit 53 takes in the voice collected by the plurality of microphones 33 and 34 and outputs the voice to the voice analysis unit 58.
  • the voice analysis unit 58 acquires the words (conversation) generated by the doctor or the nurse by a known voice recognition process, and outputs the voice recognition result to the determination unit 59.
  • the procedure phase estimation unit 59a of the determination unit 59 estimates the phase corresponding to the speech recognition result by referring to the lookup table of the storage unit 60 using the speech recognition result (step S3).
  • the voice analysis unit 58 performs voice recognition of the doctor's speech and outputs the recognition result to the determination unit 59.
  • the procedure phase estimation unit 59a of the determination unit 59 determines that the speech recognition result of the words "start” or “start” is present in the look-up table (LUT) of the storage unit 60
  • the phase corresponding to the word is An estimation result indicating that it is a phase (Phase 1) is acquired.
  • a word or phrase obtained by speech recognition does not exist in the look-up table (LUT) of the storage unit 60, the process proceeds from step S3 to step S7.
  • step S7 the control unit 51 determines whether or not there is an operation to end recording of a moving image which is a medical image. If there is no operation to end recording, the process returns to step S2, and if there is an operation to end recording. In step S8, the recording end process is performed.
  • determination unit 59 determines in step S3 that the speech recognition result is present in the lookup table, it is stored in memory 60 or a memory (not shown) in determination unit 59 in the next step S4.
  • the determination results of the previous phases are read out to determine whether or not the time series of the phase of the estimation result based on speech recognition is correct.
  • the phase (Phase 1) is the first phase after the start of the procedure, and the estimation result of this phase is correct in time series, so the determination unit 59 shifts the process from step S4 to step S5 and determines the estimation result, Information on the phase transition and the type of phase is output to the control unit 51. If the estimation result of the phase by speech recognition is not correct in time series, the determination unit 59 returns the process to step S2 without using the estimation result of this time.
  • step S6 After the control unit 51 determines in step S5, for example, on which microphone the output of the voice analysis unit 58 is due to the sound collection result, the control unit 51 identifies the speaker who issued the voice used for the phase determination, A recording request for adding index information to a moving image to be recorded is issued (step S6).
  • the recording data generation unit 54 adds index information to the recording data. That is, the index information has time information corresponding to the image portion recorded at the timing when the doctor utters "start” or "start".
  • the recording data to which the index information is added is recorded in the recording unit 55.
  • step S6 After making a recording request in step S6, the control unit 51 shifts the process to step S7 and determines whether or not a recording end operation has been performed. Thereafter, the processes of steps S2 to S7 are repeated, and the phase determination and the addition of the index information to the recording data are performed for each utterance of the doctor or the nurse.
  • the control unit 51 indexes according to this determination Information is added to the recorded data.
  • the doctor utters “start” after the addition of the index information accompanying the transition to the phase (phase 2) by the utterance “dissection start”.
  • the estimation result of the phase based on speech recognition by the procedure phase estimation unit 59a of the determination unit 59 indicates that the phase (Phase 2) returns to the phase (Phase 1), and the determination unit 59 determines that this estimation result is It determines that the time series is not correct, and ignores the estimation based on the "start” speech recognition.
  • the transition to each phase is correctly determined in the time series of the phases (Phase1, Phase2,%) Of FIG. 3, and the index information is added to the recording data at the timing according to the transition to each phase.
  • control unit 51 determines that the recording end operation has been performed in step S7, the control unit 51 instructs the recording data generation unit 54 to end the recording in step S8.
  • the recording data generation unit 54 converts the recording data into a file and records the file in the recording unit 55.
  • the external media recording and reproducing unit 56 can record the image file recorded in the recording unit 55 in the external medium 65.
  • the control unit 51 determines that the recording end operation is performed, and performs the end of recording and the conversion into an image file. It may be
  • FIG. 5 and 6 are explanatory diagrams showing examples of index information added to recording data.
  • FIG. 5 shows an example of recording index information as metadata of recording data.
  • the example of FIG. 5 shows index information added as metadata for recorded data of a predetermined procedure for a certain patient.
  • the index information includes patient information including “Date”, “Patient ID”, “Patient Name”, “BOD”, “Age”, and “Sex”.
  • the index information also includes information of the file name of the image file indicated by ⁇ title 1>.
  • the index information includes information obtained by the determination of the phase, that is, the time of the phase determination, the speaker of the phrase used for the determination, and the information of the phrase used for the phase determination.
  • the phase related to peeling has been started at time xx: xx, xx and xx according to the phrase “peeling start” uttered by the speaker OO, and the recording time of the image related to the peeling It indicates that index information indicating a time xx hour xx minute xx second) is recorded as metadata of recording data.
  • the information on the recording time indicates the time from the start of recording, and is the same information as the reproduction time.
  • the reproduction system is not shown in FIG. 1, it is also possible to read out the image file recorded in the recording unit 55 and output the reproduced image to the external monitor 66 by the video output unit 57.
  • the reproduction circuit (not shown) can jump the reproduction position and the editing position of the recorded image to the position corresponding to the index information by using the index information when reproducing the recording data. This makes it easy to move to the top position of each phase to perform reproduction.
  • the index information is used to move to the head position of each phase and play back. It can be carried out.
  • FIG. 6 shows index information in the form of a list associated with an image file in consideration of a playback device that does not have such a jump function using metadata.
  • the list of FIG. 6 has an item of playback time and an item of corresponding text. That is, in the item of reproduction time, the recording time (reproduction time) in which the phase is determined is indicated, and in the item of text, the term used for the determination of the phase is indicated. For example, in the example of FIG.
  • the first phase of the procedure xx is started after time 00:03:22 from the start of recording by the phrase "Start xx for the subject OO from this" It is determined that the index information indicating the recording time (time xx: xx: xx) of the image regarding the phase is recorded as a file different from the image file of the recording data.
  • index information in a list format as shown in FIG. 6, for example, when reproducing an image file by a personal computer or the like, it is possible to easily confirm the head position of each phase.
  • the transition to each phase of the procedure is determined by analyzing the voice emitted by the operator or the like, and index information can be reliably added to the image file to be recorded.
  • the doctor and the nurse generally exchange conversations about the procedure in order to cooperate with each other, and the phase of the procedure can be estimated by speech analysis based on the conversation. That is, it is possible to determine the switching of the phase even when there is no status change or video change of the device. Therefore, in order to add the index information, the operator does not need to perform a special operation or the like, and the index information can be added to a desired position without preventing the operator from focusing on the procedure. By using this index information, it is possible to easily move the reproduction position and the editing position to each phase at the time of reproduction, editing and the like.
  • FIG. 7 is a flowchart showing an operation flow employed in the second embodiment of the present invention.
  • the hardware configuration in the present embodiment is the same as that shown in FIG.
  • the lookup table corresponding to each procedure is adopted.
  • a plurality of lookup tables corresponding to each procedure are stored in the storage unit 60, and the lookup table can be selected according to the procedure.
  • a look-up table not only for the procedure but also for the operator is prepared, and each phase is estimated and determined using the lookup table for each procedure and each operator. .
  • the sound acquisition unit 53 can acquire sound for each microphone, that is, for each operator.
  • the voice analysis unit 58 can obtain a voice analysis result for each microphone by inputting a voice signal for each microphone from the voice acquisition unit 53.
  • the determination unit 59 determines the phase used in the determination of the phase based on the information from the control unit 51. The speaker can be identified.
  • the voice analysis unit 58 can improve voice recognition accuracy by using a personal dictionary for each user at the time of voice recognition processing, and specify the user who uttered by the voice recognition processing without specifying a microphone. Is also possible.
  • the storage unit 60 stores a lookup table for each procedure and each operator as a lookup table.
  • the storage unit 60 may store only one look-up table for each procedure or each operator.
  • the determination unit 59 identifies the procedure based on the voice recognition result, and identifies the operator based on the information from the control unit 51 or the information acquired in the process of the voice recognition process.
  • the determination unit 59 determines the phase using a look-up table for each procedure and each operator. Thereby, the estimation and determination accuracy of the phase can be improved.
  • the look-up table for each operator corresponds to the language and / or whistling of the operator. For example, when the operator speaks English, English is registered.
  • step S1 when the recording of the medical image is started, in step S21, the voice analysis unit 58 performs voice recognition, and the voice analysis unit 58 or the control unit 51 generates an operator's word that is the target of voice recognition. Identify
  • step S22 the determination unit 59 identifies the procedure based on the voice recognition result, and selects a lookup table corresponding to the identified procedure among the lookup tables stored in the storage unit 60.
  • the look-up table for each procedure and for each operator is stored in the storage unit 60, the look-up table group for each procedure and the specified procedure is selected.
  • FIG. 8 and 9 are explanatory diagrams showing an example of look-up tables of two procedures stored in the storage unit 60.
  • FIG. The look-up table (LUT 1) in FIG. 8 corresponds to the procedure of liver resection.
  • the phase (Phase 1) is generated by the operator's "start” or “start” speech. It is determined, and the phases “Phase 2", “Phase 3",... Are sequentially determined by the utterance "cut start”, “cut start”,.
  • the look-up table (LUT 2) in FIG. 9 corresponds to the procedure for bile duct stone removal, and by using this look-up table, the phase is started by the operator's etc. "start” or “start” utterance. (Phase 1) is determined, and the phases “Phase 2", “Phase 3",... Are sequentially determined by the utterance of "calculus confirmation", "stent insertion",.
  • the determination unit 59 selects the lookup table LUT1 in step S22.
  • the determination unit 59 may specify a procedure not only by specifying the procedure based on the voice recognition result, but also by specifying the procedure of the control unit 51 according to the operation of the operation unit 61, for example.
  • the determination unit 59 may specify the procedure based on the designation of the control unit 51 that has acquired the information.
  • the doctor or the like utters “start”, and the determination unit 59 specifies the doctor in step S23, and if there is a look-up table corresponding to the specified doctor, the look-up table A table is selected, and if it does not exist, a look-up table corresponding to “Hepatectomy” common to the operator is selected to perform phase determination.
  • the same effect as that of the first embodiment can be obtained, and the phase is determined with reference to the look-up table after the procedure or for each operator, thereby improving the determination accuracy. It has the effect of being able to
  • FIG. 10 is a flowchart showing an operation flow employed in the third embodiment of the present invention.
  • the same steps as in FIG. 4 will be assigned the same reference numerals and descriptions thereof will be omitted.
  • the hardware configuration in the present embodiment is the same as that shown in FIG.
  • the look-up table shows an example in which only the time-sequential phase of each procedure is registered.
  • the present embodiment shows an example in the case where not only the phase in chronological order but also the phase occurring at an arbitrary timing is registered as the lookup table.
  • FIG. 11 is an explanatory drawing showing an example of the content of the look-up table employed in the present embodiment.
  • the look-up table of FIG. 11 is stored in the storage unit 60, and, as in FIG. 3, registers the relationship between each phase of the procedure (Phase1, Phase2, etc And the corresponding word.
  • phases that do not necessarily occur in time-series are registered.
  • the hemorrhage phase, the hemostasis phase, and the cleansing phase are registered as an example of phases that do not necessarily occur in time series. In these phases, when the speech recognition result is "bleeding", "hemostasis” or "washing", it is determined that the phase has been entered.
  • the procedure phase estimation unit 59a of the determination unit 59 determines whether or not there is a phase corresponding to the speech recognition result in step S3 of FIG. 10, and if it exists, shifts the process to step S31.
  • determination unit 59 determines whether or not the phase based on the speech recognition result is a phase that occurs in time series. For example, when the phase occurring in the time series is estimated as in the phase (Phase 3) corresponding to “separation start”, the determination unit 59 determines whether the phase is the correct time series phase in the next step S 4 It is determined whether or not.
  • the determination unit 59 shifts the process from step S31 to step S5, and Speaker identification is performed without judging the sequence.
  • index information corresponding to the occurrence of the phase can be reliably added to the recording data even in the phase that may occur at an arbitrary timing such as “hemostasis”.
  • the same effect as that of the first embodiment can be obtained, and even in a phase that does not necessarily occur in time series, it is determined with certainty that index information is recorded data. There is an effect that it can be added.
  • the speech analysis unit 58 and the determination unit 59 perform the speech recognition and the phase determination.
  • these processes may be performed using artificial intelligence.
  • the collected voice is provided to an external server having an artificial intelligence function
  • the content of the conversation is determined by the artificial intelligence
  • the phase is determined
  • the determination result is taken in by the control unit
  • the index information is added to the recorded image. You may do so.
  • the present invention is not limited to the above-described embodiments as it is, and at the implementation stage, the constituent elements can be modified and embodied without departing from the scope of the invention.
  • various inventions can be formed by appropriate combinations of a plurality of components disclosed in the above-described embodiments. For example, some components of all the components shown in the embodiment may be deleted. Furthermore, components in different embodiments may be combined as appropriate.

Abstract

This image recording device is provided with: an image input unit for acquiring a medical image; a speech acquisition unit for acquiring speech produced by a health professional; a speech analysis unit for analyzing the speech acquired by the speech acquisition unit; a procedure phase estimation unit for estimating a phase of a procedure on the basis of a phrase from the analysis results of the speech analysis unit; a determination unit for confirming the estimated phase on the basis of a time series comparison of a change in the phase estimated by the procedure phase estimation unit and a phase set in advance; and a recording data generation unit which generates recording data by applying, to the medical image, index information associating the medical image and the phase relationship according to the determination results of the determination unit, and records the recording data in a recording unit.

Description

画像記録装置及び画像記録方法Image recording apparatus and image recording method
 本発明は、内視鏡等の医療装置によって得られる画像を記録する画像記録装置及び画像記録方法に関する。 The present invention relates to an image recording apparatus and an image recording method for recording an image obtained by a medical device such as an endoscope.
 従来、内視鏡は医療用分野等において広く採用されている。内視鏡によって得られる医療画像は、診断や症例の記録のために、種々のメディアに記録される。近年、記録媒体の大容量化に伴って、内視鏡からの動画像の記録も行われるようになってきた。 Conventionally, endoscopes are widely adopted in the medical field and the like. Medical images obtained by the endoscope are recorded in various media for diagnosis and case recording. In recent years, with the increase in capacity of recording media, recording of moving images from endoscopes has also been performed.
 例えば、内視鏡を用いた手技や検査では、手技や検査中における内視鏡画像、超音波画像、X線画像だけでなく、術者の手元の画像や室内の様子の画像等の種々の画像(以下、医療画像という)を動画記録することがある。このような画像記録装置においては、画像記録装置本体だけでなく、内視鏡に配設されたスコープスイッチ等によって記録操作が可能なものもある。 For example, in a procedure or examination using an endoscope, various images such as an endoscopic image, an ultrasound image, an X-ray image during the procedure or examination, an image of an operator's hand, and an image of an indoor condition Images (hereinafter referred to as medical images) may be recorded as moving images. Among such image recording apparatuses, not only the image recording apparatus main body but also those which can perform a recording operation by a scope switch or the like provided in the endoscope.
 更に、日本国特開2007-275237号公報においては、画像記録だけでなく、撮影される検査対象に関する音を録音時刻の情報と共に記録する診断支援システムも開発されている。 Furthermore, in Japanese Patent Application Laid-Open No. 2007-275237, a diagnostic support system is also developed that records not only the image recording but also the sound related to the examination object to be photographed together with the information of the recording time.
 ところで、症例を記録する目的として、医療画像を証拠画像等のバックアップ用として用いることや、教育用素材として用いることが考えられる。例えば、症例中の重要な解剖シーンについては、記録画像を学会や院内カンファレンスで共有して、若手ドクターらの教育に活用することができる。また、技術認定制度の利用のために、内視鏡の手技等を記録して、記録画像から手技の認定を行うこともできる。 By the way, as a purpose of recording a case, it is conceivable to use a medical image as a backup for an evidence image or the like, or to use it as an educational material. For example, for important anatomical scenes in cases, recorded images can be shared at academic meetings or in-hospital conferences and used for education of young doctors. In addition, in order to use the technology certification system, it is also possible to record the procedure of the endoscope and the like, and perform the procedure recognition from the recorded image.
 ところで、バックアップ用として記録する場合には症例全体を記録する必要があり、例えば、外科手術等においては、比較的長時間に渡って画像記録が行われる。これに対し、教育用途に用いる画像は、手技や検査期間のうちの一部の期間である場合が多い。このため、症例全体を記録して得た記録画像を教育用途に利用する場合には、所望のシーンの検索に手間がかかってしまうことがある。 By the way, in the case of recording for backup, it is necessary to record the entire case. For example, in a surgical operation etc., image recording is performed for a relatively long time. On the other hand, images used for educational applications are often part of the procedure or examination period. Therefore, when a recorded image obtained by recording the entire case is used for educational purposes, it may take time to search for a desired scene.
 そこで、医療画像の記録途中において、スコープスイッチ等によって編集ポイントを指定することで、記録時に編集ポイントの情報を付加して画像記録を行う方法が考えられる。しかしながら、所望のシーンの編集ポイントを付加する操作を行うことは、手術中等においては必ずしも容易ではなく、また、操作し忘れが生じることがある。 Therefore, there is a method of recording an image by adding information of the editing point at the time of recording by designating the editing point by a scope switch or the like during recording of the medical image. However, performing an operation of adding a desired scene editing point is not always easy during surgery or the like, and may cause an operation to be forgotten.
 そこで、画像の変化から手技のフェーズの切換りを推定し、推定した切換りのポイントに対応して記録画像に編集ポイントのメタ情報を付加することにより、所望のシーンの画像の検索を容易にする画像記録装置が提案されている。 Therefore, it is easy to search for an image of a desired scene by estimating the switching of the phase of the procedure from the change of the image and adding the meta information of the editing point to the recorded image corresponding to the estimated switching point. An image recording apparatus has been proposed.
 しかしながら、例えば教育用途用として必要なシーンの開始タイミングであっても、このタイミングにおいて画像に大きな変化がない場合には、編集ポイントとして判定されず、メタ情報が付加されないことがある。 However, even if it is the start timing of a scene necessary for educational use, for example, if there is no significant change in the image at this timing, it may not be determined as an editing point and meta information may not be added.
 本発明は、手技中におけるドクターやナース等の会話に基づいて手技フェーズを推定することにより、所望のタイミングでインデックス情報を付加することができる画像記録装置及び画像記録方法を提供することを目的とする。 An object of the present invention is to provide an image recording apparatus and an image recording method capable of adding index information at a desired timing by estimating a procedure phase based on a conversation such as a doctor or a nurse during a procedure. Do.
 本発明の一態様による画像記録装置は、医療画像を取得する映像入力部と、医療従事者の発生する音声を取得する音声取得部と、前記音声取得部が取得した音声を解析する音声解析部と、前記音声解析部の解析結果による語句に基づいて、手技のフェーズを推定する手技フェーズ推定部と、前記手技フェーズ推定部により推定されたフェーズの変化と予め設定されているフェーズの時系列との比較に基づいて、推定されたフェーズを確定する判定部と、前記判定部の判定結果に応じて、前記医療画像と前記フェーズとの関係を対応付けるインデックス情報を前記医療画像に付与して記録データを生成して記録部に記録する記録データ生成部とを具備する。 An image recording apparatus according to an aspect of the present invention includes a video input unit for acquiring a medical image, an audio acquisition unit for acquiring audio generated by a medical worker, and an audio analysis unit for analyzing audio acquired by the audio acquisition unit. A procedure phase estimation unit that estimates a procedure phase based on words and phrases analyzed by the voice analysis unit; a change of the phase estimated by the procedure phase estimation unit; and a time series of phases set in advance Based on the comparison of the first phase, index data for correlating the relationship between the medical image and the phase to the medical image according to the judgment result of the judgment of the estimated phase and the judgment result of the judgment portion, and recording data And a recording data generation unit for recording the information in the recording unit.
 本発明の一態様による画像記録方法は、医療画像を取得する映像入力ステップと、医療従事者の発生する音声を取得する音声取得ステップと、前記音声取得ステップにおいて取得した音声を解析する音声解析ステップと、前記音声解析ステップにおける解析結果による語句に基づいて、手技のフェーズを推定する手技フェーズ推定ステップと、前記手技フェーズ推定ステップにおいて推定されたフェーズの変化と予め設定されているフェーズの時系列との比較に基づいて、推定されたフェーズを確定する判定ステップと、前記判定ステップにおける判定結果に応じて、前記医療画像と前記フェーズとの関係を対応付けるインデックス情報を前記医療画像に付与して記録データを生成して記録部に記録する記録データ生成ステップとを具備する。 An image recording method according to an aspect of the present invention includes a video input step of acquiring a medical image, an audio acquisition step of acquiring audio generated by a medical worker, and an audio analysis step of analyzing audio acquired in the audio acquisition step. A procedure phase estimation step of estimating a procedure phase based on words and phrases obtained by analysis in the voice analysis step, a change of the phase estimated in the procedure phase estimation step, and a time series of predetermined phases Based on the comparison of the second phase, index information for correlating the relationship between the medical image and the phase according to the determination step of determining the estimated phase, and the determination result in the determination step, Recording data generation step of generating the .
本発明の第1の実施の形態に係る画像記録装置を示すブロック図。FIG. 1 is a block diagram showing an image recording apparatus according to a first embodiment of the present invention. 図1の画像記録装置が配置される手術室の様子を示す説明図。Explanatory drawing which shows the mode of the operating room in which the image recording apparatus of FIG. 1 is arrange | positioned. 記憶部60に記憶されているルックアップテーブルの内容の例を説明するための説明図。Explanatory drawing for demonstrating the example of the content of the look-up table memorize | stored in the memory | storage part 60. FIG. 第1の実施の形態の動作を説明するためのフローチャート。6 is a flowchart for explaining the operation of the first embodiment. 記録部55に生成された画像ファイルの一例を示す説明図。FIG. 6 is an explanatory view showing an example of an image file generated by a recording unit 55. 記録部55に生成された画像ファイルの一例を示す説明図。FIG. 6 is an explanatory view showing an example of an image file generated by a recording unit 55. 本発明の第2の実施の形態に採用される動作フローを示すフローチャート。The flowchart which shows the operation | movement flow employ | adopted as the 2nd Embodiment of this invention. 記憶部60に記憶されている2つの手技のルックアップテーブルの一例を示す説明図。Explanatory drawing which shows an example of the look-up table of two procedures memorize | stored in the memory | storage part 60. FIG. 記憶部60に記憶されている2つの手技のルックアップテーブルの一例を示す説明図。Explanatory drawing which shows an example of the look-up table of two procedures memorize | stored in the memory | storage part 60. FIG. 本発明の第3の実施の形態に採用される動作フローを示すフローチャート。The flowchart which shows the operation | movement flow employ | adopted in the 3rd Embodiment of this invention. 第3の実施の形態において採用されるルックアップテーブルの内容の一例を示す説明図。Explanatory drawing which shows an example of the content of the look-up table employ | adopted in 3rd Embodiment.
 以下、図面を参照して本発明の実施の形態について詳細に説明する。 Hereinafter, embodiments of the present invention will be described in detail with reference to the drawings.
(第1の実施の形態)
 図1は本発明の第1の実施の形態に係る画像記録装置を示すブロック図である。また、図2は図1の画像記録装置が配置される手術室の様子を示す説明図である。
First Embodiment
FIG. 1 is a block diagram showing an image recording apparatus according to a first embodiment of the present invention. FIG. 2 is an explanatory view showing the operating room where the image recording apparatus of FIG. 1 is disposed.
 通常、手術室には、ドクター及びナースが複数人配置されて手技が行われるようになっており、手技中において互いの連携を図るために、器具の受け渡しや容態の確認、機器の操作などに関して、要所々々で会話が交わされる。手技中において発せられる会話の内容は、手技の各フェーズに対応したものである場合がある。本実施の形態はこのような会話を利用して、手技のフェーズを推定し、推定したタイミングで編集ポイント等の特異ポイントの情報(後述するインデックス情報)を発生するようになっている。手技内の重要シーンでは、各シーンに対応した内容の会話が交わされる可能性が極めて高く、所望のシーンを特定する特異ポイントを確実に推定することができる。 Usually, multiple doctors and nurses are placed in the operating room to perform the procedure, and in order to achieve mutual cooperation during the procedure, the delivery of instruments, confirmation of the condition, operation of the instruments, etc. Conversations are exchanged at important points. The content of the conversation emitted during the procedure may correspond to each phase of the procedure. In the present embodiment, such a conversation is used to estimate the phase of the procedure, and at the estimated timing, information (specifically, index information described later) of unique points such as editing points is generated. In an important scene in a procedure, it is extremely likely that conversations of contents corresponding to each scene will be exchanged, and it is possible to reliably estimate a singular point that specifies a desired scene.
 まず、図2を参照して手術室2における画像記録装置50の配置の様子を説明する。図2に示すように、手術室2内に配置される医療システム3には、患者48が横たわる手術台10や電気メス装置13等の医療機器を制御するシステムコントローラ41が設けられている。手術室2には、第1カート11及び第2カート12が設けられており、第1カート11上にシステムコントローラ41が載置されている。 First, the arrangement of the image recording device 50 in the operating room 2 will be described with reference to FIG. As shown in FIG. 2, the medical system 3 disposed in the operating room 2 is provided with a system controller 41 that controls medical equipment such as the operating table 10 on which the patient 48 lies and the electric scalpel device 13. A first cart 11 and a second cart 12 are provided in the operating room 2, and a system controller 41 is placed on the first cart 11.
 また、第1カート11には、被制御装置である医療機器として例えば電気メス装置13、気腹装置14、ビデオプロセッサ15及び光源装置16等の装置類と、二酸化炭素を充填したガスボンベ18が載置されている。ビデオプロセッサ15は、カメラケーブル31aを介して第1の内視鏡31に接続される。 Further, on the first cart 11, devices such as an electric scalpel device 13, an insufflation device 14, a video processor 15, and a light source device 16 as medical devices which are controlled devices, and a gas cylinder 18 filled with carbon dioxide are mounted. It is placed. The video processor 15 is connected to the first endoscope 31 via a camera cable 31a.
 光源装置16は、ライトガイドケーブル31bを介して第1の内視鏡31に接続される。また、第1カート11には、表示装置19、第1の集中表示パネル20、操作パネル49等が載置されている。表示装置19は、例えばビデオプロセッサ15からの内視鏡画像等を表示する、例えばTVモニタである。 The light source device 16 is connected to the first endoscope 31 via the light guide cable 31 b. Further, on the first cart 11, the display device 19, the first central display panel 20, the operation panel 49 and the like are placed. The display device 19 is, for example, a TV monitor that displays an endoscope image or the like from the video processor 15, for example.
 集中表示パネル20は、手術中のあらゆるデータを選択的に表示させることが可能な表示手段となっている。操作パネル49は、例えば液晶ディスプレイ等の表示画面とこの表示画面上に一体的に設けられた例えばタッチセンサ等により構成され、非滅菌域にいる看護師等が操作する集中操作装置になっている。 The central display panel 20 is a display means capable of selectively displaying any data during surgery. The operation panel 49 includes, for example, a display screen such as a liquid crystal display and a touch sensor integrally provided on the display screen, and is a centralized operation device operated by a nurse or the like who is in a non-sterile area. .
 集中制御装置であるシステムコントローラ41には、手術台10と無影灯6と電気メス装置13と気腹装置14とビデオプロセッサ15と光源装置16とが、図示しない通信線を介して接続されている。 The operating table 10, the shadowless lamp 6, the electric scalpel device 13, the insufflation device 14, the video processor 15, and the light source device 16 are connected to the system controller 41 which is a central control device via a communication line (not shown) There is.
 また、第1カート11には、第1の内視鏡31や、電気メス装置13の処置具などに埋め込まれているIDタグにより、物の個別ID情報を無線にて読み取り/書き取りできるRFID(Radio Frequency Identification)端末35が設けられている。 Further, the first cart 11 can read / write the individual ID information of an object wirelessly by an ID tag embedded in the first endoscope 31 or the treatment tool of the electric scalpel device 13 or the like. A Radio Frequency Identification) terminal 35 is provided.
 一方、第2カート12には、被制御装置であるビデオプロセッサ23、光源装置24、画像処理装置25、表示装置26、第2の集中表示パネル27及び画像記録装置50が載置されている。ビデオプロセッサ23はカメラケーブル32aを介して第2の内視鏡32に接続される。光源装置24はライトガイドケーブル32bを介して第2の内視鏡32に接続される。 On the other hand, on the second cart 12, a video processor 23, which is a controlled device, a light source device 24, an image processing device 25, a display device 26, a second centralized display panel 27, and an image recording device 50 are placed. The video processor 23 is connected to the second endoscope 32 via a camera cable 32a. The light source device 24 is connected to the second endoscope 32 via the light guide cable 32b.
 表示装置26は、ビデオプロセッサ23でとらえた内視鏡画像等を表示する。第2の集中表示パネル27は、手術中のあらゆるデータを選択的に表示させることが可能になっている。 The display device 26 displays an endoscopic image or the like captured by the video processor 23. The second central display panel 27 can selectively display any data during the operation.
 これらビデオプロセッサ23と光源装置24と画像処理装置25と画像記録装置50とは、第2カート12に載置された中継ユニット28に図示しない通信線を介して接続されている。そして、この中継ユニット28は、中継ケーブル29によって、第1カート11に搭載されているシステムコントローラ41に接続されている。 The video processor 23, the light source device 24, the image processing device 25 and the image recording device 50 are connected to the relay unit 28 placed on the second cart 12 via a communication line (not shown). The relay unit 28 is connected to the system controller 41 mounted on the first cart 11 by a relay cable 29.
 こうして、システムコントローラ41は、これらの第2カート12に搭載されているビデオプロセッサ23、光源装置24、画像処理装置25及び画像記録装置50と、第1カート11に搭載されている、電気メス装置13、気腹装置14、ビデオプロセッサ15及び光源装置16と、手術台10とを集中制御することができるようになっている。システムコントローラ41は、システムコントローラ41とこれらの装置との間で通信が行われている場合、操作パネル49の表示画面上に、接続されている装置の設定状態や操作スイッチ等の設定画面を表示することができるようになっている。さらに、システムコントローラ41は、所望の操作スイッチが触れられて所定領域のタッチパネルが操作されることによって設定値の変更等の操作入力が行えるようになっている。 Thus, the system controller 41 includes the video processor 23 mounted on the second cart 12, the light source device 24, the image processing device 25 and the image recording device 50, and the electric scalpel device mounted on the first cart 11. 13, the insufflation device 14, the video processor 15, the light source device 16, and the operating table 10 can be centrally controlled. When communication is performed between the system controller 41 and these devices, the system controller 41 displays setting states of connected devices, setting screens such as operation switches, etc. on the display screen of the operation panel 49. It can be done. Furthermore, the system controller 41 can perform operation input such as change of setting values by touching a desired operation switch and operating the touch panel in a predetermined area.
 リモートコントローラ30は、滅菌域にいる執刀医等が操作する第2集中操作装置であり、通信が成立している他の装置を、システムコントローラ41を介して操作することができるようになっている。 The remote controller 30 is a second centralized control device operated by a surgeon or the like who is in a sterile area, and can operate other devices with established communication via the system controller 41. .
 また、システムコントローラ41には、通信手段である赤外線通信ポート(図示せず)が取り付けられている。この赤外線通信ポートは、表示装置19の近傍等の赤外線が照射しやすい位置に設けられ、システムコントローラ41との間がケーブルで接続されている。 Further, an infrared communication port (not shown) which is a communication means is attached to the system controller 41. The infrared communication port is provided at a position where infrared light can be easily emitted, such as in the vicinity of the display device 19, and the system controller 41 is connected with a cable.
 このシステムコントローラ41は、ケーブル9により患者モニタシステム4に接続されており、患者モニタシステム4は生体情報を解析し、この解析結果を所要の表示装置に表示させることができる。 The system controller 41 is connected to the patient monitoring system 4 by the cable 9. The patient monitoring system 4 can analyze biological information and can display the analysis result on a required display device.
 なお、手術室2内には、手術台10等の医療機器等を撮像するカメラ37も設けられている。カメラ37によって手術台10等の医療機器を撮像し、その撮像画像を解析することで、その動作状態を判定することが可能である。この判定結果やカメラ37による撮像画像はシステムコントローラ41に供給されるようになっている。 In the operating room 2, a camera 37 for imaging a medical device such as the operating table 10 is also provided. It is possible to determine the operation state by imaging a medical device such as the operating table 10 with the camera 37 and analyzing the captured image. The determination result and the image captured by the camera 37 are supplied to the system controller 41.
 ビデオプロセッサ15,23は、それぞれ内視鏡31,32の出力に基づく内視鏡画像を生成することができるようになっている。ビデオプロセッサ15,23からの内視鏡画像は画像記録装置50に供給されるようになっている。また、ビデオプロセッサ15,23と画像記録装置50とは、図示しないネットワークを介して接続されており、ビデオプロセッサ15,23から検査情報を含む各種情報が画像記録装置50に供給されるようになっている(図1では省略)。なお、ビデオプロセッサ15,23と画像記録装置50とを接続するネットワークとしては、各種通信規格の通信回線を採用することができる。 The video processors 15 and 23 can generate endoscopic images based on the outputs of the endoscopes 31 and 32, respectively. The endoscopic images from the video processors 15 and 23 are supplied to the image recording device 50. Also, the video processors 15 and 23 and the image recording device 50 are connected via a network (not shown), and various information including inspection information is supplied from the video processors 15 and 23 to the image recording device 50. (Not shown in Figure 1). As a network for connecting the video processors 15 and 23 and the image recording apparatus 50, communication lines of various communication standards can be adopted.
 本実施の形態においては、画像記録装置50には、ヘッドセット型のマイク33を接続できるようになっている。マイク33は装着者が発した音声を収音して音声信号を画像記録装置50に出力する。なお、図1では、1つのマイク33のみを示しているが、画像記録装置50には複数のマイク33を接続可能であり、画像記録装置50は、複数のマイク33からの音声を取得することができるようになっている。また、マイク33はケーブルにより有線接続された例を示しているが、Wi-Fi(登録商標)やブルートゥース(登録商標)等の無線伝送路を介して音声信号を画像記録装置50に伝送することができるものであってもよい。 In the present embodiment, a headset type microphone 33 can be connected to the image recording apparatus 50. The microphone 33 picks up the voice emitted by the wearer and outputs a voice signal to the image recording device 50. Although only one microphone 33 is shown in FIG. 1, a plurality of microphones 33 can be connected to the image recording device 50, and the image recording device 50 acquires audio from the plurality of microphones 33. It is possible to Also, although the microphone 33 is shown as being wired connected by a cable, transmitting an audio signal to the image recording apparatus 50 via a wireless transmission path such as Wi-Fi (registered trademark) or Bluetooth (registered trademark). You may be able to
 また、手術室2内には、図示しない支持部材に支持されて、マイク34も設けられている。マイク34は、図示しないケーブルやWi-Fiやブルートゥース(登録商標)等の無線伝送路を介して画像記録装置50に接続されており、手術室2内の音声を収音して音声信号を画像記録装置50に供給することができるようになっている。 In the operating room 2, a microphone 34 is also provided, supported by a support member (not shown). The microphone 34 is connected to the image recording apparatus 50 via a cable (not shown) or a wireless transmission path such as Wi-Fi or Bluetooth (registered trademark), and picks up the sound in the operating room 2 to image the sound signal. The recording apparatus 50 can be supplied.
 なお、図1では複数のマイク33,34を採用する例を説明したが、手術室2内のドクターやナースの会話を収音することができれば、マイクや音声信号の伝送路の種類及び数等は適宜設定可能である。 In addition, although the example which employ | adopts several microphones 33 and 34 was demonstrated in FIG. 1, if the doctor's and nurse's conversation in the operating room 2 can be collected, the kind and number of transmission paths of a microphone or an audio signal etc. Can be set as appropriate.
 図1は図2中の画像記録装置50の具体的な構成の一例を示している。 FIG. 1 shows an example of a specific configuration of the image recording apparatus 50 in FIG.
 画像記録装置50には制御部51が設けられている。制御部51は、画像記録装置50の各部を制御することができるようになっている。制御部51としては、図示しないCPU等のプロセッサによって構成されて、図示しないメモリに格納されたプログラムに従って動作して各部を制御するものであってもよく、フィールドプログラマブルゲートアレイ(FPGA)等によって構成されていてもよい。 The image recording device 50 is provided with a control unit 51. The control unit 51 can control each part of the image recording apparatus 50. The control unit 51 may be configured by a processor such as a CPU (not shown) and operated according to a program stored in a memory (not shown) to control each unit, and may be configured by a field programmable gate array (FPGA) or the like. It may be done.
 画像記録装置50の映像入力部52は、画像伝送に適したインターフェースであり、各種医療画像を取り込む。なお、映像入力部52は、DVI(Digital Visual Interface)端子、SDI(Serial Digital Interface)端子、RGB端子、Y/C端子、VIDEO端子等の種々の端子を採用することができる。映像入力部52は、例えば、ビデオプロセッサ15,23からの内視鏡画像や、超音波装置、術野カメラ、X線観測装置、ビデオプロセッサ15,23とは別の図示しない内視鏡プロセッサ等からの各種医療画像を取り込むことができる。 The video input unit 52 of the image recording device 50 is an interface suitable for image transmission, and takes in various medical images. The video input unit 52 can adopt various terminals such as a DVI (Digital Visual Interface) terminal, an SDI (Serial Digital Interface) terminal, an RGB terminal, a Y / C terminal, and a VIDEO terminal. The image input unit 52 is, for example, an endoscopic image from the video processor 15 or 23, an ultrasound apparatus, an operation field camera, an X-ray observation apparatus, an endoscope processor (not shown) other than the video processor 15 or 23 Can capture various medical images from
 映像入力部52が取り込んだ医療画像は、記録データ生成部54に与えられる。記録データ生成部54は、入力された医療画像に対して所定のエンコード処理を行うことで、入力された医療画像を所定の画像形式の映像信号に変換する。例えば、記録データ生成部54は、入力された医療画像をMPEG2形式やMPEG-4AVC/H.264形式等の映像信号に変換して記録データとして出力することができるようになっている。 The medical image captured by the image input unit 52 is given to the record data generation unit 54. The recording data generation unit 54 converts the input medical image into a video signal of a predetermined image format by performing a predetermined encoding process on the input medical image. For example, the recording data generation unit 54 may input the medical image into the MPEG2 format or the MPEG-4 AVC / H. It is possible to convert it into a video signal of the H.264 format or the like and output it as recording data.
 本実施の形態においては、記録データ生成部54は、制御部51に制御されて、記録データにインデックス情報を付加することができるようになっている。例えば、記録データ生成部54は、記録データである画像ファイル中にメタ情報としてインデックス情報を含めてもよく、また、画像ファイルとは別のインデックス情報を含むファイルを生成してもよく、これらの2種類のインデックス情報を記録するようになっていてもよい。 In the present embodiment, the recording data generation unit 54 is controlled by the control unit 51 and can add index information to the recording data. For example, the recording data generation unit 54 may include index information as meta information in an image file which is recording data, or may generate a file including index information different from the image file. Two types of index information may be recorded.
 例えば、手技には、麻酔、剥離、切除、切離、止血、バイパスフェーズ等の各段階(フェーズ)がある。インデックス情報は、記録データに含まれる医療画像の時間情報に対応しており、後述するように、医療画像の各シーンのうち、手技の各段階(フェーズ)の記録時間に対応した情報である。再生時においてインデックス情報を用いることで、医療画像中の各フェーズの所定位置、例えば先頭位置に再生位置をジャンプさせることが可能である。 For example, the procedure includes phases such as anesthesia, ablation, excision, dissection, hemostasis, and bypass phase. The index information corresponds to the time information of the medical image included in the recording data, and is information corresponding to the recording time of each stage (phase) of the procedure among the scenes of the medical image as described later. By using index information at the time of reproduction, it is possible to make the reproduction position jump to a predetermined position of each phase in the medical image, for example, the head position.
 記録データ生成部54は、生成した記録データを記録部55に出力するようになっている。記録部55は、制御部51に制御されて、医療画像の記録データを画像ファイルとして記録する。例えば、記録部55としては、ハードディスク装置等を採用することができる。 The recording data generation unit 54 is configured to output the generated recording data to the recording unit 55. The recording unit 55 is controlled by the control unit 51 to record the record data of the medical image as an image file. For example, a hard disk drive or the like can be employed as the recording unit 55.
 また、記録部55は、制御部51に制御されて、記録データを読出して外部メディア記録再生部56に出力することができる。外部メディア記録再生部56は、記録部55からの記録データを図示しない外部の記録媒体である外部メディア65に与えて記録させることができるようになっている。なお、外部メディア65としては、BD(ブルーレイディスク)、DVD、USBだけでなく、ネットワーク上のサーバ等を採用してもよく、また、それ以外の記録媒体を採用してもよい。 Also, the recording unit 55 can be controlled by the control unit 51 to read out the recording data and output it to the external media recording and reproducing unit 56. The external media recording and reproducing unit 56 can apply recording data from the recording unit 55 to an external medium 65, which is an external recording medium (not shown), and can record the data. As the external medium 65, not only BD (Blu-ray Disc), DVD, USB, but also a server or the like on a network may be adopted, and other recording media may be adopted.
 また、記録データ生成部54は、入力された医療画像を映像出力部57に出力することもできるようになっている。映像出力部57は、入力された医療画像を外部モニタ66に出力する。これにより、外部モニタ66において、医療画像の表示が可能である。 The record data generation unit 54 can also output the input medical image to the video output unit 57. The video output unit 57 outputs the input medical image to the external monitor 66. Thereby, the external monitor 66 can display a medical image.
 本実施の形態においては、画像記録装置50には音声取得部53も設けられている。音声取得部53には、マイク33,34からの音声信号が入力される。音声取得部53は、マイク33,34からの音声信号の伝送路が有線の場合にはケーブルが接続されるコネクタ等により構成され、マイク33,34からの音声信号の伝送路が無線の場合には無線信号を受信するアンテナや受信機等により構成されて、入力された音声信号を取り込んで音声解析部58に出力する。 In the present embodiment, the image recording apparatus 50 is also provided with a sound acquisition unit 53. The voice acquisition unit 53 receives voice signals from the microphones 33 and 34. The audio acquisition unit 53 is configured by a connector or the like to which a cable is connected when the transmission path of the audio signal from the microphones 33 and 34 is wired, and the transmission path of the audio signal from the microphones 33 and 34 is wireless. Is constituted by an antenna, a receiver or the like for receiving a radio signal, and takes in the input voice signal and outputs it to the voice analysis unit 58.
 音声取得部53は、複数の音声信号を取得可能である。音声取得部53は、各音声信号がいずれのマイクにより収音されたものであるかを検出可能である。例えば、音声取得部53は、音声信号の伝送が有線で行われる場合には、各ケーブルに接続されたコネクタによっていずれのマイクにより収音された音声信号であるかを検出することができ、音声信号の伝送が無線で行われる場合には、例えば無線通信の確立時に取得した機器ID等によっていずれのマイクにより収音された音声信号であるかを検出することができる。 The sound acquisition unit 53 can acquire a plurality of sound signals. The voice acquisition unit 53 can detect which microphone each voice signal is picked up by. For example, when transmission of an audio signal is performed by wire, the audio acquisition unit 53 can detect which microphone the audio signal is picked up by which connector is connected by the connector connected to each cable. In the case where signal transmission is performed wirelessly, it is possible to detect which microphone an audio signal is picked up by, for example, a device ID acquired at the time of establishment of wireless communication.
 音声解析部58は、入力された音声に対する解析処理を行って音声解析結果を得る。例えば、音声解析部58は、予めシステムに用意されている音響モデル、単語辞書及び言語モデル等を利用して音声認識を行う。例えば、音声解析部58は、音声認識処理によって、マイク33,34によって収音したドクターやナースの発話内容を解析し、ドクターやナースが発話した単語や会話の音声認識結果を取得して解析結果として判定部59に出力する。 The voice analysis unit 58 performs analysis processing on the input voice to obtain a voice analysis result. For example, the speech analysis unit 58 performs speech recognition using an acoustic model, a word dictionary, a language model and the like prepared in advance in the system. For example, the speech analysis unit 58 analyzes the speech contents of the doctor or nurse collected by the microphones 33 and 34 by speech recognition processing, acquires the speech recognition result of the word or speech spoken by the doctor or nurse, and analyzes the result Output to the determination unit 59 as
 判定部59は、記憶部60に記憶されているルックアップテーブルを参照することで、マイク33,34による収音音声に対応した手技のフェーズを判定するようになっている。なお、判定部59は、CPU等を用いたプロセッサによって構成されて、図示しないメモリに記憶されたプログラムに従って動作するものであってもよいし、ハードウェアの電子回路で機能の一部又は全部を実現するものであってもよい。 The determination unit 59 determines the phase of the procedure corresponding to the voice collected by the microphones 33 and 34 by referring to the lookup table stored in the storage unit 60. The determination unit 59 may be configured by a processor using a CPU or the like, and may operate in accordance with a program stored in a memory (not shown), or a part or all of the functions of the electronic circuit of hardware. It may be realized.
 図3は記憶部60に記憶されているルックアップテーブルの内容の例を説明するための説明図である。図3のルックアップテーブルは、手技の各フェーズ(Phase1,Phase2,…Phasen)と対応する語句との関係を登録したものである。図3では、例えば、Phase1に対応して「開始」及び「始める」という語句が登録され、Phase5に対応して「切離完了」、「切離終了」及び「切離終わり」という語句が登録されている。なお、図3では各フェーズと1つ又は比較的少ない語数の語句とを対応させた例を示しているが、各フェーズと比較的多くの語数からなる会話の内容とを対応させたルックアップテーブルを採用してもよい。 FIG. 3 is an explanatory diagram for explaining an example of the contents of the lookup table stored in the storage unit 60. As shown in FIG. The look-up table of FIG. 3 registers the relationship between each phase (Phase 1, Phase 2,..., Phase n) of the procedure and the corresponding word. In FIG. 3, for example, the phrases "start" and "start" are registered corresponding to Phase 1, and the phrases "separation complete", "separation complete" and "end complete" are registered corresponding to Phase 5 It is done. Although FIG. 3 shows an example in which each phase is associated with one or a relatively small number of words or phrases, a look-up table in which each phase is associated with the contents of conversation consisting of a relatively large number of words. May be adopted.
 なお、図3の各フェーズ(Phase1,Phase2,…)は、フェーズ(Phase1,Phase2,…)の順に時系列に発生するものである。また、図3は所定の手技に対応したものであり、手技が異なれば、その手技に対応した各フェーズとその各フェーズに対応した語句が登録されたルックアップテーブルが採用される。 In addition, each phase (Phase1, Phase2, ...) of FIG. 3 generate | occur | produces in order of a phase (Phase1, Phase2, ...) in time series. Further, FIG. 3 corresponds to a predetermined procedure, and if the procedure is different, each phase corresponding to the procedure and a lookup table in which words and phrases corresponding to each phase are registered are adopted.
 判定部59は手技フェーズ推定部59aを備えている。手技フェーズ推定部59aは、音声解析部58の解析結果により得られた語句を用いてルックアップテーブルを参照することで、収音音声から得た語句の内容に対応したフェーズを推定し、推定結果を制御部51に出力するようになっている。例えば、手技フェーズ推定部59aは、音声の解析結果として、「剥離」が得られた場合には、記憶部60に記憶されているルックアップテーブルを参照することで、マイク33,34によって取得された音声のタイミングでフェーズ(Phase7)に手技が移行するものと推定する。 The determination unit 59 includes a procedure phase estimation unit 59a. The procedure phase estimation unit 59a estimates the phase corresponding to the content of the phrase obtained from the collected voice by referring to the lookup table using the phrase obtained by the analysis result of the voice analysis unit 58, and the estimation result Are output to the control unit 51. For example, the procedure phase estimation unit 59a is acquired by the microphones 33 and 34 by referring to the look-up table stored in the storage unit 60 when “peeling” is obtained as the analysis result of the voice. It is estimated that the procedure shifts to the phase (Phase 7) at the timing of the voice.
 なお、図1では手技フェーズ推定部59aを判定部59内に設けた例を示したが、手技フェーズ推定部59aを判定部59とは別に設けてもよい。 Although FIG. 1 shows an example in which the procedure phase estimation unit 59a is provided in the determination unit 59, the procedure phase estimation unit 59a may be provided separately from the determination unit 59.
 画像記録装置50には、操作部61が設けられている。操作部61は、例えば、図示しない、スイッチ、ボタン、キーやタッチパネル等により構成されており、ユーザ操作を受け付けて、ユーザ操作に基づく操作信号を制御部51に出力するようになっている。制御部51は、操作部61に対するユーザ操作に基づいて、各部を制御するようになっている。例えば、制御部51は、操作部61に対するユーザ操作に基づいて、記憶部60に記憶されているルックアップテーブルを編集することができるようになっている。 The image recording device 50 is provided with an operation unit 61. The operation unit 61 includes, for example, switches, buttons, keys, and a touch panel (not shown), and is configured to receive a user operation and output an operation signal based on the user operation to the control unit 51. The control unit 51 is configured to control each unit based on a user operation on the operation unit 61. For example, the control unit 51 can edit the look-up table stored in the storage unit 60 based on a user operation on the operation unit 61.
 なお、操作部61によって患者に関する各種情報を入力することも可能であり、制御部51は、患者に関する各種情報を記録データ生成部54に与えて記録する画像ファイルのメタデータとして記録させることもできるようになっている。制御部51は、患者に関する各種情報を、図示しない外部のデータベースサーバから図示しない通信回路を介して取り込むようになっていてもよい。 Note that it is also possible to input various types of information about the patient by the operation unit 61, and the control unit 51 can also provide various types of information about the patient to the recording data generation unit 54 and record them as metadata of an image file to be recorded It is supposed to be. The control unit 51 may be configured to take in various information related to the patient from an external database server (not shown) via a communication circuit (not shown).
 また、制御部51は、外部メディア記録再生部56を介して外部メディア65に記憶されているルックアップテーブルを読出して、記憶部60に格納するようになっていてもよい。 In addition, the control unit 51 may read the look-up table stored in the external medium 65 via the external medium recording and reproducing unit 56 and store the lookup table in the storage unit 60.
 判定部59は、時系列に並んだ各フェーズの順番を考慮して、音声認識に基づくフェーズの推定結果が実際のフェーズの順番に対応したフェーズを示しているか否かを判定し、一致している場合には音声認識に基づくフェーズの推定結果が正しいものと判定し、そうでない場合には音声認識に基づくフェーズの推定結果は間違っているものと判定する。即ち、判定部59は、手技フェーズの推定結果であるフェーズの変化と予め設定されているフェーズの時系列との比較に基づいて、推定されたフェーズを確定する。 The determination unit 59 determines whether or not the estimation result of the phase based on the speech recognition indicates a phase corresponding to the actual order of the phases in consideration of the order of each phase arranged in time series. If so, it is determined that the estimation result of the phase based on speech recognition is correct, otherwise it is determined that the estimation result of the phase based on speech recognition is incorrect. That is, the determination unit 59 determines the estimated phase based on the comparison between the change of the phase, which is the estimation result of the procedure phase, and the time series of the phase set in advance.
 例えば、判定部59は、現在のフェーズがフェーズ(Phase5)で正しいと判定した後に、音声認識に基づいてフェーズ(Phase4)の推定結果が得られた場合には、当該フェーズ(Phase4)の推定結果は間違っているものと判定し、音声認識に基づいてフェーズ(Phase6)の推定結果が得られた場合には、当該フェーズ(Phase6)の推定結果は正しいものと判定する。判定部59は、音声認識に基づくフェーズの推定結果が時系列的に正しい場合にのみ、音声認識に基づいてフェーズの移行を判定したこと及び判定したフェーズの情報を含む判定結果を、記憶部60に記憶させると共に、制御部51に出力するようになっている。なお、判定部59は、判定結果については、判定部59内の図示しないメモリに記憶させるようになっていてもよい。 For example, if the determination unit 59 determines that the current phase is correct in the phase (Phase 5) and then the estimation result of the phase (Phase 4) is obtained based on speech recognition, the estimation result of the phase (Phase 4) If the estimation result of the phase (Phase 6) is obtained based on speech recognition, it is determined that the estimation result of the phase (Phase 6) is correct. The determination unit 59 stores the determination result including the determination of the phase transition based on speech recognition and the information on the determined phase only when the estimation result of the phase based on speech recognition is correct in time series. , And output to the control unit 51. The determination unit 59 may store the determination result in a memory (not shown) in the determination unit 59.
 本実施の形態においては、制御部51は、判定部59からフェーズの判定結果が与えられると、判定結果が入力されたタイミングで、記録データに当該判定結果に対応したインデックス情報を付加するように、記録データ生成部54に記録要請を発生する。これにより、記録データ生成部54は、手技のフェーズが次のフェーズに移行するタイミング、即ち、記録する医療画像のシーンが次のフェーズに移行する記録時間(再生時間)に関連付けたインデックス情報を医療画像の記録データに付加する。 In the present embodiment, when control unit 51 receives the determination result of the phase from determination unit 59, control unit 51 adds index information corresponding to the determination result to the recording data at the timing when the determination result is input. And generates a recording request to the recording data generation unit 54. Thereby, the recording data generation unit 54 performs medical treatment on the index information associated with the timing when the phase of the procedure shifts to the next phase, that is, the recording time (reproduction time) at which the scene of the medical image to be recorded shifts to the next phase. Add to recorded data of image.
 例えば、記録データ生成部54がエンコード処理における時間基準となる時間情報を、記録する医療画像のタイムコードとして記録データに含める場合には、インデックス情報は、各フェーズに移行するタイミングのタイムコードの値を含む情報であってもよい。また、インデックス情報は、記録時間の情報だけでなく、判定したフェーズの種類に関する情報を含んでいてもよい。 For example, in the case where the recording data generation unit 54 includes time information serving as a time reference in encoding processing in the recording data as the time code of the medical image to be recorded, the index information is the value of the time code of the timing to shift to each phase. It may be information including In addition, the index information may include not only information on recording time but also information on the type of the determined phase.
 また、記録データ生成部54は、インデックス情報を記録する画像ファイルとは独立した別のファイルとして生成してもよい。例えば、記録データ生成部54は、音声認識に基づいてフェーズを判定する毎に、判定の時間(記録時間、再生時間)の情報及び音声認識結果のテキスト情報をリスト化したファイルを生成してもよい。 In addition, the recording data generation unit 54 may generate the file as another file independent of the image file for recording the index information. For example, each time the recording data generation unit 54 determines a phase based on voice recognition, even if it generates a file in which information of determination time (recording time, reproduction time) and text information of voice recognition results are listed. Good.
 なお、制御部51は、音声解析部58において音声解析されている音声がいずれのマイクにより収音されたものであるかを示す情報が音声取得部53から与えられるようになっている。図示しないメモリにマイクとマイクの使用者との対応関係を登録したテーブルが記憶されている場合には、制御部51は、音声取得部53からの情報を用いて当該テーブルを参照することで、フェーズの判定に用いられた語句等の話者を特定し、インデックス情報に、当該話者の情報を含めるように記録データ生成部54を制御することが可能である。 The control unit 51 is configured to receive, from the voice acquisition unit 53, information indicating which microphone the voice subjected to voice analysis in the voice analysis unit 58 is picked up by. When a table in which the correspondence between the microphone and the user of the microphone is registered is stored in a memory (not shown), the control unit 51 refers to the table using the information from the voice acquisition unit 53, It is possible to specify a speaker such as a word and the like used in the phase determination, and control the recording data generation unit 54 so that the information of the speaker is included in the index information.
 次に、このように構成された実施の形態の動作について図4を参照して説明する。図4は第1の実施の形態の動作を説明するためのフローチャートである。 Next, the operation of the embodiment configured as described above will be described with reference to FIG. FIG. 4 is a flowchart for explaining the operation of the first embodiment.
 いま、図3の各フェーズ(Phase1,Phase2,…)を有する手技が行われるものとする。映像入力部52は、例えば、ビデオプロセッサ15からの医療画像を取り込んで記録データ生成部54に出力する。記録データ生成部54は、制御部51に制御されて、図4のステップS1において、医療画像を所定の形式の動画像に変換して記録部55への記録を開始する。また、記録データ生成部54からの医療画像は、映像出力部57によって外部モニタ66に表示される。 Now, it is assumed that a procedure having each phase (Phase 1, Phase 2,...) Of FIG. 3 is performed. For example, the video input unit 52 takes in a medical image from the video processor 15 and outputs the medical image to the recording data generation unit 54. The record data generation unit 54 is controlled by the control unit 51, converts the medical image into a moving image of a predetermined format, and starts recording in the recording unit 55 in step S1 of FIG. 4. Further, the medical image from the record data generation unit 54 is displayed on the external monitor 66 by the video output unit 57.
 一方、音声取得部53は、複数のマイク33,34によって収音された音声を取り込んで音声解析部58に出力する。音声解析部58は、ステップS2において、公知の音声認識処理によって、ドクターやナースが発した語句(会話)を取得し、音声認識結果を判定部59に出力する。判定部59の手技フェーズ推定部59aは、音声認識結果を用いて記憶部60のルックアップテーブルを参照することで、音声認識結果に対応するフェーズの推定を行う(ステップS3)。 On the other hand, the voice acquisition unit 53 takes in the voice collected by the plurality of microphones 33 and 34 and outputs the voice to the voice analysis unit 58. In step S 2, the voice analysis unit 58 acquires the words (conversation) generated by the doctor or the nurse by a known voice recognition process, and outputs the voice recognition result to the determination unit 59. The procedure phase estimation unit 59a of the determination unit 59 estimates the phase corresponding to the speech recognition result by referring to the lookup table of the storage unit 60 using the speech recognition result (step S3).
 ここで、ドクターが手技の開始に際して、「開始」又は「始める」と発話するものとする。音声解析部58は、ドクターの発話を音声認識して認識結果を判定部59に出力する。判定部59の手技フェーズ推定部59aは、「開始」又は「始める」という語句の音声認識結果が記憶部60のルックアップテーブル(LUT)に存在するものと判定すると、当該語句に対応するフェーズがフェーズ(Phase1)であることを示す推定結果を取得する。なお、音声認識により得られた語句が記憶部60のルックアップテーブル(LUT)に存在しない場合には、ステップS3から処理はステップS7に移行する。制御部51は、ステップS7において、医療画像である動画像の記録終了操作があったか否かを判定し、記録終了操作がない場合には処理をステップS2に戻し、記録終了操作があった場合にはステップS8において記録終了処理を行う。 Here, it is assumed that the doctor utters "start" or "start" at the start of the procedure. The voice analysis unit 58 performs voice recognition of the doctor's speech and outputs the recognition result to the determination unit 59. When the procedure phase estimation unit 59a of the determination unit 59 determines that the speech recognition result of the words "start" or "start" is present in the look-up table (LUT) of the storage unit 60, the phase corresponding to the word is An estimation result indicating that it is a phase (Phase 1) is acquired. When a word or phrase obtained by speech recognition does not exist in the look-up table (LUT) of the storage unit 60, the process proceeds from step S3 to step S7. In step S7, the control unit 51 determines whether or not there is an operation to end recording of a moving image which is a medical image. If there is no operation to end recording, the process returns to step S2, and if there is an operation to end recording. In step S8, the recording end process is performed.
 判定部59は、ステップS3において音声認識結果がルックアップテーブルに存在することを示す判定を行った場合には、次のステップS4において、記憶部60又は判定部59内の図示しないメモリに記憶されているこれまでのフェーズの判定結果を読出して、音声認識に基づく推定結果のフェーズの時系列が正しいか否かを判定する。フェーズ(Phase1)は手技開始後の最初のフェーズであり、このフェーズの推定結果は時系列的に正しいので、判定部59は、ステップS4からステップS5に処理を移行して推定結果を確定し、フェーズの移行及びフェーズの種類に関する情報を制御部51に出力する。なお、判定部59は、音声認識によるフェーズの推定結果が時系列的に正しくない場合には、今回の推定結果を用いることなく、処理をステップS2に戻す。 If determination unit 59 determines in step S3 that the speech recognition result is present in the lookup table, it is stored in memory 60 or a memory (not shown) in determination unit 59 in the next step S4. The determination results of the previous phases are read out to determine whether or not the time series of the phase of the estimation result based on speech recognition is correct. The phase (Phase 1) is the first phase after the start of the procedure, and the estimation result of this phase is correct in time series, so the determination unit 59 shifts the process from step S4 to step S5 and determines the estimation result, Information on the phase transition and the type of phase is output to the control unit 51. If the estimation result of the phase by speech recognition is not correct in time series, the determination unit 59 returns the process to step S2 without using the estimation result of this time.
 制御部51は、ステップS5において、例えば音声解析部58の出力がいずれのマイクの収音結果によるものであるかに基づいて、フェーズの判定に用いた音声を発した話者を特定した後、記録する動画像にインデックス情報を付加するための記録要請を行う(ステップS6)。 After the control unit 51 determines in step S5, for example, on which microphone the output of the voice analysis unit 58 is due to the sound collection result, the control unit 51 identifies the speaker who issued the voice used for the phase determination, A recording request for adding index information to a moving image to be recorded is issued (step S6).
 この記録要請に従って、記録データ生成部54は、記録データにインデックス情報を付加する。即ち、インデックス情報は、ドクターが「開始」又は「始める」と発話したタイミングにおいて記録される画像部分に対応する時間情報を有している。インデックス情報が付加された記録データは、記録部55に記録される。 In response to the recording request, the recording data generation unit 54 adds index information to the recording data. That is, the index information has time information corresponding to the image portion recorded at the timing when the doctor utters "start" or "start". The recording data to which the index information is added is recorded in the recording unit 55.
 制御部51は、ステップS6において記録要請を行った後、処理をステップS7に移行して、記録終了操作があったか否かを判定する。以後、ステップS2~S7の処理が繰り返されて、ドクターやナースの発話毎に、フェーズの判定及び記録データに対するインデックス情報の付加が行われる。 After making a recording request in step S6, the control unit 51 shifts the process to step S7 and determines whether or not a recording end operation has been performed. Thereafter, the processes of steps S2 to S7 are repeated, and the phase determination and the addition of the index information to the recording data are performed for each utterance of the doctor or the nurse.
 例えば、ドクターが「切除開始」と発話すると、このタイミングで、判定部59により手技のフェーズが時系列的に正しいフェーズ(Phase2)に移行したことが判定され、制御部51により、この判定に従ってインデックス情報が記録データに付加される。 For example, when the doctor utters "dissection start", at this timing, it is determined by the determination unit 59 that the phase of the procedure has shifted to the correct phase (Phase 2) in time series, and the control unit 51 indexes according to this determination Information is added to the recorded data.
 例えば、「切除開始」の発話によるフェーズ(Phase2)への移行に伴うインデックス情報の付加後に、ドクターが「開始」と発話するものとする。この場合には、判定部59の手技フェーズ推定部59aによる音声認識に基づくフェーズの推定結果はフェーズ(Phase2)からフェーズ(Phase1)に戻ることを示すものとなり、判定部59は、この推定結果が時系列的に正しくないものと判定して、「開始」の音声認識に基づく推定を無視する。こうして、図3のフェーズ(Phase1,Phase2,…)の時系列順に、各フェーズへの移行が正しく判定され、各フェーズへの移行に応じたタイミングでインデックス情報が記録データに付加される。 For example, it is assumed that the doctor utters “start” after the addition of the index information accompanying the transition to the phase (phase 2) by the utterance “dissection start”. In this case, the estimation result of the phase based on speech recognition by the procedure phase estimation unit 59a of the determination unit 59 indicates that the phase (Phase 2) returns to the phase (Phase 1), and the determination unit 59 determines that this estimation result is It determines that the time series is not correct, and ignores the estimation based on the "start" speech recognition. In this way, the transition to each phase is correctly determined in the time series of the phases (Phase1, Phase2,...) Of FIG. 3, and the index information is added to the recording data at the timing according to the transition to each phase.
 制御部51は、ステップS7において記録終了操作が行われたものと判定すると、ステップS8において記録データ生成部54に記録の終了を指示する。記録データ生成部54は、記録データをファイル化して記録部55に記録する。外部メディア記録再生部56は、記録部55に記録された画像ファイルを外部メディア65に記録することができる。なお、制御部51は、ドクターの「終了」という発話に基づくフェーズの判定結果が与えられた場合に、記録終了操作が行われたものと判定して、記録の終了及び画像ファイル化を行うようになっていてもよい。 If the control unit 51 determines that the recording end operation has been performed in step S7, the control unit 51 instructs the recording data generation unit 54 to end the recording in step S8. The recording data generation unit 54 converts the recording data into a file and records the file in the recording unit 55. The external media recording and reproducing unit 56 can record the image file recorded in the recording unit 55 in the external medium 65. In addition, when the determination result of the phase based on the doctor's "end" utterance is given, the control unit 51 determines that the recording end operation is performed, and performs the end of recording and the conversion into an image file. It may be
 図5及び図6は記録データに付加されるインデックス情報の例を示す説明図である。図5はインデックス情報を記録データのメタデータとして記録する例を示している。図5の例は、ある患者に対する所定の手技の記録データについて、メタデータとして付加されるインデックス情報を示している。図5の例では、インデックス情報は、「Date」、「Patient ID」、「Patient Name」、「BOD」、「Age」、「Sex」を含む患者情報を含んでいる。また、インデックス情報は、<タイトル1>で示す画像ファイルのファイル名の情報も含む。 5 and 6 are explanatory diagrams showing examples of index information added to recording data. FIG. 5 shows an example of recording index information as metadata of recording data. The example of FIG. 5 shows index information added as metadata for recorded data of a predetermined procedure for a certain patient. In the example of FIG. 5, the index information includes patient information including “Date”, “Patient ID”, “Patient Name”, “BOD”, “Age”, and “Sex”. The index information also includes information of the file name of the image file indicated by <title 1>.
 本実施の形態においては、インデックス情報は、フェーズの判定によって得られた情報、即ち、フェーズ判定の時刻、及び当該判定に用いた語句の発話者、フェーズ判定に用いた語句の情報を含む。図5の例では、発話者○○が発話した「剥離開始」の語句によって、時間xx時xx分xx秒に、剥離に関するフェーズが開始されたことが判定され、当該剥離に関する画像の記録時間(時間xx時xx分xx秒)を示すインデックス情報が記録データのメタデータとして記録されたことを示している。なお、記録時間の情報は、記録開始からの時間を示しており、再生時間と同一の情報である。 In the present embodiment, the index information includes information obtained by the determination of the phase, that is, the time of the phase determination, the speaker of the phrase used for the determination, and the information of the phrase used for the phase determination. In the example of FIG. 5, it is determined that the phase related to peeling has been started at time xx: xx, xx and xx according to the phrase “peeling start” uttered by the speaker OO, and the recording time of the image related to the peeling It indicates that index information indicating a time xx hour xx minute xx second) is recorded as metadata of recording data. Note that the information on the recording time indicates the time from the start of recording, and is the same information as the reproduction time.
 図1では再生系については図示を省略しているが、記録部55に記録された画像ファイルを読み出して映像出力部57により外部モニタ66に再生画像を出力させることも可能である。この場合には、図示しない再生回路は、記録データの再生時に、当該インデックス情報を利用して、記録画像の再生位置、編集位置をインデックス情報に対応する位置にジャンプさせることができる。これにより、各フェーズの先頭位置に移動して再生を行うことが容易となる。また、このようなメタデータを利用したジャンプ機能を有する再生機器であれば、外部メディア65に記録された画像ファイルの再生時に、インデックス情報を利用して各フェーズの先頭位置に移動して再生を行うことができる。 Although the reproduction system is not shown in FIG. 1, it is also possible to read out the image file recorded in the recording unit 55 and output the reproduced image to the external monitor 66 by the video output unit 57. In this case, the reproduction circuit (not shown) can jump the reproduction position and the editing position of the recorded image to the position corresponding to the index information by using the index information when reproducing the recording data. This makes it easy to move to the top position of each phase to perform reproduction. In addition, in the case of a playback device having a jump function using such metadata, when playing back an image file recorded on the external medium 65, the index information is used to move to the head position of each phase and play back. It can be carried out.
 図6はこのようなメタデータを利用したジャンプ機能を有していない再生機器を考慮して、画像ファイルに対応付けられたリスト形式のインデックス情報を示している。図6のリストは、再生時間の項目と対応するテキストの項目とを有する。即ち、再生時間の項目において、フェーズが判定された記録時間(再生時間)を示し、テキストの項目において、フェーズの判定に用いられた語句を示している。例えば、図6の例では、「これより被験者○○に対してxxを開始します」の語句によって、記録開始からの時間00時03分22秒後に、手技xxの最初のフェーズが開始されたことが判定され、当該フェーズに関する画像の記録時間(時間xx時xx分xx秒)を示すインデックス情報が記録データの画像ファイルとは別ファイルとして記録されたことを示している。 FIG. 6 shows index information in the form of a list associated with an image file in consideration of a playback device that does not have such a jump function using metadata. The list of FIG. 6 has an item of playback time and an item of corresponding text. That is, in the item of reproduction time, the recording time (reproduction time) in which the phase is determined is indicated, and in the item of text, the term used for the determination of the phase is indicated. For example, in the example of FIG. 6, the first phase of the procedure xx is started after time 00:03:22 from the start of recording by the phrase "Start xx for the subject OO from this" It is determined that the index information indicating the recording time (time xx: xx: xx) of the image regarding the phase is recorded as a file different from the image file of the recording data.
 図6のようなリスト形式のインデックス情報を用いることで、例えば、パーソナルコンピュータ等によって、画像ファイルを再生する際に、各フェーズの先頭位置の確認を容易に行うことができる。 By using index information in a list format as shown in FIG. 6, for example, when reproducing an image file by a personal computer or the like, it is possible to easily confirm the head position of each phase.
 このように本実施の形態においては、術者等が発した音声の解析により、手技の各フェーズへの移行を判定しており、記録する画像ファイルに確実にインデックス情報を付加することができる。手技中はドクター及びナースは相互の連携を図るために、手技に関する会話を交すことが一般的であり、この会話を元にした音声解析によって、手技のフェーズを推定することができる。即ち、装置のステータス変化や映像変化が生じない場合でもフェーズの切換を判定することが可能である。従って、インデックス情報を付加するために、術者は特別な操作等を行う必要はなく、術者が手技に集中することを妨げることなく、所望の位置にインデックス情報を付加することができる。このインデックス情報を利用することで、再生時や編集時等において、簡単に各フェーズに再生位置、編集位置を移動させることが可能である。 As described above, in the present embodiment, the transition to each phase of the procedure is determined by analyzing the voice emitted by the operator or the like, and index information can be reliably added to the image file to be recorded. During the procedure, the doctor and the nurse generally exchange conversations about the procedure in order to cooperate with each other, and the phase of the procedure can be estimated by speech analysis based on the conversation. That is, it is possible to determine the switching of the phase even when there is no status change or video change of the device. Therefore, in order to add the index information, the operator does not need to perform a special operation or the like, and the index information can be added to a desired position without preventing the operator from focusing on the procedure. By using this index information, it is possible to easily move the reproduction position and the editing position to each phase at the time of reproduction, editing and the like.
(第2の実施の形態)
 図7は本発明の第2の実施の形態に採用される動作フローを示すフローチャートである。本実施の形態におけるハードウェア構成は図1と同様である。第1の実施の形態においても、ルックアップテーブルは、各手技に対応したものが採用された。本実施の形態は、記憶部60に各手技に対応した複数のルックアップテーブルを格納し、手技に応じてルックアップテーブルを選択可能にしたものである。更に、本実施の形態においては、手技だけでなく、術者に応じたルックアップテーブルを用意し、手技毎及び術者毎のルックアップテーブルを用いて、各フェーズを推定し判定するものである。
Second Embodiment
FIG. 7 is a flowchart showing an operation flow employed in the second embodiment of the present invention. The hardware configuration in the present embodiment is the same as that shown in FIG. Also in the first embodiment, the lookup table corresponding to each procedure is adopted. In the present embodiment, a plurality of lookup tables corresponding to each procedure are stored in the storage unit 60, and the lookup table can be selected according to the procedure. Furthermore, in the present embodiment, a look-up table not only for the procedure but also for the operator is prepared, and each phase is estimated and determined using the lookup table for each procedure and each operator. .
 上述したように、音声取得部53は、マイク毎、即ち、術者毎に音声を取得することができる。音声解析部58は、音声取得部53からマイク毎の音声信号が入力されることにより、マイク毎に音声解析結果を取得することができる。マイクとマイクの使用者との対応関係を登録したテーブルが図示しないメモリに記憶されている場合には、制御部51からの情報に基づいて、判定部59は、フェーズの判定に用いる語句等の話者を特定することができる。 As described above, the sound acquisition unit 53 can acquire sound for each microphone, that is, for each operator. The voice analysis unit 58 can obtain a voice analysis result for each microphone by inputting a voice signal for each microphone from the voice acquisition unit 53. When the table in which the correspondence between the microphone and the user of the microphone is registered is stored in the memory (not shown), the determination unit 59 determines the phase used in the determination of the phase based on the information from the control unit 51. The speaker can be identified.
 なお、音声解析部58は、音声認識処理に際してユーザ毎の個人辞書を用いることにより音声認識精度を向上させることができると共に、マイクを特定することなく、音声認識処理によって発話したユーザを特定することも可能である。 The voice analysis unit 58 can improve voice recognition accuracy by using a personal dictionary for each user at the time of voice recognition processing, and specify the user who uttered by the voice recognition processing without specifying a microphone. Is also possible.
 記憶部60には、ルックアップテーブルとして手技毎及び術者毎のルックアップテーブルが記憶されている。なお、記憶部60には、手技毎又は術者毎のいずれか一方のルックアップテーブルのみが記憶される場合もある。判定部59は、音声認識結果に基づいて手技を特定すると共に、制御部51からの情報又は音声認識処理の過程で取得した情報に基づいて術者を特定する。判定部59は、手技毎及び術者毎のルックアップテーブルを用いて、フェーズの判定を行う。これにより、フェーズの推定及び判定精度を向上させることができる。なお、術者毎のルックアップテーブルは、術者の国語や口癖に対応したものであり、例えば、術者が英語を話す場合には英語が登録される。 The storage unit 60 stores a lookup table for each procedure and each operator as a lookup table. The storage unit 60 may store only one look-up table for each procedure or each operator. The determination unit 59 identifies the procedure based on the voice recognition result, and identifies the operator based on the information from the control unit 51 or the information acquired in the process of the voice recognition process. The determination unit 59 determines the phase using a look-up table for each procedure and each operator. Thereby, the estimation and determination accuracy of the phase can be improved. Note that the look-up table for each operator corresponds to the language and / or whistling of the operator. For example, when the operator speaks English, English is registered.
 図7のフローは、図4のステップS2,S5を省略すると共に、ステップS21~S23を追加した点が図4のフローと異なる。ステップS1において、医療画像の記録が開始されると、ステップS21において、音声解析部58は音声認識を行い、音声解析部58又は制御部51は音声認識の対象となった語句を発した術者を特定する。 The flow of FIG. 7 differs from the flow of FIG. 4 in that steps S2 and S5 of FIG. 4 are omitted and steps S21 to S23 are added. In step S1, when the recording of the medical image is started, in step S21, the voice analysis unit 58 performs voice recognition, and the voice analysis unit 58 or the control unit 51 generates an operator's word that is the target of voice recognition. Identify
 判定部59は、ステップS22において、音声認識結果に基づいて手技を特定し、記憶部60に記憶されているルックアップテーブルのうち特定した手技に対応したルックアップテーブルを選択する。なお、手技毎で且つ術者毎のルックアップテーブルが記憶部60に記憶されている場合には、特定された手技で且つ術者毎のルックアップテーブル群が選択されることになる。 In step S22, the determination unit 59 identifies the procedure based on the voice recognition result, and selects a lookup table corresponding to the identified procedure among the lookup tables stored in the storage unit 60. When the look-up table for each procedure and for each operator is stored in the storage unit 60, the look-up table group for each procedure and the specified procedure is selected.
 図8及び図9は記憶部60に記憶されている2つの手技のルックアップテーブルの一例を示す説明図である。図8のルックアップテーブル(LUT1)は、肝切除の手技に対応したものであり、このルックアップテーブルを用いることで、術者等の「開始」又は「始める」の発話によってフェーズ(Phase1)が判定され、「切除開始」、「切離開始」、…の発話によって、順次フェーズ(Phase2)、フェーズ(Phase3)、…が判定される。また、図9のルックアップテーブル(LUT2)は、胆管結石除去の手技に対応したものであり、このルックアップテーブルを用いることで、術者等の「開始する」又は「スタート」の発話によってフェーズ(Phase1)が判定され、「結石確認」、「ステント挿入」、…の発話によって、順次フェーズ(Phase2)、フェーズ(Phase3)、…が判定される。 8 and 9 are explanatory diagrams showing an example of look-up tables of two procedures stored in the storage unit 60. FIG. The look-up table (LUT 1) in FIG. 8 corresponds to the procedure of liver resection. By using this look-up table, the phase (Phase 1) is generated by the operator's "start" or "start" speech. It is determined, and the phases "Phase 2", "Phase 3",... Are sequentially determined by the utterance "cut start", "cut start",. Also, the look-up table (LUT 2) in FIG. 9 corresponds to the procedure for bile duct stone removal, and by using this look-up table, the phase is started by the operator's etc. "start" or "start" utterance. (Phase 1) is determined, and the phases "Phase 2", "Phase 3",... Are sequentially determined by the utterance of "calculus confirmation", "stent insertion",.
 例えば、ドクターが「肝切除」と発することによって、判定部59はステップS22においてルックアップテーブルLUT1を選択する。なお、判定部59は音声認識結果によって手技を特定するだけでなく、例えば操作部61の操作に応じた制御部51の手技の指定によって、手技を特定するようになっていてもよい。また、図示しない外部のデータベースサーバから手技に関する情報が得られる場合には、判定部59は、当該情報を取得した制御部51の指定に基づいて手技を特定するようになっていてもよい。 For example, when the doctor issues “Hepatectomy”, the determination unit 59 selects the lookup table LUT1 in step S22. The determination unit 59 may specify a procedure not only by specifying the procedure based on the voice recognition result, but also by specifying the procedure of the control unit 51 according to the operation of the operation unit 61, for example. In addition, when information on the procedure is obtained from an external database server (not shown), the determination unit 59 may specify the procedure based on the designation of the control unit 51 that has acquired the information.
 手技の特定の後、ドクター等が「開始」と発話することで、判定部59は、ステップS23において、当該ドクターを特定し、特定したドクターに対応したルックアップテーブルが存在する場合にそのルックアップテーブルを選択し、存在しない場合には、術者に共通の「肝切除」に対応するルックアップテーブルを選択して、フェーズの判定を行う。 After the specification of the procedure, the doctor or the like utters “start”, and the determination unit 59 specifies the doctor in step S23, and if there is a look-up table corresponding to the specified doctor, the look-up table A table is selected, and if it does not exist, a look-up table corresponding to “Hepatectomy” common to the operator is selected to perform phase determination.
 以後のステップS3,S4,S6-S8の処理は第1の実施の形態と同様である。 The processing of the subsequent steps S3, S4 and S6-S8 is the same as that of the first embodiment.
 このように本実施の形態においては、第1の実施の形態と同様の効果が得られると共に、手技後や術者毎のルックアップテーブルを参照してフェーズを判定しており、判定精度を向上させることができるという効果がある。 As described above, in this embodiment, the same effect as that of the first embodiment can be obtained, and the phase is determined with reference to the look-up table after the procedure or for each operator, thereby improving the determination accuracy. It has the effect of being able to
(第3の実施の形態)
 図10は本発明の第3の実施の形態に採用される動作フローを示すフローチャートである。図10において図4と同一の手順には同一符号を付して説明を省略する。本実施の形態におけるハードウェア構成は図1と同様である。第1及び第2の実施の形態においては、ルックアップテーブルは、各手技の時系列順のフェーズのみが登録されている例を示した。本実施の形態は、ルックアップテーブルとして、時系列順のフェーズだけでなく、任意のタイミングで発生するフェーズも登録されている場合の例を示している。
Third Embodiment
FIG. 10 is a flowchart showing an operation flow employed in the third embodiment of the present invention. In FIG. 10, the same steps as in FIG. 4 will be assigned the same reference numerals and descriptions thereof will be omitted. The hardware configuration in the present embodiment is the same as that shown in FIG. In the first and second embodiments, the look-up table shows an example in which only the time-sequential phase of each procedure is registered. The present embodiment shows an example in the case where not only the phase in chronological order but also the phase occurring at an arbitrary timing is registered as the lookup table.
 図11は本実施の形態において採用されるルックアップテーブルの内容の一例を示す説明図である。図11のルックアップテーブルは、記憶部60に記憶されており、図3と同様に、手技の各フェーズ(Phase1,Phase2,…)と対応する語句との関係を登録したものである。更に、図11のルックアップテーブルは、非時系列の欄の丸印にて示すように、時系列に発生するとは限らないフェーズが登録されている。図11の例では、時系列に発生するとは限らないフェーズの例として、出血フェーズ、止血フェーズ及び洗浄フェーズが登録されている。これらのフェーズは、それぞれ音声認識結果が「出血」、「止血」、「洗浄」である場合に、そのフェーズに移行したものと判定されることになる。 FIG. 11 is an explanatory drawing showing an example of the content of the look-up table employed in the present embodiment. The look-up table of FIG. 11 is stored in the storage unit 60, and, as in FIG. 3, registers the relationship between each phase of the procedure (Phase1, Phase2,...) And the corresponding word. Furthermore, in the look-up table of FIG. 11, as indicated by the non-time-series column circles, phases that do not necessarily occur in time-series are registered. In the example of FIG. 11, the hemorrhage phase, the hemostasis phase, and the cleansing phase are registered as an example of phases that do not necessarily occur in time series. In these phases, when the speech recognition result is "bleeding", "hemostasis" or "washing", it is determined that the phase has been entered.
 判定部59の手技フェーズ推定部59aは、図10のステップS3において、音声認識結果に対応するフェーズが存在する否かを判定し、存在する場合には処理をステップS31に移行する。判定部59は、ステップS31において、音声認識結果に基づくフェーズが時系列に発生するフェーズであるか否かを判定する。例えば、「切離開始」に対応したフェーズ(Phase3)のように、時系列に発生するフェーズが推定された場合には、判定部59は、次のステップS4において正しい時系列のフェーズであるか否かを判定する。 The procedure phase estimation unit 59a of the determination unit 59 determines whether or not there is a phase corresponding to the speech recognition result in step S3 of FIG. 10, and if it exists, shifts the process to step S31. In step S31, determination unit 59 determines whether or not the phase based on the speech recognition result is a phase that occurs in time series. For example, when the phase occurring in the time series is estimated as in the phase (Phase 3) corresponding to “separation start”, the determination unit 59 determines whether the phase is the correct time series phase in the next step S 4 It is determined whether or not.
 一方、例えば、「止血」に対応した止血フェーズのように、時系列に発生するとは限らないフェーズが推定された場合には、判定部59は、ステップS31から処理をステップS5に移行し、時系列の判定を行うことなく話者の特定を行う。 On the other hand, for example, when a phase that does not necessarily occur in time series is estimated as in the hemostasis phase corresponding to “hemostasis”, the determination unit 59 shifts the process from step S31 to step S5, and Speaker identification is performed without judging the sequence.
 他の作用は第1の実施の形態と同様である。これにより、本実施の形態においては、「止血」等の任意のタイミングで発生する可能性があるフェーズについても、確実に当該フェーズの発生に対応したインデックス情報を記録データに付加することができる。 The other actions are the same as in the first embodiment. Thus, in the present embodiment, index information corresponding to the occurrence of the phase can be reliably added to the recording data even in the phase that may occur at an arbitrary timing such as “hemostasis”.
 このように本実施の形態においては、第1の実施の形態と同様の効果が得られると共に、時系列に発生するとは限らないフェーズであっても確実に判定して、インデックス情報を記録データに付加することができるという効果がある。 As described above, in this embodiment, the same effect as that of the first embodiment can be obtained, and even in a phase that does not necessarily occur in time series, it is determined with certainty that index information is recorded data. There is an effect that it can be added.
 なお、本実施の形態は上記第1の実施の形態に適用する例を説明したが、手技毎及び術者毎のルックアップテーブルに非時系列のフェーズであるか否かの情報を含めることにより、第2の実施の形態にも同様に適用可能であることは明らかである。 Although the present embodiment has been described by way of example applied to the first embodiment, by including information on whether or not it is a non-time-sequential phase in the look-up table for each procedure and for each operator. It is apparent that the present invention is similarly applicable to the second embodiment.
 また、上記各実施の形態においては、音声解析部58及び判定部59において、音声認識及びフェーズの判定を行ったが、これらの処理を人工知能を用いて行ってもよい。例えば、人工知能の機能を有する外部サーバに収音音声を与えて、人工知能により会話の内容を判定してフェーズを判定し、判定結果を制御部において取り込んで、インデックス情報を記録画像に付加するようにしてもよい。 In each of the above embodiments, the speech analysis unit 58 and the determination unit 59 perform the speech recognition and the phase determination. However, these processes may be performed using artificial intelligence. For example, the collected voice is provided to an external server having an artificial intelligence function, the content of the conversation is determined by the artificial intelligence, the phase is determined, the determination result is taken in by the control unit, and the index information is added to the recorded image. You may do so.
 本発明は、上記各実施形態にそのまま限定されるものではなく、実施段階ではその要旨を逸脱しない範囲で構成要素を変形して具体化できる。また、上記各実施形態に開示されている複数の構成要素の適宜な組み合わせにより、種々の発明を形成できる。例えば、実施形態に示される全構成要素の幾つかの構成要素を削除してもよい。さらに、異なる実施形態にわたる構成要素を適宜組み合わせてもよい。 The present invention is not limited to the above-described embodiments as it is, and at the implementation stage, the constituent elements can be modified and embodied without departing from the scope of the invention. In addition, various inventions can be formed by appropriate combinations of a plurality of components disclosed in the above-described embodiments. For example, some components of all the components shown in the embodiment may be deleted. Furthermore, components in different embodiments may be combined as appropriate.
 本出願は、2017年11月14日に日本国に出願された特願2017-219040号を優先権主張の基礎として出願するものであり、上記の開示内容は、本願明細書、請求の範囲に引用されるものとする。 This application is based on Japanese Patent Application No. 2017-219040, filed on Nov. 14, 2017, as the basis for claiming priority, and the above disclosure is made with the present specification and claims. It shall be quoted.

Claims (7)

  1.  医療画像を取得する映像入力部と、
     医療従事者の発生する音声を取得する音声取得部と、
     前記音声取得部が取得した音声を解析する音声解析部と、
     前記音声解析部の解析結果による語句に基づいて、手技のフェーズを推定する手技フェーズ推定部と、
     前記手技フェーズ推定部により推定されたフェーズの変化と予め設定されているフェーズの時系列との比較に基づいて、推定されたフェーズを確定する判定部と、
     前記判定部の判定結果に応じて、前記医療画像と前記フェーズとの関係を対応付けるインデックス情報を前記医療画像に付与して記録データを生成して記録部に記録する記録データ生成部とを具備したことを特徴とする画像記録装置。
    A video input unit for acquiring a medical image,
    A voice acquisition unit for acquiring a voice generated by a medical worker,
    A voice analysis unit that analyzes the voice acquired by the voice acquisition unit;
    A procedure phase estimation unit that estimates the phase of the procedure based on the word / phrase based on the analysis result of the voice analysis unit;
    A determination unit that determines the estimated phase based on a comparison between the change of the phase estimated by the procedure phase estimation unit and the time series of the preset phase;
    According to the determination result of the determination unit, there is provided a recording data generation unit that adds recording information to the medical image by adding index information that associates the relationship between the medical image and the phase to the medical image. An image recording apparatus characterized by
  2.  前記手技フェーズ推定部は、前記音声解析部の解析結果により得られた語句と対応するフェーズとが関連付けられたテーブルを参照することで、前記フェーズを推定することを特徴とする請求項1に記載の画像記録装置。 2. The method according to claim 1, wherein the procedure phase estimation unit estimates the phase by referring to a table in which words and phrases obtained by the analysis result of the voice analysis unit are associated with the corresponding phase. Image recording device.
  3.  前記テーブルは、手技及び医療従事者の少なくとも一方毎に設定されることを特徴とする請求項2に記載の画像記録装置。 The image recording apparatus according to claim 2, wherein the table is set for each of at least one of a procedure and a medical worker.
  4.  前記判定部は、
     前記推定されたフェーズが手技中の任意のタイミングで発生するものであると判定した場合には、前記フェーズの変化と前記フェーズの時系列との比較を行うことなく、推定されたフェーズを確定することを特徴とする請求項1に記載の画像記録装置。
    The determination unit is
    If it is determined that the estimated phase occurs at an arbitrary timing during the procedure, the estimated phase is determined without comparing the change of the phase with the time series of the phase. The image recording apparatus according to claim 1,
  5.  前記記録データ生成部は、前記インデックス情報を前記記録データのメタデータとして前記記録部に記録することを特徴とする請求項1に記載の画像記録装置。 The image recording apparatus according to claim 1, wherein the recording data generation unit records the index information as metadata of the recording data in the recording unit.
  6.  前記記録データ生成部は、前記インデックス情報を前記記録データとは別のデータとして前記記録部に記録することを特徴とする請求項1に記載の画像記録装置。 The image recording apparatus according to claim 1, wherein the recording data generation unit records the index information in the recording unit as data different from the recording data.
  7.  医療画像を取得する映像入力ステップと、
     医療従事者の発生する音声を取得する音声取得ステップと、
     前記音声取得ステップにおいて取得した音声を解析する音声解析ステップと、
     前記音声解析ステップにおける解析結果による語句に基づいて、手技のフェーズを推定する手技フェーズ推定ステップと、
     前記手技フェーズ推定ステップにおいて推定されたフェーズの変化と予め設定されているフェーズの時系列との比較に基づいて、推定されたフェーズを確定する判定ステップと、
     前記判定ステップにおける判定結果に応じて、前記医療画像と前記フェーズとの関係を対応付けるインデックス情報を前記医療画像に付与して記録データを生成して記録部に記録する記録データ生成ステップとを具備したことを特徴とする画像記録方法。
    A video input step for acquiring a medical image;
    A voice acquisition step of acquiring a voice generated by a medical worker;
    A speech analysis step of analyzing the speech acquired in the speech acquisition step;
    A procedure phase estimation step of estimating the phase of the procedure based on the words and phrases according to the analysis result in the voice analysis step;
    A determination step of determining an estimated phase based on a comparison between a change in the phase estimated in the procedure phase estimation step and a time series of a preset phase;
    And recording data generation step of generating recording data and recording the recording data on a recording unit by adding index information correlating the relationship between the medical image and the phase to the medical image according to the determination result in the determining step. An image recording method characterized by
PCT/JP2018/031838 2017-11-14 2018-08-28 Image recording device and image recording method WO2019097804A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2017219040 2017-11-14
JP2017-219040 2017-11-14

Publications (1)

Publication Number Publication Date
WO2019097804A1 true WO2019097804A1 (en) 2019-05-23

Family

ID=66538992

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2018/031838 WO2019097804A1 (en) 2017-11-14 2018-08-28 Image recording device and image recording method

Country Status (1)

Country Link
WO (1) WO2019097804A1 (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006164251A (en) * 2004-11-09 2006-06-22 Toshiba Corp Medical information system, medical information system program, and medical information processing method for performing information processing for management of medical practice
JP2011167301A (en) * 2010-02-17 2011-09-01 Asahikawa Medical College Surgery video image accumulating device, surgery video image accumulating method, and program

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006164251A (en) * 2004-11-09 2006-06-22 Toshiba Corp Medical information system, medical information system program, and medical information processing method for performing information processing for management of medical practice
JP2011167301A (en) * 2010-02-17 2011-09-01 Asahikawa Medical College Surgery video image accumulating device, surgery video image accumulating method, and program

Similar Documents

Publication Publication Date Title
JP5690450B2 (en) Image recording device
EP2682050B1 (en) Medical information recording apparatus
US20180092509A1 (en) Image recording device
WO2019181432A1 (en) Operation assistance system, information processing device, and program
WO2018163600A1 (en) Medical information management device, medical information management method, and medical information management system
JP2004181229A (en) System and method for supporting remote operation
WO2012165381A1 (en) Medical information recording device
JP2011036372A (en) Medical image recording apparatus
US10275905B2 (en) Color integration system for medical images and recording and color management apparatus for medical images
US11483515B2 (en) Image recording and reproduction apparatus, image recording method, and endoscope system
JPWO2015114901A1 (en) Medical video recording / playback system and medical video recording / playback device
JP2011036370A (en) Medical image recording apparatus
US20220008161A1 (en) Information processing device, presentation method, and surgical system
JPWO2017187676A1 (en) Control device, control method, program, and sound output system
WO2019097804A1 (en) Image recording device and image recording method
WO2014196292A1 (en) Medical assistance device and method for processing setting information for medical equipment by scene
US20070083480A1 (en) Operation information analysis device and method for analyzing operation information
JP2005124824A (en) Operation supporting system
JP7451707B2 (en) Control device, data log display method, and medical centralized control system
US20230149100A1 (en) Control apparatus, medical central control system, and surgery-related information display method
WO2022201800A1 (en) Surgical operation room system, image recording method, program, and medical information processing system
JP5551522B2 (en) Medical information recording device
JP2018175757A (en) Image recording regeneration device and endoscope system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18878461

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18878461

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP