WO2015083411A1 - 情報処理装置、情報処理方法、およびプログラム - Google Patents
情報処理装置、情報処理方法、およびプログラム Download PDFInfo
- Publication number
- WO2015083411A1 WO2015083411A1 PCT/JP2014/074257 JP2014074257W WO2015083411A1 WO 2015083411 A1 WO2015083411 A1 WO 2015083411A1 JP 2014074257 W JP2014074257 W JP 2014074257W WO 2015083411 A1 WO2015083411 A1 WO 2015083411A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- section
- sleep
- user
- sensor
- information processing
- Prior art date
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/48—Other medical applications
- A61B5/4806—Sleep evaluation
- A61B5/4809—Sleep detection, i.e. determining whether a subject is asleep or not
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0002—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
- A61B5/0015—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network characterised by features of the telemetry system
- A61B5/0022—Monitoring a patient using a global network, e.g. telephone networks, internet
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1116—Determining posture transitions
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/48—Other medical applications
- A61B5/4806—Sleep evaluation
- A61B5/4812—Detecting sleep stages or cycles
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/48—Other medical applications
- A61B5/4806—Sleep evaluation
- A61B5/4815—Sleep quality
Definitions
- the present disclosure relates to an information processing apparatus, an information processing method, and a program.
- Patent Document 1 describes a technique for presenting an evaluation regarding sleep quality to a user based on a measurement result of a biological signal during sleep.
- Patent Document 1 In recent years, with the downsizing of sensor devices, it is becoming possible for users to wear sensor devices on a daily basis, including during sleep, and acquire activity logs.
- the technique described in Patent Document 1 is based on the assumption that the user is in a sleep state. That is, these techniques analyze what a user's sleep state is, and do not determine whether or not the user is in a sleep state. Therefore, it cannot be said to be sufficient to detect or analyze the sleep state together with the daily activities of the user.
- the present disclosure proposes a new and improved information processing apparatus, information processing method, and program capable of detecting or analyzing a sleep state together with a user's daily activities.
- the stable section detection unit that detects the stable section of the posture of the sensor device, and the second sensor data.
- a specific motion section detecting unit that detects a specific motion section in which the user's specific motion has occurred, and detecting a sleep section in which the user is in a sleep state based on a relationship between the stable section and the specific motion section.
- an information processing device including a sleep section detecting unit.
- position of the said sensor apparatus is detected, Based on 2nd sensor data And detecting a specific motion section in which the user's specific motion has occurred, and a processor detecting a sleep section in which the user is in a sleep state based on a relationship between the stable section and the specific motion section.
- the sleep state can be detected or analyzed together with the daily activities of the user.
- FIG. 1 is a block diagram illustrating a schematic configuration of a system according to a first embodiment of the present disclosure. It is a figure for demonstrating the principle of the sleep area detection in 1st Embodiment of this indication. It is a block diagram showing a schematic functional composition of an information processor which can perform sleep section detection in a 1st embodiment of this indication. It is a figure for demonstrating the processing flow of the sleep area detection in 1st Embodiment of this indication. It is a block diagram showing a schematic structure of a system concerning a 2nd embodiment of this indication. It is a figure for demonstrating use of the selective detector in 4th Embodiment of this indication.
- FIG. 3 is a block diagram illustrating a hardware configuration example of an information processing apparatus according to an embodiment of the present disclosure.
- FIG. 1 is a block diagram illustrating a schematic configuration of a system according to the first embodiment of the present disclosure.
- the system 10 includes a sensor device 100 and a smartphone 200.
- the sensor device 100 includes a sensor 110, a preprocessing unit 120, and a memory 130.
- the smartphone 200 includes a position detection unit 210, a sensor 220, an action recognition unit 230, a processing unit 240, an integrated analysis unit 250, a storage 260, and an application 270.
- the sensor device 100 is a wearable device.
- the sensor device 100 is directly attached to the user by, for example, wrapping around a wrist, ankle, finger, or the like.
- the sensor device 100 may be indirectly attached to the user by being fixed to clothes using a clip or the like.
- the sensor device 100 is a device that a user wears, the sensor device 100 is not always worn by the user.
- the sensor device 100 may be removed during bathing or during preparation. Therefore, as will be described later, when the detection value of the acceleration sensor included in the sensor 110 does not fluctuate, when the user wears the sensor device 100 to stabilize the posture, and when the sensor device 100 is removed, There is a case where it is left still.
- the sensor device 100 stores a processor (which can implement the preprocessing unit 120) that processes sensor data acquired by the sensor 110, and sensor data or data after processing.
- a storage device (which may implement the memory 130) and a communication device (not shown) for transmitting sensor data or processed data to the smartphone 200 are included.
- the sensor device 100 including a processor may be an information processing device according to an embodiment of the present disclosure.
- the information processing apparatus can be realized using a hardware configuration example described later. Hereinafter, each component of the sensor device 100 will be further described.
- the sensor 110 includes various sensors provided in the sensor device 100, and senses the behavior of the user wearing the sensor device 100.
- the sensor 110 includes an acceleration sensor. Based on the detection value of the acceleration sensor by the acceleration sensor, it is possible to specify the posture and body movement of the user wearing the sensor device 100.
- the sensor 110 may also include other sensors such as an angular velocity sensor, a gyro sensor, an optical sensor, a sound sensor, or an atmospheric pressure sensor. The detection values of these sensors can also be used to specify the user's posture and body movement.
- the sensor 110 may include a camera for acquiring a user's image and a microphone for acquiring sound emitted by the user.
- the sensor data acquired by the sensor 110 may be temporarily stored in the memory 130 through processing by the preprocessing unit 120, or may be stored in the memory 130 as it is without performing preprocessing.
- the pre-processing unit 120 performs pre-processing of the sensor data acquired by the sensor 110. For example, the preprocessing unit 120 extracts a feature amount from the sensor data. Further, the preprocessing unit 120 may remove noise included in the sensor data or resample the sensor data.
- the sensor data processing by the preprocessing unit 120 is preferably capable of restoring raw data of sensor data or data close thereto from the processed data.
- part or all of the analysis processing executed by the processing unit 240 of the smartphone 200 described later may be executed by the preprocessing unit 120.
- the memory 130 stores sensor data that has undergone processing by the preprocessing unit 120 or raw data of sensor data that has not undergone preprocessing.
- the data stored in the memory 130 is transmitted to the smartphone 200 by a communication device (not shown).
- a communication device for communication between the sensor device 100 and the smartphone 200, for example, wireless communication such as Bluetooth (registered trademark) can be used.
- the smartphone 200 is a terminal device possessed by the user separately from the sensor device 100.
- the smartphone 200 can be realized by, for example, a hardware configuration of an information processing device described later.
- the smartphone 200 may be replaced by another terminal device that can realize the same function, such as a tablet terminal.
- each component of the smartphone 200 will be further described.
- the position detection unit 210 is realized by, for example, a GPS (Global Positioning System) receiver. In this case, the position detection unit 210 acquires the position information of the smartphone 200 by receiving radio waves from the satellite. Alternatively, the position detection unit 210 may be realized by a communication device that performs wireless communication such as Wi-Fi. In this case, the position detection unit 210 acquires the position information of the smartphone 200 based on the position information of the wireless communication base station and the reception state of the radio wave from the base station.
- GPS Global Positioning System
- the sensor 220 may include various sensors such as an acceleration sensor, an angular velocity sensor, a gyro sensor, an optical sensor, a sound sensor, an atmospheric pressure sensor, a camera, or a microphone, similar to the sensor 110 of the sensor device 100.
- the smartphone 200 is placed on a desk during the user's work, placed in a bag while moving, or in bed while sleeping. It is placed nearby. Therefore, the sensor 110 of the sensor device 100 is mainly used to acquire information indicating the posture and movement of the user.
- the smartphone 200 allows a certain amount of volume and weight, and therefore the sensor 220 has more types than the sensor 110. It is possible to include a sensor or a sensor with higher accuracy than the sensor 110.
- the behavior recognition unit 230 is realized by a processor, for example.
- the behavior recognition unit 230 recognizes the user's behavior based on the position information of the smartphone 200 detected by the position detection unit 210 and / or the sensor data provided from the sensor 220.
- the user's behavior recognized by the behavior recognition unit 230 is used in combination with the data analysis result from the sensor device 100.
- the user's behavior recognized by the behavior recognition unit 230 may be used by the application 270 or the like separately from the data analysis result from the sensor device 100.
- the processing unit 240 is realized by a processor, for example.
- the processing unit 240 performs an analysis process on data received from the sensor device 100 by a communication device (not shown). For example, the processing unit 240 detects a sleep section in which a user wearing the sensor device 100 is in a sleep state by analyzing sensor data of the sensor 110. Details of the sleep section detection process will be described later. As described above, part or all of the analysis processing executed by the processing unit 240 may be realized by being distributed to the preprocessing unit 120 of the sensor device 100.
- the integrated analysis unit 250 is realized by a processor, for example.
- the integrated analysis unit 250 integrates the result of the user's behavior recognition by the behavior recognition unit 230 and the detection result of the user's sleep section by the processing unit 240. In this way, by integrating the results of user behavior recognition based on different data, for example, supplementing the missing user behavior in each result or adjusting the results to match each other results. Accuracy can be improved, or erroneous detection can be prevented.
- the result of user behavior recognition integrated by the integrated analysis unit 250 is stored in the storage 260. Apart from this, the analysis result of the user's behavior by the behavior recognition unit 230 and the detection result of the sleep section by the analysis processing of the processing unit 240 may also be stored in the storage 260.
- the storage 260 is realized by, for example, a memory or a storage device.
- the storage 260 accumulates the results of user behavior recognition obtained by the processing of the sensor device 100 and the smartphone 200. For example, the user's behavior log, position information log, number of steps, sleep time log, and the like are accumulated in the storage 260. Such data is generated based on, for example, the result of action recognition in the action recognition unit 230 and the detection result of the sleep section in the processing unit 240. Logs temporarily or persistently stored in the storage 260 are used by the application 270.
- the application 270 is application software that is executed on the smartphone 200 and uses a user action log.
- the application 270 is executed by the processor of the smartphone 200, and uses an output device such as a display or a speaker, or an input device such as a touch panel, as necessary.
- the application 270 may transmit a control signal to an external device via a communication device, or may transmit information to another user. A specific example of the application 270 will be described later.
- FIG. 2 is a diagram for describing the principle of sleep interval detection according to the first embodiment of the present disclosure.
- FIG. 2 shows a change over time of the detection value of the triaxial acceleration sensor included in the sensor 110 of the sensor device 100.
- the time of the illustrated example will be described by dividing it into sections P1 to P7.
- the detection value of the acceleration sensor continuously fluctuates greatly. In this section, it is estimated that the user is performing various activities while waking up. On the other hand, in the sections P2 and P6, the detection value of the acceleration sensor does not change and is constant. In the section P4, it is repeated that the detected value of the acceleration sensor is constant for a while and then fluctuates in a short time and becomes constant for a while.
- the user wears the sensor device 100 to stabilize the posture, and the sensor device 100 is removed and left still somewhere. There may be cases. In order to detect a user's sleep period, it is necessary to distinguish between these cases.
- the inventors focused on the fact that when the user is in a sleep state, the user's body is not always stationary, but occasionally becomes stationary again after the posture changes. Such user behavior is observed, for example, as turning over. Rolling over is a very natural behavior seen by many users.
- the user's sleep state is divided into a first section in which the user's posture is unchanged (a section in which the detected value of the acceleration sensor is not changed and constant), and a second section in which the user's posture is changed ( It is defined as a state in which the detected value of the acceleration sensor fluctuates and is sufficiently shorter than the first interval).
- the detection value of the acceleration sensor remains constant and does not change, and then the sensor device 100 is attached to the user again, and the detection value continues. Fluctuations are expected to resume. Therefore, the case where the user wears the sensor device 100 and stabilizes the posture and the case where the sensor device 100 is removed and left still somewhere are the second before and after the first section. Can be identified by whether or not there is a section.
- the detected value of the acceleration sensor is constant without being changed, but the detected values fluctuate before and after that (sections P1, P3). , P5, P7), the detected values continuously fluctuate, and the lengths of these sections are not necessarily shorter than those of the sections P2, P6. Therefore, in these sections, it is presumed that the user did not wear the sensor device 100 to stabilize the posture, but the sensor device 100 was removed and left still somewhere.
- a first section where the detection value of the acceleration sensor is constant without fluctuation and a second section where the detection value of the acceleration sensor fluctuates in a relatively short time between them are repeated. Therefore, in the section P4, it is presumed that the user was wearing the sensor device 100 to stabilize the posture.
- the fact that the user wears the sensor device 100 and stabilizes the posture is not the same as that the user is in a sleep state.
- the user's posture is stable even when the user is watching TV or a book, it is considered that the time during which the user is stable is shorter than that when the user is in a sleep state, or the posture is slightly changed.
- the average value or dispersion value of acceleration for determining that the posture of the sensor device is in a stable state is appropriately set, or the stable state is longer than a predetermined length.
- the orientation of the sensor device 100 may identify whether the user is in a sleep state or in another stable state. .
- the state where the user is lying down but not falling asleep is also conceivable, in such a case, the user may be considered to be in a sleeping state.
- FIG. 3 is a block diagram illustrating a schematic functional configuration of the information processing apparatus capable of performing sleep interval detection according to the first embodiment of the present disclosure.
- the information processing device 300 includes a sensor data acquisition unit 301, a stable interval detection unit 303, a specific motion interval detection unit 305, a sleep interval detection unit 307, an external device control unit 309, and notification information. And an output unit 311.
- the information processing device 300 is realized by a processor included in the sensor device 100 or the smartphone 200.
- the information processing apparatus 300 may be realized in the smartphone 200.
- the sensor data acquisition unit 301, the stable section detection unit 303, the specific action section detection unit 305, and the sleep section detection unit 307 correspond to the processing unit 240.
- the external device control unit 309 and the notification information output unit 311 correspond to the application 270.
- the sleep section detection unit 307 may correspond to both the processing unit 240 and the integrated analysis unit 250.
- the information processing device 300 may be realized by being distributed between the sensor device 100 and the smartphone 200.
- the sensor data acquisition unit 301, the stable section detection unit 303, the specific motion section detection unit 305, and the sleep section detection unit 307 are sensor devices.
- the external device control unit 309 and the notification information output unit 311 correspond to the application 270 of the smartphone 200.
- the configuration for realizing the information processing apparatus 300 is not limited to the above example, and various other examples are possible within the scope apparent to those skilled in the art from the description in this specification.
- all of the functions of the information processing device 300 may be realized in the sensor device 100.
- the sleep section detection unit 307 includes the preprocessing unit 120 of the sensor device 100, the processing unit 240 of the smartphone 200, and / or Alternatively, it can be realized by the integrated analysis unit 250.
- the sensor data acquisition unit 301 can be realized by a processor as a software interface for acquiring sensor data provided from the sensor device 100.
- the sensor data is output by various sensors such as an acceleration sensor, an angular velocity sensor, a gyro sensor, an optical sensor, a sound sensor, an atmospheric pressure sensor, a camera, or a microphone included in the sensor 110, and the pre-processing unit 120 according to need. It can be processed data.
- the sensor data acquired by the sensor data acquisition unit 301 includes first sensor data used by the stable section detection unit 303 and second sensor data used by the specific operation section detection unit 305. . These data may be the same data, for example, a detection value of a common acceleration sensor, or different data.
- the stable section detection unit 303 detects the stable section of the posture of the sensor device 100 based on the first sensor data acquired by the sensor data acquisition unit 301.
- the stable section is a section that is a candidate for the first section described with reference to FIG.
- the stable section of the posture of the sensor device 100 can be defined as a section in which the posture of the sensor device 100 is unchanged as in the present embodiment, for example.
- the stable interval detection unit 303 detects the stable interval as an interval in which the detection value does not vary.
- the stable section may be limited to a section longer than a predetermined time (for example, several minutes).
- the stable section detected here is a stable section of the posture of the sensor device 100, whether the user wears the sensor device 100 to stabilize the posture, the sensor device Whether 100 is removed and left still is not specified only by the detection result of the stable section. Therefore, in the present embodiment, detection of a specific motion section is performed by the specific motion section detection unit 305 described below.
- the specific operation section detection unit 305 detects a specific operation section in which a specific operation of the user has occurred based on the second sensor data acquired by the sensor data acquisition unit 301.
- the specific operation section is a section corresponding to the second section described with reference to FIG.
- the second sensor data includes a detection value of the acceleration sensor, similarly to the first sensor data.
- the specific operation section is a section in which the posture of the user has changed.
- the possibility that the sensor device 100 is removed in a section other than the stable section (including the specific operation section) is Low enough to be ignored. Therefore, in the present embodiment, it is possible to regard a change in the posture of the sensor device 100 indicated by the second sensor data as a change in the posture of the user outside the stable period.
- the specific motion section detection unit 305 occurs after (1) a stable section having a certain length (for example, several minutes or more), and (2) compared with the stable section. Identified on the condition that the detected value of the acceleration sensor changes in a sufficiently short time, and (3) another stable interval in which the detected value of the acceleration sensor stabilizes at a value different from the previous stable interval occurs. An operation interval may be detected.
- the sleep interval detection unit 307 determines the sleep interval in which the user was in the sleep state based on the relationship between the stable interval detected by the stable interval detection unit 303 and the specific operation interval detected by the specific operation interval detection unit 305. To detect. In the present embodiment, the sleep interval detection unit 307 detects a sleep interval including the stable interval and the specific operation interval when the stable interval and the specific operation interval are repeated.
- the section P4 corresponds to the sleep section, but in this section P4, the first section (stable section) in which the detected value of the acceleration sensor does not vary and is relatively short, and the interval is relatively short.
- the second section (specific operation section) in which the detection value of the acceleration sensor varies with time is repeated.
- the sleep interval detection unit 307 includes: (1) from the first stable interval that occurs following the interval that is not the specific operation interval, (2) the specific operation interval that follows the first stable interval, and ( 3) After an intermediate stable section sandwiched between the specific motion sections before and after, (4) The last stable section that occurs after the last specific motion section and is followed by a section that is not a specific motion section is the sleep section It may be detected.
- the start time of the stable section (1) as the user's sleep start time and the end time of the stable section (4) as the user's sleep end time, respectively.
- the sleep interval detection unit 307 determines the sleep interval in consideration of the consistency with the user behavior recognized by the behavior recognition unit 230 based on the position information and the sensor data by the function of the integrated analysis unit 250 of the smartphone 200. You may decide. At this time, it can be said that the sleep section detection unit further detects the sleep section based on the user's behavior recognized based on the sensor data different from the first and second sensor data.
- Both the external device control unit 309 and the notification information output unit 311 described below have a functional configuration that uses the sleep section detection result by the sleep section detection unit 307. Only one of these functional configurations may be included in the information processing apparatus 300, or both may be included in the information processing apparatus 300. In this embodiment, since the method of using the sleep section detection result is not particularly limited, a different functional configuration from the external device control unit 309 and the notification information output unit 311 exists in order to use the sleep section detection result. May be. In addition, the detection result of the sleep section by the sleep section detection unit 307 may be stored in the storage as it is and used in a device different from the information processing device 300.
- the external device control unit 309 outputs a control signal of the external device when the sleep interval of the user is detected by the sleep interval detection unit 307. For example, when the start of a sleep interval is detected, the external device control unit 309 outputs a control signal for stopping or pausing the external device.
- the sleep interval detection unit 307 can detect the sleep interval only after detecting a specific action interval that occurs afterwards. Therefore, in this embodiment, the external device control unit 309 cannot transmit a control signal in real time when the user's sleep state is started. Therefore, the control signal transmitted by the external device control unit 309 in the present embodiment stops an external device that is considered unnecessary when the user enters a sleep state, such as a television or an air conditioner (depending on the temperature). It may be sent to pause.
- the external device control unit 309 may output a control signal for the external device to set an alarm based on the start time of the sleep interval.
- the external device here is merely an external device for the device shown as the information processing device 300 in FIG. 3, and thus other functions of the smartphone 200 when the information processing device 300 is realized in the smartphone 200, for example.
- An element including an element realized by a processor similar to the information processing apparatus 300
- the notification information output unit 311 outputs information for notifying other users of the detection result of the sleep interval by the sleep interval detection unit 307. For example, the notification information output unit 311 outputs information indicating the start, continuation, or end of the sleep interval. In this case, for example, a friend on social media can be notified that the user has woken up, slept, or slept. This notification may be a push type or a pull type. In the case of pull-type notification, for example, the user's icon expressed in the virtual space expresses the sleep state, and other users see the virtual space to recognize that the user is sleeping or waking up May be possible.
- the notification information output unit 311 may output information indicating the length of the sleep interval in a predetermined period.
- a friend or the like may be notified that the user is short of sleep depending on whether the sleep time per day is greater or less than the average.
- a notification may be output via a virtual space as in the above example, or wrinkles may be superimposed on an actual user's face using AR (Augmented Reality) technology.
- FIG. 4 is a diagram for describing a processing flow of sleep interval detection according to the first embodiment of the present disclosure.
- the processing flow shown in FIG. 4 corresponds to the functional configuration of the information processing apparatus 300 described above with reference to FIG. That is, it can be said that a series of processing executed by the feature amount calculation unit 401 to the consistency adjustment unit 413 described below expresses the function of each component of the information processing apparatus 300 from different viewpoints. Therefore, a series of processes can be executed by the sensor device 100 or the processor of the smartphone 200. Hereinafter, processing of each component of the processing flow will be further described.
- the feature amount calculation unit 401 calculates a feature amount from the sensor data.
- the calculated feature amount may be capable of being restored to raw data or data close thereto.
- the feature amount calculation unit 401 may calculate an average value and a variance value of acceleration with a predetermined time (for example, 1 minute) as a time unit.
- the stable section detection unit 403 detects a stable section in which the user's posture is estimated to be stable based on the feature amount calculated from the sensor data by the feature amount calculation unit 401. As described above, since the stable section detected here is a stable section of the posture of the sensor device 100, whether the user is wearing the sensor device 100 to stabilize the posture, the sensor device 100 is removed. However, it is not specified only by the detection result of the stable section.
- the filter processing unit 405 executes a filter process for the stable section detected by the stable section detection unit 403. For example, the filter processing unit 405 filters the detected stable interval by its length. More specifically, when the length of the section is a predetermined time (for example, 5 minutes) or less, the filter processing unit 405 excludes the section from the stable section. Considering that the stable interval is used for detecting the user's sleep interval, there may be a sleep interval with a short time, but the stable interval in such a case does not include the above-described specific operation interval, etc. For reasons, it may be difficult to distinguish from a pseudo stable interval that occurs for reasons other than sleep. Therefore, in the present embodiment, a section having a short time is excluded from the stable section.
- the specific action determination unit 407 detects a specific action section in which a specific action of the user has occurred based on the feature quantity calculated by the feature quantity calculation unit 401 from the sensor data. Further, the specific action determination unit 407 determines whether or not the stable period is a sleep period based on the relationship between the specific period and the stable period detected by the stable period detection unit 403 and filtered by the filter processing unit 405. To do. More specifically, the specific action determination unit 407, when there is a specific action section consecutively at least before or after the stable section, displays the stable section and the specific action section continuous thereto as a sleep section. May be specified.
- the result R1 indicating the sleep interval is acquired by the processing of the feature amount calculation unit 401, the stable interval detection unit 403, the filter processing unit 405, and the specific action determination unit 407 so far.
- the result R1 except that the stable interval that is too short is removed by the filter processing unit 405, the detection result of the sleep interval is not filtered, and the relationship between the purely stable interval and the specific motion interval is pure. It can be said that the sleep interval is specified based on the above.
- the result R1 may be used as an input for processing of the section combining unit 409 described below, or may be output as it is.
- the section combining unit 409 combines the sleep sections specified by the result R1 that are close to each other. More specifically, for example, the interval combining unit 409 combines these intervals when the interval between sleep intervals is equal to or shorter than a predetermined time (for example, 30 minutes).
- a predetermined time for example, 30 minutes.
- the filter processing unit 411 executes filter processing on the sleep interval after combination by the interval combination unit 409. For example, the filter processing unit 411 filters the combined sleep intervals by their lengths. More specifically, the filter processing unit 411 excludes the section from the sleep section when the length of the sleep section candidate is equal to or shorter than a predetermined time (for example, 30 minutes). There is a possibility that a sleep interval that does not reach a sufficient length even when combined is a false detection of a stable interval that occurs for reasons other than sleep, for example. Also, depending on how the detection result is used, there is no problem even if a short sleep section such as a snooze or a song is not detected. Therefore, in the present embodiment, the filter processing unit 411 excludes sleep sections that do not reach a sufficient length.
- a predetermined time serving as a threshold is arbitrarily set, and in the example described above (5 minutes and 30 minutes) Is not limited.
- the threshold time in the filter processing unit 411 (used to exclude a sleep interval that is still too short even if combined) is It may be longer than the threshold time in the filter processing unit 405 (used to exclude a stable section that is too short).
- the result R2 indicating the sleep interval is acquired by the processing of the interval combination unit 409 and the filter processing unit 411 with respect to the result R1 so far.
- the sleep intervals specified by the result R1 are combined with ones having close intervals (interval combining unit 409), and sleep intervals that do not reach a sufficient length even after being combined are excluded ( Filter processing unit 411). Therefore, in the result R2, it can be said that the weak filter using the interval and the length indicated by the detection result itself is applied to the detection result of the sleep section.
- the result R2 may be used as an input of the consistency adjustment unit 413 described below, or may be output as it is.
- the consistency adjustment unit 413 adjusts the consistency of the sleep recognition section specified by the result R2 with the action recognition result by another method. More specifically, for example, the consistency adjustment unit 413 is specified by the result of action recognition performed by the action recognition unit 230 based on the detection results of the position detection unit 210 and the sensor 220 of the smartphone 200 and the result R2. Adjust the consistency with the sleep interval. For example, in the sleep section specified by the result R2, when the action recognition unit 230 recognizes the action of “walking” (walking a longer distance, not in the house), either the sleep section or the action recognition result Or both are considered wrong.
- the consistency adjustment unit 413 may give priority to either the sleep interval or the action recognition result in a fixed manner. For example, when a sleep interval is given priority, an action such as “walking” recognized in the sleep interval is ignored. Conversely, when the action recognition result is prioritized, it is determined that the sleep section in which the action such as “walking” is recognized is erroneously detected. Or the consistency adjustment part 413 may select the result employ
- the result R3 indicating the sleep interval is acquired through the processing of the consistency adjustment unit 413 as described above.
- the result R3 it can be said that a stronger filter is applied to the sleep section specified by the result R2 in consideration of the consistency with the action recognition result by another method.
- the result R3 is output for use by the application 270 in the smartphone 200, for example.
- the result R1 and / or the result R2 can also be output together with the result R3 or instead of the result R3. Therefore, in the present embodiment, for example, the result R1 is output and the processing of the section combining unit 409 to the consistency adjusting unit 413 may not be executed, or the result R2 is output and the processing of the consistency adjusting unit 413 is executed. It does not have to be done.
- FIG. 5 is a block diagram illustrating a schematic configuration of a system according to the second embodiment of the present disclosure.
- the system 20 includes a sensor device 500, a smartphone 600, and a server 700.
- the sensor device 500 includes a sensor 110, a preprocessing unit 120, and a memory 130.
- Smartphone 200 includes a position detection unit 210, a sensor 220, and an application 270.
- the server 700 includes an action recognition unit 230, a processing unit 240, an integrated analysis unit 250, and a storage 260.
- the sensor device 500 is a wearable device similar to the sensor device 100 according to the first embodiment.
- the sensor 110, the preprocessing unit 120, and the memory 130 included in the sensor device 100 are the same as those in the first embodiment.
- the sensor device 500 is different from the sensor device 100 according to the first embodiment in that the data stored in the memory 130 is directly or indirectly transmitted to the server 700 by a communication device (not shown).
- the sensor device 500 transmits data to the smartphone 600 using wireless communication such as Bluetooth (registered trademark), for example, and the communication device (not shown) of the smartphone 600 transfers the data to the server 700 via the network. Also good.
- the sensor device 500 may be able to communicate directly with the server 700 via a network.
- the smartphone 600 is a terminal device possessed by the user separately from the sensor device 500, similarly to the smartphone 200 according to the first embodiment.
- the position detection unit 210, the sensor 220, and the application 270 included in the smartphone 600 are the same components as those in the first embodiment.
- the smartphone 600 does not have processing elements such as the behavior recognition unit 230, the processing unit 240, and the integrated analysis unit 250. That is, the smartphone 600 also includes a processor in this embodiment, but the above elements are not realized by the processor.
- the detection results by the position detection unit 210 and the sensor 220 are transmitted to the server 700 via a network by a communication device (not shown).
- Data used by the application 270 is also received from the server 700 via a network by a communication device (not shown).
- the smartphone 600 may relay data transmission from the sensor device 500 to the server 700.
- the server 700 is realized by one or a plurality of information processing apparatuses on the network.
- the action recognition unit 230, the processing unit 240, the integrated analysis unit 250, and the storage 260 included in the server 700 are the same components as those included in the smartphone 200 in the first embodiment. Since the functions of these components are the same as those in the first embodiment, detailed description thereof is omitted.
- the server 700 corresponds to an information processing apparatus that executes sleep section detection.
- the configuration of the system including the server in the embodiment of the present disclosure is not limited to the example illustrated in FIG.
- a part of the action recognition unit 230, the processing unit 240, and the integrated analysis unit 250 may be realized in the sensor device 500 or the smartphone 600.
- an apparatus in which the processing unit 240 is realized can be an information processing apparatus according to an embodiment of the present disclosure.
- the smartphone 600 when the server 700 detects the user's sleep interval based on the sensor data acquired in the sensor device 500, the smartphone 600 may not necessarily be included in the system. In this case, the action recognition unit 230 and the integrated analysis unit 250 may not be included in the server 700, and the sleep section data detected by the processing unit 240 based on the sensor data may be stored in the storage 260 as it is.
- the specific operation section detecting unit 305 detects a section in which the user's image or sound indicates a feature related to sleep.
- the second sensor data acquired by the sensor data acquisition unit 301 includes a user image or sound acquired by a camera or a microphone.
- the specific motion section detection unit 305 may detect a section in which it is recognized that the user has closed his eyes based on the image data included in the second sensor data, as the specific motion section. The fact that the user has closed his / her eyes may be detected by, for example, another sensor that detects the movement of the eyeball or an electrooculogram.
- the specific motion section detection unit 305 may detect a section in which it is recognized that the user is making a specific sound based on the audio data included in the second sensor data as the specific motion section.
- the specific sound referred to here can be, for example, a sound corresponding to sleep or snoring.
- the sleep interval detection unit 307 determines the stable interval when the stable interval detected by the stable interval detection unit 303 and the specific motion interval detected by the specific motion interval detection unit 305 are in parallel. Detecting a sleep interval including. That is, in this embodiment, the posture of the sensor device 100 is regularly changed in the stable section, and the user's image or sound is related to sleep (close eyes, sleep, make a habit, Etc.) are detected as sleep intervals.
- the present embodiment may have a configuration in which the first to third embodiments described above are combined.
- a plurality of detectors for detecting a user's sleep section are selectively used.
- FIG. 6 is a diagram for explaining the use of a selective detector according to the fourth embodiment of the present disclosure.
- one or more detectors used by the detector selector unit 801 are selected when the sensor data is input and the sleep section detection process is started.
- the detector may include a body motion detector 703, a neck motion detector 805, and a closed eye detector 807.
- the detection results by one or more detectors selected by the detector selector unit 801 are combined in the determination result combining unit 809 and output as a result of sleep interval detection.
- the body motion detector 803 is, for example, from sensing by an acceleration sensor included in the sensor 110, pre-processing of a detection value in the pre-processing unit 120, detection of a user's body motion based on the detection value in the processing unit 240, and a sleep section based on the detection (For example, detection of a sleep section using detection of a section in which the user's posture is unchanged). Therefore, when the detector selector unit 801 selects the body motion detector 803, a series of processes for detecting a sleep interval based on the detection of the user's body motion as described above is executed.
- the neck motion detector 805 is, for example, sensing from an acceleration sensor included in the sensor 110, pre-processing of a detection value in the pre-processing unit 120, detection of a user's neck motion based on the detection value in the processing unit 240, and sleep based on the detection Including detection of a section (for example, detection of a sleep section using detection of a section in which the user's neck (head) is regularly operated (a section that is going to be awkward)). Therefore, when the detector selector unit 801 selects the neck motion detector 805, a series of processes for detecting a sleep section based on the detection of the user's neck (head) motion as described above is executed. become.
- the closed eye detector 807 detects, for example, whether or not the user is closing his / her eyes based on image data preprocessing in the preprocessing unit 120 and image data in the processing unit 240 from image capturing by a camera included in the sensor 110. And the detection of the sleep interval based on it (for example, the detection of the sleep interval on the condition that the user's body movement is regular and the user closes his eyes). Therefore, when the detector selector unit 801 selects the closed eye detector 807, a series of processes for detecting a sleep interval based on detection of whether or not the user has closed eyes as described above is executed. become.
- the determination result combining unit 809 combines detection results from the selected one or more detectors and determines a sleep interval.
- the determination result combining unit 809 can be realized as the processing unit 240 in the smartphone 200, for example. If there is a single selected detector, the determination result combining unit 809 may output the detection result of the selected detector as it is. In addition, when there are a plurality of selected detectors, the determination result combining unit 809 may take a logical sum of the detection results of the sleep intervals by the plurality of selected detectors, for example. In this case, the section detected as the sleep section by any one detector is output as the sleep section. Or the determination result coupling
- the determination result combining unit 809 obtains a score indicating the probability of the detection result in each selected detector, weights the detection result based on the score, and adds the sleep interval based on the score. You may decide.
- each detector for example, the body motion detector 803, the neck motion detector 805, and the closed eye detector 807 outputs a score indicating the likelihood of the detection result to the determination result combining unit 809.
- the detection can be continued using the smartphone 200 while the sensor device 100 is removed, By switching the sensor to be used according to the remaining battery level of the sensor device 100, it is possible to adopt an optimal sleep interval detection method according to the situation.
- Control of external devices As an example in which the external device control unit 309 of the information processing device 300 outputs the control signal of the external device when the start of the sleep interval is detected, a control signal for stopping or pausing the external device as described above is used. There is an example of outputting, and an example in which an external device outputs a control signal for setting an alarm based on the start time of the sleep interval.
- the external device control unit 309 may output a control signal for causing the external device to execute a predetermined operation when the start of the sleep interval is detected.
- the external device control unit 309 may output a control signal for causing an information processing terminal such as a personal computer to execute processing such as backup, upload, and analysis. If these processes are executed while the user is using the information processing terminal, there is a possibility that the operation of the information processing terminal desired by the user may be hindered due to the consumption of the resources of the apparatus. Therefore, if these processes are executed when the start of the user's sleep section is detected, the processes can be completed without hindering the user's operation.
- notification information output unit 311 of the information processing device 300 outputs information for notifying other users of the detection result of the sleep section, information indicating the start, continuation, or end of the sleep section as described above. There is an example of outputting, and an example of outputting information indicating the length of the sleep interval in a predetermined period.
- the notification information output unit 311 acquires information on the user's schedule, and wakes up the user when it is determined that the user's sleep state impedes execution of the schedule.
- a notification requesting a friend to take an action may be output.
- the external device control unit 309 controls the external device (including other functions implemented by the same processor as described above) to sound an alarm at the time when the user should wake up to execute the schedule. If the signal is output and the user still does not wake up (the end of the sleep period is not detected), the notification information output unit 311 may output a notification to another user.
- Still another example of application of the sleep section detection result in the embodiment of the present disclosure occurs at a time when the user is originally sleeping in a certain country or region.
- the percentage of users sleeping in a certain region is statistically grasped based on the detection result of the sleep section of the user in the region, and infrastructure such as a base station for wireless communication is set in a low power consumption mode. You may use as a standard of whether to operate.
- FIG. 7 is a block diagram illustrating a hardware configuration example of the information processing apparatus according to the embodiment of the present disclosure.
- the illustrated information processing apparatus 900 can realize, for example, the sensor device, smartphone, or server in the above-described embodiment.
- the information processing apparatus 900 includes a CPU (Central Processing unit) 901, a ROM (Read Only Memory) 903, and a RAM (Random Access Memory) 905.
- the information processing apparatus 900 may include a host bus 907, a bridge 909, an external bus 911, an interface 913, an input device 915, an output device 917, a storage device 919, a drive 921, a connection port 923, and a communication device 925.
- the information processing apparatus 900 may include an imaging device 933 and a sensor 935 as necessary.
- the information processing apparatus 900 may include a processing circuit called DSP (Digital Signal Processor) or ASIC (Application Specific Integrated Circuit) instead of or in addition to the CPU 901.
- DSP Digital Signal Processor
- ASIC Application Specific Integrated Circuit
- the CPU 901 functions as an arithmetic processing device and a control device, and controls all or a part of the operation in the information processing device 900 according to various programs recorded in the ROM 903, the RAM 905, the storage device 919, or the removable recording medium 927.
- the ROM 903 stores programs and calculation parameters used by the CPU 901.
- the RAM 905 primarily stores programs used in the execution of the CPU 901, parameters that change as appropriate during the execution, and the like.
- the CPU 901, the ROM 903, and the RAM 905 are connected to each other by a host bus 907 configured by an internal bus such as a CPU bus. Further, the host bus 907 is connected to an external bus 911 such as a PCI (Peripheral Component Interconnect / Interface) bus via a bridge 909.
- PCI Peripheral Component Interconnect / Interface
- the input device 915 is a device operated by the user, such as a mouse, a keyboard, a touch panel, a button, a switch, and a lever.
- the input device 915 may be, for example, a remote control device that uses infrared rays or other radio waves, or may be an external connection device 929 such as a mobile phone that supports the operation of the information processing device 900.
- the input device 915 includes an input control circuit that generates an input signal based on information input by the user and outputs the input signal to the CPU 901. The user operates the input device 915 to input various data and instruct processing operations to the information processing device 900.
- the output device 917 is a device that can notify the user of the acquired information visually or audibly.
- the output device 917 can be, for example, a display device such as an LCD (Liquid Crystal Display), a PDP (Plasma Display Panel), an organic EL (Electro-Luminescence) display, an audio output device such as a speaker and headphones, and a printer device.
- the output device 917 outputs the result obtained by the processing of the information processing device 900 as video such as text or an image, or outputs it as audio such as voice or sound.
- the storage device 919 is a data storage device configured as an example of a storage unit of the information processing device 900.
- the storage device 919 includes, for example, a magnetic storage device such as an HDD (Hard Disk Drive), a semiconductor storage device, an optical storage device, or a magneto-optical storage device.
- the storage device 919 stores programs executed by the CPU 901, various data, various data acquired from the outside, and the like.
- the drive 921 is a reader / writer for a removable recording medium 927 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory, and is built in or externally attached to the information processing apparatus 900.
- the drive 921 reads information recorded on the attached removable recording medium 927 and outputs the information to the RAM 905.
- the drive 921 writes a record in the attached removable recording medium 927.
- the connection port 923 is a port for directly connecting a device to the information processing apparatus 900.
- the connection port 923 can be, for example, a USB (Universal Serial Bus) port, an IEEE 1394 port, a SCSI (Small Computer System Interface) port, or the like.
- the connection port 923 may be an RS-232C port, an optical audio terminal, an HDMI (registered trademark) (High-Definition Multimedia Interface) port, or the like.
- the communication device 925 is a communication interface configured with, for example, a communication device for connecting to the communication network 931.
- the communication device 925 may be, for example, a communication card for wired or wireless LAN (Local Area Network), Bluetooth (registered trademark), or WUSB (Wireless USB).
- the communication device 925 may be a router for optical communication, a router for ADSL (Asymmetric Digital Subscriber Line), or a modem for various communication.
- the communication device 925 transmits and receives signals and the like using a predetermined protocol such as TCP / IP with the Internet and other communication devices, for example.
- the communication network 931 connected to the communication device 925 is a wired or wireless network, such as the Internet, a home LAN, infrared communication, radio wave communication, or satellite communication.
- the imaging device 933 uses various members such as an imaging element such as a CCD (Charge Coupled Device) or a CMOS (Complementary Metal Oxide Semiconductor), and a lens for controlling the formation of a subject image on the imaging element. It is an apparatus that images a real space and generates a captured image.
- the imaging device 933 may capture a still image or may capture a moving image.
- the sensor 935 is various sensors such as an acceleration sensor, a gyro sensor, a geomagnetic sensor, an optical sensor, and a sound sensor.
- the sensor 935 acquires information about the state of the information processing apparatus 900 itself, such as the posture of the information processing apparatus 900, and information about the surrounding environment of the information processing apparatus 900, such as brightness and noise around the information processing apparatus 900, for example. To do.
- the sensor 935 may include a GPS sensor that receives a GPS (Global Positioning System) signal and measures the latitude, longitude, and altitude of the apparatus.
- GPS Global Positioning System
- Each component described above may be configured using a general-purpose member, or may be configured by hardware specialized for the function of each component. Such a configuration can be appropriately changed according to the technical level at the time of implementation.
- an information processing apparatus for example, an information processing apparatus, a system, an information processing method executed by the information processing apparatus or system, a program for causing the information processing apparatus to function, and a program are recorded It may include tangible media that is not temporary.
- a stable section detection unit that detects a stable section of the posture of the sensor device based on first sensor data provided from a sensor device for a user to wear;
- a specific action section detector that detects a specific action section in which a specific action of the user has occurred, based on second sensor data;
- An information processing apparatus comprising: a sleep section detection unit that detects a sleep section in which the user is in a sleep state based on a relationship between the stable section and the specific operation section.
- the specific motion section is a section in which the posture of the user has changed, The information processing unit according to (2), wherein the sleep section detection unit detects a sleep section including the stable section and the specific operation section when the stable section and the specific operation section are repeated. apparatus. (4) The information processing apparatus according to (3), wherein each of the first sensor data and the second sensor data includes acceleration sensor data. (5) The information processing apparatus according to (1), wherein the stable section is a section in which the attitude of the sensor device has regularly changed. (6) The specific operation section is a section in which the image or sound of the user has shown characteristics related to sleep, The information processing apparatus according to (5), wherein the sleep section detection unit detects a sleep section including the stable section when the stable section and the specific operation section are in parallel.
- the first sensor data includes acceleration sensor data
- the second sensor data includes image data
- the information processing apparatus according to (6), wherein the specific operation section is a section in which it is recognized that the user has closed his eyes based on the image data.
- the first sensor data includes acceleration sensor data
- the second sensor data includes audio data;
- the information processing apparatus according to (6), wherein the specific operation section is a section in which it is recognized that the user is generating a specific sound based on the audio data.
- the sleep interval detecting unit detects the sleep interval based on a user action recognized based on sensor data different from the first and second sensor data.
- the information processing apparatus according to any one of (8).
- the information processing apparatus according to any one of (1) to (9), further including an external device control unit that outputs a control signal of the external device when the start of the sleep period is detected.
- the information processing apparatus according to (10), wherein the external device control unit outputs the control signal for stopping or pausing the external device.
- the external device control unit outputs the control signal for the external device to set an alarm based on a start time of the sleep interval.
- the information processing apparatus according to any one of (1) to (12), further including a notification information output unit that outputs information for notifying another user of the detection result of the sleep period.
- the information processing apparatus according to (13), wherein the notification information output unit outputs information indicating start, continuation, or end of the sleep interval.
- the information processing apparatus wherein the notification information output unit outputs information indicating a length of the sleep section in a predetermined period.
- An information processing method comprising: a processor detecting a sleep section in which the user is in a sleep state based on a relationship between the stable section and the specific operation section.
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- Heart & Thoracic Surgery (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- Biophysics (AREA)
- Pathology (AREA)
- Physics & Mathematics (AREA)
- Biomedical Technology (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Molecular Biology (AREA)
- Surgery (AREA)
- Animal Behavior & Ethology (AREA)
- Physiology (AREA)
- Dentistry (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Anesthesiology (AREA)
- Computer Networks & Wireless Communication (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
Abstract
Description
1.第1の実施形態
1-1.システム構成
1-2.睡眠区間検出の原理
1-3.情報処理装置の機能構成
1-4.睡眠区間検出の処理フロー
2.第2の実施形態
3.第3の実施形態
4.第4の実施形態
5.検出結果の応用例
6.ハードウェア構成
7.補足
(1-1.システム構成)
図1は、本開示の第1の実施形態に係るシステムの概略的な構成を示すブロック図である。図1を参照すると、システム10は、センサ装置100と、スマートフォン200とを含む。センサ装置100は、センサ110と、前処理部120と、メモリ130とを含む。スマートフォン200は、位置検出部210と、センサ220と、行動認識部230と、処理部240と、統合解析部250と、ストレージ260と、アプリケーション270とを含む。
センサ装置100は、ウェアラブルの装置である。センサ装置100は、例えば手首や足首、指などに巻き付けるなどしてユーザに直接的に装着される。あるいは、センサ装置100は、クリップなどを用いて衣服に固定されることによって、ユーザに間接的に装着されてもよい。なお、センサ装置100は、ユーザが装着するための装置ではあるものの、必ずしも常にユーザによって装着されているわけではない。例えば、センサ装置100は、入浴中や身支度の途中などには取り外される場合がある。従って、後述するように、センサ110に含まれる加速度センサの検出値が変動しない場合には、ユーザがセンサ装置100を装着して体勢を安定させている場合と、センサ装置100が取り外されてどこかに静置されている場合とがありうる。
スマートフォン200は、センサ装置100とは別にユーザが所持している端末装置である。スマートフォン200は、例えば後述する情報処理装置のハードウェア構成によって実現されうる。また、スマートフォン200は、例えばタブレット端末など、同様の機能を実現しうる他の端末装置によって代替されてもよい。以下、スマートフォン200の各構成要素についてさらに説明する。
次に、上記の第1の実施形態における睡眠区間検出の原理について説明する。
図3は、本開示の第1の実施形態における睡眠区間検出を実行可能な情報処理装置の概略的な機能構成を示すブロック図である。図3を参照すると、情報処理装置300は、センサデータ取得部301と、安定区間検出部303と、特定動作区間検出部305と、睡眠区間検出部307と、外部装置制御部309と、通知情報出力部311とを含む。本実施形態において、情報処理装置300は、センサ装置100またはスマートフォン200が備えるプロセッサによって実現される。
図4は、本開示の第1の実施形態における睡眠区間検出の処理フローについて説明するための図である。図4に示された処理フローは、上記で図3を参照して説明した情報処理装置300の機能構成に対応する。つまり、以下で説明される特徴量算出部401~整合性調整部413によって実行される一連の処理は、情報処理装置300の各構成要素の機能を異なる観点から表現したものともいえる。従って、一連の処理は、センサ装置100またはスマートフォン200のプロセッサによって実行されうる。以下、処理フローの各構成要素の処理について、さらに説明する。
図5は、本開示の第2の実施形態に係るシステムの概略的な構成を示すブロック図である。図5を参照すると、システム20は、センサ装置500と、スマートフォン600と、サーバ700とを含む。センサ装置500は、センサ110と、前処理部120と、メモリ130とを含む。スマートフォン200は、位置検出部210と、センサ220と、アプリケーション270とを含む。サーバ700は、行動認識部230と、処理部240と、統合解析部250と、ストレージ260とを含む。
センサ装置500は、上記の第1の実施形態に係るセンサ装置100と同様にウェアラブルの装置である。センサ装置100に含まれるセンサ110、前処理部120、およびメモリ130についても第1の実施形態と同様である。ただし、センサ装置500は、メモリ130に格納されたデータを、図示しない通信装置によって直接的または間接的にサーバ700に送信する点で、第1の実施形態に係るセンサ装置100とは異なる。センサ装置500は、例えばBluetooth(登録商標)などの無線通信を利用してスマートフォン600にデータを送信し、スマートフォン600の通信装置(図示せず)がネットワークを介してデータをサーバ700に転送してもよい。あるいは、センサ装置500は、ネットワークを介して直接的にサーバ700と通信することが可能であってもよい。
スマートフォン600は、上記の第1の実施形態に係るスマートフォン200と同様に、センサ装置500とは別にユーザが所持している端末装置である。スマートフォン600に含まれる位置検出部210、センサ220、およびアプリケーション270は、第1の実施形態と同様の構成要素である。その一方で、本実施形態において、スマートフォン600は、行動認識部230、処理部240、および統合解析部250といった処理要素を有さない。つまり、本実施形態でもスマートフォン600はプロセッサを備えるが、上記の要素はプロセッサによって実現されない。位置検出部210およびセンサ220による検出結果は、図示しない通信装置によってネットワークを介してサーバ700に送信される。また、アプリケーション270によって利用されるデータも、図示しない通信装置によってネットワークを介してサーバ700から受信される。さらに、上記のように、スマートフォン600は、センサ装置500からサーバ700へのデータの送信を中継してもよい。
サーバ700は、ネットワーク上の1または複数の情報処理装置によって実現される。サーバ700に含まれる行動認識部230、処理部240、統合解析部250、およびストレージ260は、第1の実施形態においてスマートフォン200に含まれていたものと同様の構成要素である。これらの構成要素の機能については上記の第1の実施形態と同様であるため、詳細な説明は省略する。
次に、本開示の第3の実施形態について説明する。本実施形態は、装置構成的には上記の第1または第2の実施形態と同様であるが、図3を参照して説明した情報処理装置300において実行される睡眠区間検出の処理が異なる。従って、以下の説明ではかかる相違点について主に説明し、これまでに説明した実施形態と共通する点(装置構成)については詳細な説明を省略する。なお、以下の第3および第4の実施形態の説明では第1の実施形態の符号(センサ装置100およびスマートフォン200)を引用するが、第2の実施形態(センサ装置500、スマートフォン600、サーバ700)においても同様の構成が可能である。
次に、本開示の第4の実施形態について説明する。本実施形態は、例えば、上記の第1~第3の実施形態を組み合わせた構成をとりうる。本実施形態では、ユーザの睡眠区間を検出するための複数の検出器が選択的に使用される。
次に、本開示の実施形態における、睡眠区間の検出結果の応用例について説明する。上述のように、本開示の実施形態では、例えば、睡眠区間の開始が検出された場合に外部装置の制御信号を出力したり、睡眠区間の検出結果を他のユーザに通知するための情報を出力したりすることによって睡眠区間の検出結果が応用される。以下では、そのような睡眠区間の検出結果の応用例について、より具体的な例を含めて説明する。
情報処理装置300の外部装置制御部309が、睡眠区間の開始が検出された場合に外部装置の制御信号を出力する例としては、上述したように外部装置を停止または休止するための制御信号を出力する例や、外部装置が睡眠区間の開始時刻を基準にしてアラームを設定するための制御信号を出力する例がある。
情報処理装置300の通知情報出力部311が、睡眠区間の検出結果を他のユーザに通知するための情報を出力する例としては、上述したように睡眠区間の開始、継続または終了を示す情報を出力する例や、所定の期間における睡眠区間の長さを示す情報を出力する例がある。
次に、図7を参照して、本開示の実施形態に係る情報処理装置のハードウェア構成について説明する。図7は、本開示の実施形態に係る情報処理装置のハードウェア構成例を示すブロック図である。図示された情報処理装置900は、例えば、上記の実施形態におけるセンサ装置、スマートフォン、またはサーバなどを実現しうる。
本開示の実施形態は、例えば、上記で説明したような情報処理装置、システム、情報処理装置またはシステムで実行される情報処理方法、情報処理装置を機能させるためのプログラム、およびプログラムが記録された一時的でない有形の媒体を含みうる。
(1)ユーザが装着するためのセンサ装置から提供される第1のセンサデータに基づいて前記センサ装置の姿勢の安定区間を検出する安定区間検出部と、
第2のセンサデータに基づいて、前記ユーザの特定の動作が発生した特定動作区間を検出する特定動作区間検出部と、
前記安定区間と前記特定動作区間との関係に基づいて前記ユーザが睡眠状態にあった睡眠区間を検出する睡眠区間検出部と
を備える情報処理装置。
(2)前記安定区間は、前記センサ装置の姿勢が不変であった区間である、前記(1)に記載の情報処理装置。
(3)前記特定動作区間は、前記ユーザの体勢が変化した区間であり、
前記睡眠区間検出部は、前記安定区間と前記特定動作区間とが繰り返されている場合に、前記安定区間と前記特定動作区間とを含む睡眠区間を検出する、前記(2)に記載の情報処理装置。
(4)前記第1のセンサデータおよび前記第2のセンサデータはいずれも加速度センサデータを含む、前記(3)に記載の情報処理装置。
(5)前記安定区間は、前記センサ装置の姿勢が規則的に変化していた区間である、前記(1)に記載の情報処理装置。
(6)前記特定動作区間は、前記ユーザの画像または音声が睡眠に関係する特徴を示していた区間であり、
前記睡眠区間検出部は、前記安定区間と前記特定動作区間とが並行している場合に、前記安定区間を含む睡眠区間を検出する、前記(5)に記載の情報処理装置。
(7)前記第1のセンサデータは加速度センサデータを含み、
前記第2のセンサデータは画像データを含み、
前記特定動作区間は、前記画像データに基づいて前記ユーザが目を閉じていたことが認識された区間である、前記(6)に記載の情報処理装置。
(8)前記第1のセンサデータは加速度センサデータを含み、
前記第2のセンサデータは音声データを含み、
前記特定動作区間は、前記音声データに基づいて前記ユーザが特定の音を発していることが認識された区間である、前記(6)に記載の情報処理装置。
(9)前記睡眠区間検出部は、前記第1および前記第2のセンサデータとは異なるセンサデータに基づいて認識されたユーザの行動にさらに基づいて前記睡眠区間を検出する、前記(1)~(8)のいずれか1項に記載の情報処理装置。
(10)前記睡眠区間の開始が検出された場合に外部装置の制御信号を出力する外部装置制御部をさらに備える、前記(1)~(9)のいずれか1項に記載の情報処理装置。
(11)前記外部装置制御部は、前記外部装置を停止または休止するための前記制御信号を出力する、前記(10)に記載の情報処理装置。
(12)前記外部装置制御部は、前記外部装置が前記睡眠区間の開始時刻を基準にしてアラームを設定するための前記制御信号を出力する、前記(10)に記載の情報処理装置。
(13)前記睡眠区間の検出結果を他のユーザに通知するための情報を出力する通知情報出力部をさらに備える、前記(1)~(12)のいずれか1項に記載の情報処理装置。
(14)前記通知情報出力部は、前記睡眠区間の開始、継続または終了を示す情報を出力する、前記(13)に記載の情報処理装置。
(15)前記通知情報出力部は、所定の期間における前記睡眠区間の長さを示す情報を出力する、前記(13)に記載の情報処理装置。
(16)ユーザが装着するためのセンサ装置から提供される第1のセンサデータに基づいて、前記センサ装置の姿勢の安定区間を検出することと、
第2のセンサデータに基づいて、前記ユーザの特定の動作が発生した特定動作区間を検出することと、
プロセッサが、前記安定区間と前記特定動作区間との関係に基づいて前記ユーザが睡眠状態にあった睡眠区間を検出することと
を含む情報処理方法。
(17)ユーザが装着するためのセンサ装置から提供される第1のセンサデータに基づいて、前記センサ装置の姿勢の安定区間を検出する機能と、
第2のセンサデータに基づいて、前記ユーザの特定の動作が発生した特定動作区間を検出する機能と、
前記安定区間と前記特定動作区間との関係に基づいて前記ユーザが睡眠状態にあった睡眠区間を検出する機能と
をコンピュータに実現させるためのプログラム。
100 センサ装置
110 センサ
120 前処理部
130 メモリ
200 スマートフォン
210 位置検出部
220 センサ
230 行動認識部
240 処理部
250 統合解析部
260 ストレージ
270 アプリケーション
300 情報処理装置
301 取得部
303 安定区間検出部
305 特定動作区間検出部
307 睡眠区間検出部
309 外部装置制御部
311 通知情報出力部
Claims (17)
- ユーザが装着するためのセンサ装置から提供される第1のセンサデータに基づいて前記センサ装置の姿勢の安定区間を検出する安定区間検出部と、
第2のセンサデータに基づいて、前記ユーザの特定の動作が発生した特定動作区間を検出する特定動作区間検出部と、
前記安定区間と前記特定動作区間との関係に基づいて前記ユーザが睡眠状態にあった睡眠区間を検出する睡眠区間検出部と
を備える情報処理装置。 - 前記安定区間は、前記センサ装置の姿勢が不変であった区間である、請求項1に記載の情報処理装置。
- 前記特定動作区間は、前記ユーザの体勢が変化した区間であり、
前記睡眠区間検出部は、前記安定区間と前記特定動作区間とが繰り返されている場合に、前記安定区間と前記特定動作区間とを含む睡眠区間を検出する、請求項2に記載の情報処理装置。 - 前記第1のセンサデータおよび前記第2のセンサデータはいずれも加速度センサデータを含む、請求項3に記載の情報処理装置。
- 前記安定区間は、前記センサ装置の姿勢が規則的に変化していた区間である、請求項1に記載の情報処理装置。
- 前記特定動作区間は、前記ユーザの画像または音声が睡眠に関係する特徴を示していた区間であり、
前記睡眠区間検出部は、前記安定区間と前記特定動作区間とが並行している場合に、前記安定区間を含む睡眠区間を検出する、請求項5に記載の情報処理装置。 - 前記第1のセンサデータは加速度センサデータを含み、
前記第2のセンサデータは画像データを含み、
前記特定動作区間は、前記画像データに基づいて前記ユーザが目を閉じていたことが認識された区間である、請求項6に記載の情報処理装置。 - 前記第1のセンサデータは加速度センサデータを含み、
前記第2のセンサデータは音声データを含み、
前記特定動作区間は、前記音声データに基づいて前記ユーザが特定の音を発していることが認識された区間である、請求項6に記載の情報処理装置。 - 前記睡眠区間検出部は、前記第1および前記第2のセンサデータとは異なるセンサデータに基づいて認識されたユーザの行動にさらに基づいて前記睡眠区間を検出する、請求項1に記載の情報処理装置。
- 前記睡眠区間の開始が検出された場合に外部装置の制御信号を出力する外部装置制御部をさらに備える、請求項1に記載の情報処理装置。
- 前記外部装置制御部は、前記外部装置を停止または休止するための前記制御信号を出力する、請求項10に記載の情報処理装置。
- 前記外部装置制御部は、前記外部装置が前記睡眠区間の開始時刻を基準にしてアラームを設定するための前記制御信号を出力する、請求項10に記載の情報処理装置。
- 前記睡眠区間の検出結果を他のユーザに通知するための情報を出力する通知情報出力部をさらに備える、請求項1に記載の情報処理装置。
- 前記通知情報出力部は、前記睡眠区間の開始、継続または終了を示す情報を出力する、請求項13に記載の情報処理装置。
- 前記通知情報出力部は、所定の期間における前記睡眠区間の長さを示す情報を出力する、請求項13に記載の情報処理装置。
- ユーザが装着するためのセンサ装置から提供される第1のセンサデータに基づいて、前記センサ装置の姿勢の安定区間を検出することと、
第2のセンサデータに基づいて、前記ユーザの特定の動作が発生した特定動作区間を検出することと、
プロセッサが、前記安定区間と前記特定動作区間との関係に基づいて前記ユーザが睡眠状態にあった睡眠区間を検出することと
を含む情報処理方法。 - ユーザが装着するためのセンサ装置から提供される第1のセンサデータに基づいて、前記センサ装置の姿勢の安定区間を検出する機能と、
第2のセンサデータに基づいて、前記ユーザの特定の動作が発生した特定動作区間を検出する機能と、
前記安定区間と前記特定動作区間との関係に基づいて前記ユーザが睡眠状態にあった睡眠区間を検出する機能と
をコンピュータに実現させるためのプログラム。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP14868106.7A EP3078326A4 (en) | 2013-12-03 | 2014-09-12 | Information-processing apparatus, information-processing method, and program |
JP2015551409A JPWO2015083411A1 (ja) | 2013-12-03 | 2014-09-12 | 情報処理装置、情報処理方法、およびプログラム |
US15/029,404 US20160249852A1 (en) | 2013-12-03 | 2014-09-12 | Information processing apparatus, information processing method, and program |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2013250187 | 2013-12-03 | ||
JP2013-250187 | 2013-12-03 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015083411A1 true WO2015083411A1 (ja) | 2015-06-11 |
Family
ID=53273189
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2014/074257 WO2015083411A1 (ja) | 2013-12-03 | 2014-09-12 | 情報処理装置、情報処理方法、およびプログラム |
Country Status (4)
Country | Link |
---|---|
US (1) | US20160249852A1 (ja) |
EP (1) | EP3078326A4 (ja) |
JP (1) | JPWO2015083411A1 (ja) |
WO (1) | WO2015083411A1 (ja) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2018000232A (ja) * | 2016-06-27 | 2018-01-11 | 日本電信電話株式会社 | 変化点検知装置、変化点検知方法および変化点検知プログラム |
JP2020036781A (ja) * | 2018-09-05 | 2020-03-12 | 日本電信電話株式会社 | 生体情報解析装置、生体情報解析方法、および生体情報解析システム |
JP2021041088A (ja) * | 2019-09-13 | 2021-03-18 | 株式会社東芝 | 電子装置及び方法 |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105573495B (zh) * | 2015-12-14 | 2020-06-23 | 联想(北京)有限公司 | 一种信息处理方法及穿戴式设备 |
US10621992B2 (en) * | 2016-07-22 | 2020-04-14 | Lenovo (Singapore) Pte. Ltd. | Activating voice assistant based on at least one of user proximity and context |
US10664533B2 (en) | 2017-05-24 | 2020-05-26 | Lenovo (Singapore) Pte. Ltd. | Systems and methods to determine response cue for digital assistant based on context |
US11406788B2 (en) * | 2017-08-08 | 2022-08-09 | Sony Corporation | Information processing apparatus and method |
CN109480782B (zh) * | 2018-11-16 | 2023-10-20 | 深圳和而泰智能家电控制器有限公司 | 一种睡眠状态检测方法、装置及设备 |
CN109875518A (zh) * | 2019-04-02 | 2019-06-14 | 浙江和也健康科技有限公司 | 一种多用户睡眠状况同步监测系统及监测方法 |
US11202121B2 (en) | 2020-05-13 | 2021-12-14 | Roku, Inc. | Providing customized entertainment experience using human presence detection |
US11395232B2 (en) * | 2020-05-13 | 2022-07-19 | Roku, Inc. | Providing safety and environmental features using human presence detection |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0453527A (ja) * | 1990-06-20 | 1992-02-21 | Matsushita Electric Ind Co Ltd | 安眠装置 |
JPH04212331A (ja) * | 1990-06-21 | 1992-08-03 | Mitsubishi Denki Eng Kk | 眠り検出装置 |
JPH0515517A (ja) * | 1991-07-11 | 1993-01-26 | Matsushita Electric Ind Co Ltd | 睡眠検出装置 |
JP2004097496A (ja) * | 2002-09-09 | 2004-04-02 | Yamatake Corp | 睡眠モニタリングシステム及びモニタリング装置 |
JP2005199078A (ja) * | 2005-02-07 | 2005-07-28 | Toshiba Corp | 状態監視装置 |
JP2005312913A (ja) * | 2004-03-30 | 2005-11-10 | Toshiba Corp | 生体情報計測装置 |
JP2010133692A (ja) * | 2008-10-31 | 2010-06-17 | Mitsubishi Electric Corp | 空気調和機 |
JP2013052165A (ja) | 2011-09-06 | 2013-03-21 | Sony Corp | 情報処理装置、情報処理方法、およびプログラム |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5853005A (en) * | 1996-05-02 | 1998-12-29 | The United States Of America As Represented By The Secretary Of The Army | Acoustic monitoring system |
US8831735B2 (en) * | 2005-08-31 | 2014-09-09 | Michael Sasha John | Methods and systems for semi-automatic adjustment of medical monitoring and treatment |
US9089713B2 (en) * | 2005-08-31 | 2015-07-28 | Michael Sasha John | Methods and systems for semi-automatic adjustment of medical monitoring and treatment |
US7733224B2 (en) * | 2006-06-30 | 2010-06-08 | Bao Tran | Mesh network personal emergency response appliance |
AU2009246442B2 (en) * | 2008-05-14 | 2015-02-12 | Heartmiles, Llc. | Physical activity monitor and data collection unit |
US8094013B1 (en) * | 2009-03-31 | 2012-01-10 | Lee Taek Kyu | Baby monitoring system |
US20110230790A1 (en) * | 2010-03-16 | 2011-09-22 | Valeriy Kozlov | Method and system for sleep monitoring, regulation and planning |
EP2691020A2 (en) * | 2011-03-30 | 2014-02-05 | Koninklijke Philips N.V. | Contactless sleep disorder screening system |
EP2524647A1 (en) * | 2011-05-18 | 2012-11-21 | Alain Gilles Muzet | System and method for determining sleep stages of a person |
EP2852361B1 (en) * | 2012-05-22 | 2019-07-03 | Hill-Rom Services, Inc. | Adverse event mitigation systems, methods and devices |
-
2014
- 2014-09-12 WO PCT/JP2014/074257 patent/WO2015083411A1/ja active Application Filing
- 2014-09-12 JP JP2015551409A patent/JPWO2015083411A1/ja active Pending
- 2014-09-12 EP EP14868106.7A patent/EP3078326A4/en not_active Withdrawn
- 2014-09-12 US US15/029,404 patent/US20160249852A1/en not_active Abandoned
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0453527A (ja) * | 1990-06-20 | 1992-02-21 | Matsushita Electric Ind Co Ltd | 安眠装置 |
JPH04212331A (ja) * | 1990-06-21 | 1992-08-03 | Mitsubishi Denki Eng Kk | 眠り検出装置 |
JPH0515517A (ja) * | 1991-07-11 | 1993-01-26 | Matsushita Electric Ind Co Ltd | 睡眠検出装置 |
JP2004097496A (ja) * | 2002-09-09 | 2004-04-02 | Yamatake Corp | 睡眠モニタリングシステム及びモニタリング装置 |
JP2005312913A (ja) * | 2004-03-30 | 2005-11-10 | Toshiba Corp | 生体情報計測装置 |
JP2005199078A (ja) * | 2005-02-07 | 2005-07-28 | Toshiba Corp | 状態監視装置 |
JP2010133692A (ja) * | 2008-10-31 | 2010-06-17 | Mitsubishi Electric Corp | 空気調和機 |
JP2013052165A (ja) | 2011-09-06 | 2013-03-21 | Sony Corp | 情報処理装置、情報処理方法、およびプログラム |
Non-Patent Citations (1)
Title |
---|
See also references of EP3078326A4 * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2018000232A (ja) * | 2016-06-27 | 2018-01-11 | 日本電信電話株式会社 | 変化点検知装置、変化点検知方法および変化点検知プログラム |
JP2020036781A (ja) * | 2018-09-05 | 2020-03-12 | 日本電信電話株式会社 | 生体情報解析装置、生体情報解析方法、および生体情報解析システム |
WO2020050042A1 (ja) * | 2018-09-05 | 2020-03-12 | 日本電信電話株式会社 | 生体情報解析装置、生体情報解析方法、および生体情報解析システム |
JP7180216B2 (ja) | 2018-09-05 | 2022-11-30 | 日本電信電話株式会社 | 生体情報解析装置、生体情報解析方法、および生体情報解析システム |
JP2021041088A (ja) * | 2019-09-13 | 2021-03-18 | 株式会社東芝 | 電子装置及び方法 |
Also Published As
Publication number | Publication date |
---|---|
EP3078326A1 (en) | 2016-10-12 |
EP3078326A4 (en) | 2017-08-09 |
US20160249852A1 (en) | 2016-09-01 |
JPWO2015083411A1 (ja) | 2017-03-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2015083411A1 (ja) | 情報処理装置、情報処理方法、およびプログラム | |
US10990613B2 (en) | Information processing apparatus and information processing method | |
JP6756328B2 (ja) | 情報処理装置、情報処理方法、およびプログラム | |
WO2016143404A1 (ja) | 情報処理装置、情報処理方法、およびプログラム | |
US8532737B2 (en) | Real-time video based automated mobile sleep monitoring using state inference | |
US20180048802A1 (en) | Method of operating a wearable lifelogging device | |
JP6459972B2 (ja) | 表示制御装置、表示制御方法、およびプログラム | |
WO2015196584A1 (zh) | 一种智能录制系统 | |
WO2018072339A1 (zh) | 一种虚拟现实头盔和切换虚拟现实头盔显示信息的方法 | |
WO2017175432A1 (ja) | 情報処理装置、情報処理方法、およびプログラム | |
WO2017071059A1 (zh) | 可穿戴设备的通信方法、装置及系统 | |
KR102496225B1 (ko) | 영상 인코딩 방법 및 이를 지원하는 전자 장치 | |
JP6402718B2 (ja) | 情報処理装置、制御方法およびプログラム | |
JP2015118185A (ja) | 情報処理装置、情報処理方法、およびプログラム | |
KR20160123294A (ko) | 도메인 인식 카메라 시스템 | |
EP3092630B1 (en) | Dual mode baby monitoring priority application | |
CN112069949A (zh) | 一种基于人工智能的婴儿睡眠监测系统及监测方法 | |
CN114543313A (zh) | 空调控制方法、服务器、空调及用户终端 | |
WO2016088611A1 (ja) | 情報処理装置、情報処理方法及びコンピュータプログラム | |
WO2022151887A1 (zh) | 睡眠监测方法及相关装置 | |
JP5669302B2 (ja) | 行動情報収集システム | |
WO2016143415A1 (ja) | 情報処理装置、情報処理方法およびプログラム | |
JP6435595B2 (ja) | トレーニング支援システム、サーバー、端末、カメラ、方法並びにプログラム | |
JP2017004372A (ja) | 情報処理装置、情報処理方法及びプログラム | |
CN114007141A (zh) | 一种智能终端、服务器及睡眠检测显示方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14868106 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2015551409 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15029404 Country of ref document: US |
|
REEP | Request for entry into the european phase |
Ref document number: 2014868106 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2014868106 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |