US20160283579A1 - Information processing device, information processing method, and computer program - Google Patents

Information processing device, information processing method, and computer program Download PDF

Info

Publication number
US20160283579A1
US20160283579A1 US15/173,793 US201615173793A US2016283579A1 US 20160283579 A1 US20160283579 A1 US 20160283579A1 US 201615173793 A US201615173793 A US 201615173793A US 2016283579 A1 US2016283579 A1 US 2016283579A1
Authority
US
United States
Prior art keywords
action
segment
information
unit
display
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/173,793
Inventor
Masatomo Kurata
Masanori Katsu
Takashi Ogata
Yasutaka Fukumoto
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sony Corp
Original Assignee
Sony Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=47357016&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=US20160283579(A1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Application filed by Sony Corp filed Critical Sony Corp
Priority to US15/173,793 priority Critical patent/US20160283579A1/en
Assigned to SONY CORPORATION reassignment SONY CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: FUKUMOTO, YASUTAKA, KURATA, MASATOMO, KATSU, MASANORI, OGATA, TAKASHI
Priority to US15/254,482 priority patent/US10740057B2/en
Publication of US20160283579A1 publication Critical patent/US20160283579A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • G06F17/30601
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/25Integrating or interfacing systems involving database management systems
    • G06F16/252Integrating or interfacing systems involving database management systems between a Database Management System and a front-end application
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/28Databases characterised by their database models, e.g. relational or object models
    • G06F16/284Relational databases
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/28Databases characterised by their database models, e.g. relational or object models
    • G06F16/284Relational databases
    • G06F16/285Clustering or classification
    • G06F16/287Visualization; Browsing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9537Spatial or temporal dependent retrieval, e.g. spatiotemporal queries
    • G06F17/3056
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0346Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/04817Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance using icons
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0482Interaction with lists of selectable items, e.g. menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04842Selection of displayed objects or displayed text elements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04847Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • G06Q10/109Time management, e.g. calendars, reminders, meetings or time accounting
    • G06Q10/1093Calendar-based scheduling for persons or groups
    • G06Q10/1095Meeting or appointment
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • GPHYSICS
    • G07CHECKING-DEVICES
    • G07CTIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
    • G07C1/00Registering, indicating or recording the time of events or elapsed time, e.g. time-recorders for work people
    • G07C1/10Registering, indicating or recording the time of events or elapsed time, e.g. time-recorders for work people together with the recording, indicating or registering of other data, e.g. of signs of identity
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/02Services making use of location information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/02Services making use of location information
    • H04W4/029Location-based management or tracking services
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/52Network services specially adapted for the location of the user terminal
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/535Tracking the activity of the user
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72448User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions
    • H04M1/72454User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions according to context-related or environment-related conditions

Definitions

  • the present disclosure relates to an information processing device that processes a user's action records, an information processing method, and a computer program.
  • a technology to recognize a user's operation action from sensor information acquired by using various sensing technologies is proposed.
  • the recognized user's operation action is automatically recorded as an action log and can be represented by various techniques, for example, reproducing the operation action by animation such as an avatar, showing a user's movement locus on a map, or using an index abstracting various operation actions for representation.
  • Patent Literature 1 JP 2008-3655A
  • an action log is reproduced by animation such as an avatar using an action recording device like, for example, a motion capture
  • an action log generated by using a small sensor containing recording device such as a smartphone limits the types of action that can be recorded/recognized and thus, it is difficult to present an action record that is valuable to the user. Therefore, an action log is generally shown as a user's movement locus on a map or displayed as an action record converted to the amount of activity like a health index.
  • an information processing device including an action recognition unit that recognizes an operation action of a user based on sensor information, and an action representation generation unit that analyzes operation action data showing the operation action of the user recognized by the action recognition unit to generate an action segment represented by a meaning and content of the operation action from the operation action data.
  • an information processing device including an action recognition unit that recognizes an operation action of a user based on sensor information, an action representation generation unit that generates an action segment constituting an action log from operation action data showing the operation action of the user recognized by the action recognition unit based on operation action estimation information that decides the operation action, and a feedback adjustment unit that corrects the operation action estimation information based on correction feedback from the user to the action segment generated by the action representation generation unit.
  • an information processing method including a step for recognizing an operation action of a user based on sensor information, and a step for analyzing operation action data showing the recognized operation action of the user to generate an action segment represented by a meaning and content of the operation action from the operation action data.
  • an information processing method including a step for recognizing an operation action of a user based on sensor information, a step for generating an action segment constituting an action log from operation action data showing the recognized operation action of the user based on operation action estimation information that decides the operation action, and a step for correcting the operation action estimation information based on correction feedback from the user to the action segment.
  • a computer program for causing a computer to function as an information processing device including an action recognition unit that recognizes an operation action of a user based on sensor information, and an action representation generation unit that analyzes operation action data showing the operation action of the user recognized by the action recognition unit to generate an action segment represented by a meaning and content of the operation action from the operation action data.
  • a computer program for causing a computer to function as an information processing device including an action recognition unit that recognizes an operation action of a user based on sensor information, an action representation generation unit that generates an action segment constituting an action log from operation action data showing the operation action of the user recognized by the action recognition unit based on operation action estimation information that decides the operation action, and a feedback adjustment unit that corrects the operation action estimation information based on correction feedback from the user to the action segment generated by the action representation generation unit.
  • operation action data showing a user's operation action recognized by an action recognition unit based on sensor information is analyzed by an action representation generation unit to generate an action segment represented by the meaning and content of the operation action from the operation action data.
  • an action representation generation unit By displaying an action log with the action segment represented by the meaning and content of the operation action, information can be presented to the user in a manner that is easy to understand.
  • a recorded action log can be presented to the user in a manner that is easy to understand.
  • FIG. 1 is an explanatory view showing the configuration of an action log display system according to an embodiment of the present disclosure.
  • FIG. 2 is a functional block diagram showing a functional configuration of the action log display system according to the embodiment.
  • FIG. 3 is an explanatory view showing an example of a context level dictionary.
  • FIG. 4 is an explanatory view showing a case when an action segment is generated from operation action data by contextual analysis and the display of the action log is changed by changing a segmentation grain size of the action segment.
  • FIG. 5 is an explanatory view showing another example of the case when the action segment is generated from operation action data by the contextual analysis and the display of the action log is changed by changing the segmentation grain size of the action segment.
  • FIG. 6 is an explanatory view showing a case when an action segment is generated from operation action data by combining the contextual analysis and a time width and the display of the action log is changed by changing the segmentation grain size of the action segment.
  • FIG. 7 is an explanatory view showing a case when an action segment is generated from operation action data by combining the contextual analysis, the time width, and position changes and the display of the action log is changed by changing the segmentation grain size of the action segment.
  • FIG. 8 is a flow chart showing the flow of overall processing of action recognition.
  • FIG. 9 is a flow chart showing processing by a living action recognition unit.
  • FIG. 10 is a flow chart showing processing by a hierarchical structure judgment unit.
  • FIG. 11 is a flow chart showing action segment generation processing.
  • FIG. 12 is a flow chart showing the action segment generation processing.
  • FIG. 13 is an explanatory view illustrating a method of attaching hierarchical information to the action segment.
  • FIG. 14 is functional block diagram showing the functional configuration of an analysis server.
  • FIG. 15 is an explanatory view showing an example of the representation of the action log.
  • FIG. 16 is an explanatory view showing a display example of the action log when an action log display application is activated.
  • FIG. 17 is an explanatory view showing a display example of a calendar.
  • FIG. 18 is an explanatory view showing a display example when a map is caused to display position information corresponding to the action log by operating a map button.
  • FIG. 19 is an explanatory view showing a state in which a correction screen to correct the action segment to be corrected is displayed.
  • FIG. 20 is an explanatory view showing an example of the correction screen to correct an operation action.
  • FIG. 21 is an explanatory view showing an example of a method of combining action segments.
  • FIG. 22 is an explanatory view showing an example of another method of deciding operation content of the action segment after combination.
  • FIG. 23 is an explanatory view showing an example of a division method by time settings of the action segment.
  • FIG. 24 is an explanatory view showing an example of the division method based on hierarchical information of the action segment.
  • FIG. 25 is an explanatory view showing a case when display roughness is changed by using a slider.
  • FIG. 26 is an explanatory view showing a case when the display roughness is changed by using a zoom button.
  • FIG. 27 is an explanatory view showing a display change of the action segment when a button of “work details” is checked in an action type selection list.
  • FIG. 28 is an explanatory view showing a display change of the action segment when a button of “shopping details” is checked in the action type selection list.
  • FIG. 29 is an explanatory view showing a display change of the action segment when a button of “movement details” is checked in the action type selection list.
  • FIG. 30 is an explanatory view showing a display change of the action segment when a button of “uniform details” is checked in the action type selection list.
  • FIG. 31 is an explanatory view showing a positional configuration example of a display grain size setting unit provided with a slider that sets the display roughness of the action segment for each type of action.
  • FIG. 32 is an explanatory view showing a method of deleting an action segment from the action log.
  • FIG. 33 is an explanatory view showing a method of posting content of the action segment of the action log to a posting site.
  • FIG. 34 is an explanatory view showing a positional configuration example of a setting screen to make various settings about the action log display application.
  • FIG. 35 is a flow chart showing an example of action recognition processing by the living action recognition unit.
  • FIG. 36 is an explanatory view showing operation action estimation information showing a relationship between a weighting factor depending on the location and a probability distribution of each action.
  • FIG. 37 is an explanatory view providing an overview of reflection processing of correction feedback.
  • FIG. 38 is a flow chart showing the reflection processing of correction feedback of an action.
  • FIG. 39 is an explanatory view illustrating corrections of the operation action estimation information based on the processing in FIG. 38 .
  • FIG. 40 is a flow chart showing other reflection processing of correction feedback of the action.
  • FIG. 41 is a flow chart showing the reflection processing of correction feedback of the action and position information.
  • FIG. 42 is an explanatory view illustrating personal modeling of an action pattern by a typical action pattern generation unit.
  • FIG. 43 is an explanatory view illustrating a position display technique by determining a medium/means of transport.
  • FIG. 44 is an explanatory view illustrating line estimation processing.
  • FIG. 45 is an explanatory view illustrating station name selection processing.
  • FIG. 46 is a block diagram showing a hardware configuration example of an action recording device according to the embodiment.
  • FIG. 1 is an explanatory view showing an outline configuration of the action log display system according to the present embodiment.
  • the action log display system realizes a representation technique that presents an action log recorded by a recording device 100 containing a small sensor (hereinafter, referred to as an “action recording device”) to the user in a manner that is easy to understand.
  • a recording device 100 for example, a mobile terminal such as a mobile phone, PDA (Personal Digital Assistant), and smartphone can be used.
  • the action recording device 100 is provided with at least one sensor to sense conditions or an action of a user holding the device.
  • the action recording device 100 estimates an operation action of the user based on sensor information acquired by the sensor and transmits the operation action to an action log server 200 as an action log. In this manner, the action log of the user is accumulated in the action log server 200 .
  • An action log analyzed by the action recording device 100 and stored in the action log server 200 records an operation like, for example, “meal”, “movement”, and “sleep” together with the action time, position information and the like.
  • An action log display system further analyzes an action log representing the operation content by an analysis server 300 to recognize the meaning of action and generates information (action segment) to which the meaning of action is added.
  • the action segment is unit information as an easy-to-understand representation for the user of an action log. Instead of simply presenting an action log to the user, the action segment can present an action log in a manner that conveys the meaning of action.
  • An action log analyzed by the analysis server 300 and presented to the user can be corrected by the user.
  • data of the presented action log can be combined, divided, or deleted by generating an action segment.
  • the presented action log can also be posted to a posting site.
  • FIG. 2 shows the functional configuration of the action log display system according to the present embodiment.
  • the action log display system includes, as described above, the action recording device 100 that records a user's operation action, the action log server 200 that manages an action log recorded by the action recording device 100 , and the analysis server 300 that analyzes an action log to generate an action segment.
  • the action recording device 100 includes sensors 110 , an action recognition unit 120 , a client interface unit 130 , an action representation processing unit 140 , a display unit 150 , and an input unit 160 .
  • the sensors 110 are devices that sense a user's action or conditions and are installed in the action recording device 100 .
  • the sensors 110 for example, an acceleration sensor, gyro sensor, magnetic field sensor, atmospheric pressure sensor, illuminance sensor, temperature sensor, microphone and the like can be used.
  • a latitude/longitude acquisition sensor that acquires the latitude/longitude can also be installed.
  • the latitude/longitude acquisition sensor for example, not only GPS (Global Positioning System) or WiFi, but also base station information of other communication networks or information such as RFID and images may be used.
  • the sensors 110 output detected information to the action recognition unit 120 as sensor information.
  • the action recognition unit 120 estimates a user's action based on sensor information.
  • the action recognition unit 120 includes a sensor controller 122 and an operation action recognition unit 124 .
  • the sensor controller 122 controls the sensor 110 , the CPU or the overall system to cause sensing by the sensor 110 to operate effectively.
  • the sensor controller 122 controls the above devices based on recognition results by the sensor 110 or the operation action recognition unit 124 .
  • the operation action recognition unit 124 recognizes a user's action or conditions by performing signal processing or statistical processing of sensor information.
  • the action recording device 100 holds a correspondence between an action model as information about a user's action obtained as a result of processing sensor information and an operation action in advance.
  • the operation action recognition unit 124 identifies an operation action corresponding to the parameters. Then, the operation action recognition unit 124 associates the identified operation action and the action time period, action time, position information and the like and outputs the associated information to the client interface unit 130 as operation action data.
  • the operation action data is uploaded from the client interface unit 130 to the action log server 200 .
  • the client interface unit 130 transmits/receives information between the action recording device 100 , and the action log server 200 and the analysis server 300 .
  • the client interface unit 130 transmits operation action data input from the action recognition unit 120 to the action log server 200 or outputs an analysis result received from the analysis server 300 to the action representation processing unit 140 .
  • the client interface unit 130 transmits feedback information from the user input through the input unit 160 to the analysis server 300 .
  • the action representation processing unit 140 is a functional unit that displays an action log or processes feedback information from the user and includes a display processing unit 142 and an input information processing unit 144 .
  • the display processing unit 142 performs processing to display an analysis result by the analysis server 300 input from the client interface unit 130 in the display unit 150 .
  • the input information processing unit 144 performs processing to transmit feedback information from the user for an action log input from the input unit 160 to the analysis server 300 via the client interface unit 130 .
  • the display unit 150 is an output device that displays information and can be configured by, for example, a liquid crystal display, organic EL display or the like. For example, an action log processed for display by the display processing unit 142 is displayed in the display unit 150 .
  • the input unit 160 is an input device to input information and, for example, a touch panel, keyboard, hardware button or the like can be used.
  • a touch panel for example, a touch panel, keyboard, hardware button or the like
  • the display surface of the display unit 150 is provided with a touch panel as the input unit 160 .
  • the user can input information by, for example, bringing an operation body such as a finger or touch pen into contact with the display surface of the display unit 150 or moving the operation body brought into contact with the display surface.
  • Information input from the input unit 160 is output to the input information processing unit 144 .
  • the action log server 200 includes a log server interface unit 210 and an action log DB 220 .
  • the log server interface unit 210 transmits/receives information between the action log server 200 , and the action recording device 100 and the analysis server 300 .
  • the log server interface unit 210 records operation action data received from the action recording device 100 in the action log DB 220 or acquires operation action data in accordance with a transmission request from the analysis server 300 from the action log DB 220 and transmits the operation action data to the analysis server 300 .
  • the action log DB 220 is a storage unit that stores operation action data of the user acquired by the action recording device 100 .
  • operation action data stored in the action log DB 220 as described above, the operation action identified by the operation action recognition unit 124 and the action time period, action time, position information and the like are associated and stored in the action log DB 220 in, for example, chronological order.
  • the analysis server 300 includes an analysis server interface unit 310 , an action representation generation unit 320 , and a data management unit 330 .
  • the analysis server interface unit 310 transmits/receives information between the analysis server 300 , and the action recording device 100 and the action log server 200 .
  • the analysis server interface unit 310 receives an analysis instruction (analysis request) of an action log from the action recording device 100 or transmits a transmission request of necessary operation action data in accordance with an analysis request.
  • the analysis server interface unit 310 also receives feedback information from the user of an action log from the action recording device 100 .
  • the action representation generation unit 320 analyzes operation action data to understand the meaning thereof and generates an action segment to which the meaning and content is added.
  • the action representation generation unit 320 includes a living action recognition unit 321 and a hierarchical structure judgment unit 322 .
  • the living action recognition unit 321 generates an action segment from an action log including operation action data.
  • the living action recognition unit 321 analyses the meaning and content of operation action data arranged in chronological order based on relationships between data and the time period, time and the like of data. Then, the living action recognition unit 321 selects data classified as the most detailed meaning and content of analyzed meaning and content as the action segment.
  • the generated action segment is output to the data management unit 330 and held there.
  • the hierarchical structure judgment unit 322 judges a hierarchical structure about the meaning and content of an action segment generated by the living action recognition unit 321 and attaches hierarchical information representing a hierarchical relationship of the meaning and content to the action segment.
  • Hierarchical information is hierarchical meaning information attached to an action segment by processing described later.
  • Hierarchical information may be, for example, information using a normalized value as a key or information using ID identifying the level of meaning information as a direct key.
  • An action segment to which hierarchical information is attached is also called a hierarchical information attached action segment.
  • Hierarchical relationships of the meaning and content of action are stored in the data management unit 330 .
  • the hierarchical structure judgment unit 322 outputs a hierarchical information attached action segment to the data management unit 330 via the living action recognition unit 321 .
  • the function of the action representation generation unit 320 and details of processing content thereby will be described later.
  • the data management unit 330 manages an action segment generated by the action representation generation unit 320 .
  • the data management unit 330 includes a data acquisition unit 331 , a feedback adjustment unit 332 , an analysis parameter DB 333 , a unit data storage DB 334 , and a hierarchical information attached data storage DB 335 .
  • the data acquisition unit 331 transmits/receives data to/from the action representation generation unit 320 .
  • the data acquisition unit 331 records an action segment transmitted from the action representation generation unit 320 in the unit data storage DB 334 or records a hierarchical information attached action segment in the hierarchical information attached data storage DB 335 .
  • the data acquisition unit 331 acquires the specified action segment in accordance with a request from the action representation generation unit 320 from the unit data storage DB 334 or the hierarchical information attached data storage DB 335 and outputs the action segment to the action representation generation unit 320 .
  • the feedback adjustment unit 332 reflects feedback information received from the action recording device 100 in analysis parameters used for analyzing the meaning and content of operation action data.
  • the feedback information represents content of processing such as corrections made by the user on an action log displayed in the display unit 150 of the action recording device 100 .
  • the feedback adjustment unit 332 corrects analysis parameters using feedback information so that the meaning and content of a user's action can be recognized more correctly.
  • the analysis parameter DB 333 is a storage unit that holds analysis parameters used for analyzing the meaning and content of operation action data.
  • a correspondence between an operation action and the meaning and content is stored as analysis parameters.
  • Information stored in the analysis parameter DB 333 can be referenced by both of the living action recognition unit 321 and the hierarchical structure judgment unit 322 . Analysis parameters are updated when necessary based on feedback information from the user.
  • the unit data storage DB 334 stores an action segment generated by the action representation generation unit 320 .
  • the action segment stored in the unit data storage DB 334 is a segment (unit segment) of the minimum unit necessary for recognition.
  • the hierarchical information attached data storage DB 335 stores an action segment in which hierarchical information is attached to an action segment generated by the action representation generation unit 320 .
  • the action segment stored in the hierarchical information attached data storage DB 335 is a hierarchical information attached action segment to which hierarchical information representing a hierarchical structure of the meaning and content of action is attached by the hierarchical structure judgment unit 322 .
  • the recording timing of an action segment to which hierarchical information is attached may be, for example, when requested by an application or analysis results of a plurality of segmentation grain sizes may be recorded in advance by the action representation generation unit 320 .
  • the action representation generation unit 320 and the data management unit 330 function as information processing devices that analyze the meaning and content of operation action data generated by the action recording device 100 to present information that is easy for the user to understand.
  • the meaning and content of an operation action generated by the action recording device 100 is analyzed by the analysis server 300 to generate an action segment based on the meaning and content of action.
  • Hierarchical information about the meaning and content of action can also be attached to an action segment and the display form of an action log can also be changed easily based on the hierarchical information.
  • generation processing of an action segment will be described based on FIGS. 3 to 13 .
  • the action recording device 100 analyzes an operation action, for example, “meal”, “movement”, or “sleep”.
  • the analysis server 300 analyzes content of each operation action more deeply using operation action data containing the operation action.
  • the analysis of the meaning and content of the operation action is conducted by using, for example, as shown in FIG. 3 , a context level dictionary.
  • the operation action of “movement” is taken up, As shown in FIG. 3 , the meaning and content thereof changes depending on what kind of movement. For example, when “moving on foot”, the user can take action of continuing to “walk” or then change to an action of “halting”. When, for example, “waiting for means of transport” on the move, action of continuing to “wait for means of transport” can be taken. Further, when, for example, “moving by train”, the movement can be made an action of “movement by train on a single line”. Further, a state in which a state of “movement by rain” continues or a state of “train stopped” in which the train on which the user moves stops can be assumed. Alternatively, action of “trains changed” after “movement by train” can be taken or a state transition to “waiting for means of transport” can also take place.
  • a further action of “walking” can be associated with a “walking” action or “changing trains” action and a further action of “stopping” can be associated with a “halting” action, a “waiting for means of transport” state, or a “train stop” state.
  • a “train” as a means of transport can further be associated with a state of “moving by train”.
  • action meta information at an operation action level could change to, as shown in FIG. 3 , a higher level of action meta information depending on the context.
  • An action log display system in the present embodiment can analyze the meaning and content of an action based on the relationship between an operation action and the meaning and content and so can present an action log that is easy for the user to understand.
  • the segmentation grain size of an action log described later can easily be changed.
  • the ontology/semantic technology is applied to the recognition of the meaning and content of an operation action to segment the action recognition that judges “context” in an “action” and operation action data.
  • Ontology systematically represents the concept of relations between words and in the present embodiment, for example, as shown in FIG. 3 , the concept of relations between actions is systematized. Then, using the systematized concept of relations between actions, the meaning of an action or the relationship of actions is understood by applying the semantic technology and recognition processing of the meaning and content of the operation action is performed based on the understanding.
  • an action segment fitting to the feeling of the user can be generated.
  • FIG. 4 is an explanatory view showing a case when an action segment is generated from operation action data by the contextual analysis and the display of the action log is changed by changing a segmentation grain size of the action segment.
  • FIG. 5 is an explanatory view showing another example of the case when the action segment is generated from operation action data by the contextual analysis and the display of the action log is changed by changing the segmentation grain size of the action segment.
  • an action log including operation action data is acquired by the operation action recognition unit 124 .
  • the operation action data is arranged from left to right in chronological order.
  • the operation action data is an operation action recognized based on sensor information of the sensors 110 and an action log is represented by a simple operation action like “walked”, “stopped”, and “got on a train”.
  • the living action recognition unit 321 recognizes the meaning of each piece of operation action data or the relationship between operation action data using a dictionary as shown in FIG. 3 from such an action log.
  • a state of short “stopped” of a predetermined time or shorter between operation action data of “got on a train” is estimated to be a “train stopped (at a station)” state.
  • a state of short “walked” of a predetermined time or shorter between operation action data of “got on a train” is estimated to be a “changing trains” action.
  • “stopped” of operation action data immediately before operation action data of “got on a train” is estimated to be a state of “waited for a train”.
  • the action can be estimated more appropriately.
  • an action of “movement by train” for example, the meaning of action of “going to office” or “going to school” can be estimated if the action time is a morning hour (for example, from 6 am to 10 am) and the meaning of action of “going home” can be estimated if the action time is an evening hour (for example, from 5 pm to 8 pm).
  • the meaning of action of “breakfast” can be estimated if the action time is a morning hour
  • “lunch” can be estimated if the action time is an hour around noon
  • “supper” can be estimated if the action time is an evening hour.
  • an action log including action segments as shown in FIG. 4 is generated by an action log including operation action data being analyzed by the living action recognition unit 321 .
  • An action segment is data representing an operation action to which operation content is added and is a chunk of consecutive operation actions having the same meaning and content.
  • An action segment generated to match operation action data includes unit segments to which detailed operation content is added. Therefore, if an action log is represented by action segments, as shown in FIG. 4 , the moving state while riding on a train can also be known.
  • the display of the action segment can be changed.
  • a plurality of action segments that can be considered to be one action is combined to produce an action segment representing rough operation content.
  • an action segment approaches the unit segment.
  • the segmentation grain size of an action log represented by unit segments is increased.
  • a state of a grain size 1 - 1 reached by increasing the segmentation grain size from the segmentation grain size of the unit segment a series of operations of “got on a train”, “train stopped”, and “got on a train” are represented by one action segment of “got on a train on some line”.
  • a state of a grain size 1 - 2 reached by further increasing the segmentation grain size a series of operations of “got on a train on some line”, “trains changed”, and “got on a train on some line” are represented by one action segment of “movement by train”.
  • a series of operations of “walked”, “waited for a train”, “movement by train” and “walked” are represented by one action segment of “moved”.
  • the action log can be displayed at an operation action level that is easy for the user to view.
  • Action model X representing a conspicuously unsteady motion appearing in a shopping action or the like and an operation action of “had a meal” are recognized by the operation action recognition unit 124 as operation action data.
  • “Action model X” normally means the class of shopping, but the meaning thereof changes in accordance with a prior or subsequent determination result.
  • action model X is present before and after “had a meal”.
  • the living action recognition unit 321 recognizes the action model X as “accepted” and “paid the bill” as operations performed before and after an operation of “have a meal” from operation action data before or after the action model X.
  • an action log that is easy for the user to understand can be presented by the meaning and content being added to the user's unsteady motion from the prior or subsequent operation.
  • a series of actions of “accepted”, “had a meal”, and “paid the bill” can be represented as one action segment of “had a meal” by increasing the segmentation grain size.
  • FIG. 6 is an explanatory view showing a case when an action segment is generated from operation action data by combining the contextual analysis and the time width and the display of the action log is changed by changing the segmentation grain size of the action segment.
  • an action log including operation action data is acquired by the operation action recognition unit 124 .
  • an action log is represented by simple operation actions like “desk work”, “walked”, “advance arrangements”, and “meal” as operation action data.
  • the living action recognition unit 321 recognizes the meaning of each piece of operation action data or the relationship between operation action data using a dictionary as shown in FIG. 3 and also recognizes the meaning and content thereof by considering the time width of the operation action data.
  • FIG. 6 shows an example of the action log in a company. What kind of walking the operation action data of “walked” is about can be recognized from prior or subsequent operation action data, but can also be estimated based on a walking time. An operation action of short “walked” in a company normally does not have any special meaning. However, if the walking continues for a predetermined time or longer, the walking is estimated not to be simple movement on the floor, but to be movement between premises. Thus, action segments generated by combining the contextual analysis using a dictionary and the time width of operation action data are as shown in FIG. 6 . The operation action data of “walked” is divided into “short walking” and “movement between premises” depending on the time width thereof.
  • the display of the action log including the action segments can easily be changed by changing the segmentation grain size.
  • a series of operations of “desk work”, “short walking”, “advance arrangements”, “short walking”, and “desk work” are represented as one action segment of “worked”.
  • “short walking” is combined into one action segment of “worked” and thus, the action segment of “movement between premises” may be displayed simply as “movement”.
  • FIG. 7 is an explanatory view showing a case when an action segment is generated from operation action data by combining the contextual analysis, the time width, and position changes and the display of the action log is changed by changing the segmentation grain size of the action segment.
  • an action log including operation action data is acquired by the operation action recognition unit 124 .
  • an action log is represented by simple operation actions like “did shopping”, and “walked” as operation action data.
  • the living action recognition unit 321 recognizes the meaning of each piece of operation action data or the relationship between operation action data using a dictionary as shown in FIG. 3 and also recognizes the meaning and content thereof by considering the time width of the operation action data and position changes of the action recording device 100 (that is, the user).
  • FIG. 7 shows an example of the action log in shopping. What kind of walking the operation action data of “walked” is about can be recognized from prior or subsequent operation action data, but can also be estimated in detail based on a walking time and position changes.
  • the user is estimated to “move between shops”. Also, if, for example, the operation actions before and after the operation action data of “walked” are “did shopping” and the movement time t is t 2 (for example, 20 s) or more and shorter than t 1 , and no position change of the action recording device 100 is measured, the user is estimated to “move between floors” during shopping.
  • t 1 for example, 35 s
  • t 2 for example, 20 s
  • the operation actions before and after the operation action data of “walked” are “did shopping” and the movement time t is t 3 (for example, 5 s) or more and shorter than t 2 , and no position change of the action recording device 100 is measured, the user is estimated to “move in a shop” during shopping.
  • action segments including three action segments of “movement in a shop (SG 1 )”, “movement on floors (SG 2 )”, and “movement between shops (SG 3 )” are generated from operation action data of “walked”.
  • the display of the action log including the action segments can easily be changed by changing the segmentation grain size. For example, in a state of a grain size 3 - 1 reached by increasing the segmentation grain size from the segmentation grain size of the unit segment, among action segments concerning walking, the action segment SG 1 with the shortest walking time is combined with the action segments of “did shopping” preceding and succeeding the action segment SG 1 .
  • These action segments are represented by an action segment as a series of operations of “did shopping”. In this case, “movement in a shop” is combined into one action segment of “did shopping”, other action segments concerning walking may be displayed simply as “movement”.
  • the action segment SG 2 with the shortest walking time next to the action segment SG 1 is combined with the action segments of “did shopping” preceding and succeeding the action segment SG 2 .
  • the action segment SG 3 with the longest walking time is also combined with the action segments of “did shopping” preceding and succeeding the action segment SG 3 . Accordingly, a series of operations of “did shopping” and “walked” are represented as one action segment.
  • FIG. 8 is a flow chart showing the flow of overall processing of action recognition.
  • FIG. 9 is a flow chart showing processing by the living action recognition unit 321 .
  • FIG. 10 is a flow chart showing processing by the hierarchical structure judgment unit 322 .
  • Action recognition processing includes, as shown in FIG. 8 , operation action data creation processing (S 100 , S 110 ) performed by the action recognition unit 120 of the action recording device 100 and action segment generation processing (S 120 to S 140 ) performed by the action representation generation unit 320 and the data management unit 330 of the analysis server 300 .
  • the operation action recognition unit 124 of the action recognition unit 120 having acquired sensor information from the sensors 110 starts creation processing of operation action data (S 100 ).
  • the operation action data creation processing can be performed by using an existing technique.
  • the operation action recognition unit 124 outputs the operation action data to the action log server 200 (S 110 ).
  • operation action data constituting an action log of the user is accumulated in the action log server 200 .
  • the action recognition unit 120 may generate, as operation action data, not only action information at the operation level, but also information including, for example, time information, location information, an operation history of devices and the like.
  • the action representation generation unit 320 of the analysis server 300 analyzes the meaning and content of the operation action data through the living action recognition unit 321 (S 120 ).
  • the living action recognition unit 321 segments the operation action data into data of a preset unit length and attaches living action meta information to each piece of segmented data.
  • the living action recognition unit 321 After segmenting the operation action data into the unit time T in chronological order, the living action recognition unit 321 first determines whether an integrated value of the unit length (T) and the parameter i is smaller than the length (time) of the operation action data (S 121 ). If it is determined in step S 121 that the integrated value of the unit length (T) and the parameter i is smaller than the length (time) of the operation action data, the living action recognition unit 321 attaches living action meta information to the segmented data between time T*i and time T*(i+1) (step S 122 ).
  • the symbol “*” indicates integration processing.
  • the meaning and content (living action meta information) applicable to the segmented data at the time can be attached by using, for example, ruled-based branching processing.
  • living action meta information can also be attached the segmented data using machine learning such as the Hidden Markov Model (HMM) or Neural Network.
  • HMM Hidden Markov Model
  • Neural Network The number of pieces of living action meta information attached to the segmented data is not limited to one and a plurality of pieces of living action meta information may be attached.
  • the living action recognition unit 321 When living action meta information is attached to the segmented data of the operation action data in step S 122 , the living action recognition unit 321 adds 1 to the parameter i (S 123 ) to repeat the processing from step S 121 . If it is determined in step S 121 that the integrated value of the unit length (T) and the parameter i is equal to or greater than the length (time) of the operation action data, the living action recognition unit 321 outputs each piece of segmented data to which living action meta information is attached by the processing in step S 122 as living action data (S 124 ). An output result of the living action recognition unit 321 may be recorded in a predetermined storage unit (not shown) or may be output directly to the functional unit (in this case, the hierarchical structure judgment unit 322 ) that performs the next processing.
  • the hierarchical structure judgment unit 322 attaches hierarchical information representing a hierarchical relationship about the meaning and content to the living action data. The processing will be described based on FIG. 10 .
  • the hierarchical structure judgment unit 322 first determines whether the input living action data satisfies at least one of conditions that no hierarchical information is attached and it is possible to further shift to a higher level from the attached hierarchical information (S 131 ).
  • step S 131 If none of these conditions is satisfied in step S 131 , that is, the highest hierarchical information is already attached, the hierarchical structure judgment unit 322 terminates the processing shown in FIG. 10 .
  • the hierarchical structure judgment unit 322 combines adjacent segmented data of the same action of each piece of segmented data arranged in chronological order (S 132 ).
  • step S 132 processing to regard data discretized in step S 120 as one operation action when the same action continues is performed.
  • a chunk (segment) generated by combining segmented data may be recorded in a predetermined storage unit (not shown) (S 133 ).
  • the hierarchical structure judgment unit 322 determines whether j is smaller than the number of segments generated in step S 131 and dictionary information about the action of the j-th segment can be acquired (S 134 ). If both of these conditions are satisfied in step S 134 , the hierarchical structure judgment unit 322 selects and attaches dictionary information optimum to the relevant segment of the acquired dictionary information (S 135 ). Then, the hierarchical structure judgment unit 322 determines whether still higher hierarchical information can be attached to the segment (j) based on the selected dictionary information and temporal context (S 136 ).
  • step S 136 If it is determined in step S 136 that higher hierarchical information can be attached, the hierarchical structure judgment unit 322 attaches higher hierarchical information to the segment (j) and adds 1 to the parameter j (S 137 ). Then, the processing in step S 134 and thereafter is repeated.
  • step S 134 if it is determined in step S 134 that j is equal to or greater than the number of segments generated in step S 132 or dictionary information about the action of the j-th segment cannot be acquired, the hierarchical structure judgment unit 322 repeats the processing in step S 131 and thereafter. Also when it is determined in step S 136 that higher hierarchical information cannot be attached to the segment (j), the hierarchical structure judgment unit 322 repeats the processing in step S 131 and thereafter.
  • the hierarchical structure judgment unit 322 segments segmented data generated in step S 120 by piecing together segmented data as the same action that is temporally consecutive. Then, related dictionary information is acquired and attached to each segment and whether still higher hierarchical information can be added is judged. Thus, by performing the processing in steps S 131 to S 137 , an action segment (unit segment) in the minimum unit to which hierarchical information is attached is generated.
  • FIGS. 11 and 12 show a flow chart showing an example of action segment generation processing.
  • the hierarchical structure judgment unit 322 determines whether “another action” sandwiched between the same action of segmented data is present (S 200 ). If “another action” sandwiched between the same action is present, “another action” is merged into the action before and after “another action” (S 202 ). However, merge processing of “another action” sandwiched between operation actions of “walking” whose precision is sufficiently high and which is likely to be a point of change is not performed. After “another action” is merged in step S 202 or there is no “another action” sandwiched between the same action in step S 200 , the hierarchical structure judgment unit 322 merges the same actions that are consecutive (S 204 ).
  • the hierarchical structure judgment unit 322 merges the segment (k) into the action immediately before (S 214 ). On the other hand, if the action time of the segment (k) is equal to or longer than the predetermined time T 2 in step S 212 , the hierarchical structure judgment unit 322 decides the operation action of the segment as “another action” (S 216 ).
  • the hierarchical structure judgment unit 322 determines whether the processing of steps S 208 to S 216 has been performed for all segments (S 218 ) and if there is any unprocessed segment, the hierarchical structure judgment unit 322 adds 1 to k (S 220 ) and then repeats the processing in step S 208 and thereafter,
  • steps S 208 to S 216 have been performed for all segments, as shown in FIG. 12 .
  • processing to merge consecutive segments of the same action is performed (S 222 ).
  • a vehicle action of only the data unit time for example, only 1 min
  • the action of the segment is set as “another action” (S 224 ).
  • the hierarchical structure judgment unit 322 determines whether action content of the segment is “walking” (S 226 ) and, if the action content is other than “walking”, accumulates the segment in the buffer (S 228 ). On the other hand, if the action content of the segment is “walking”, the hierarchical structure judgment unit 322 determines whether any vehicle action is accumulated in the buffer (S 230 ). If a vehicle action is accumulated in the buffer, the hierarchical structure judgment unit 322 sets the operation action of the segment as an action of vehicle with the maximum share from “walking” (S 323 ). On the other hand, if no vehicle action is accumulated in the buffer, If a vehicle action is accumulated in the buffer, the hierarchical structure judgment unit 322 sets the operation action of the segment as “another action” (S 324 ).
  • FIG. 13 it is assumed that the operation content of six unit segments (SG 01 to SG 06 ) is “train”, “train”, “walking”, “walking”, “train”, and “train”.
  • the unit segment SG 01 and the unit segment SG 02 are merged into an action segment SG 07
  • the unit segment SG 03 and the unit segment SG 04 are merged into an action segment SG 08
  • the unit segment SG 05 and the unit segment SG 06 are merged into an action segment SG 09 by the processing shown in FIG. 10 .
  • the state is set as the grain size 1 - 0 .
  • the action segments SG 07 to SG 09 in the grain size 1 - 0 becomes an action segment SG 17 of “got on a train on some line (action B)”, an action segment SG 18 of “trains changed (action C)”, and an action segment SG 19 of “got on a train on some line (action C)” respectively.
  • the grain size at this point is set as 1 - 1 . If the action segments SG 17 to SG 19 are further merged from the grain size 1 - 1 , one action segment SG 10 of “got on a train (action A)” is obtained.
  • the grain size at this point is set as 1 - 2 .
  • An action segment holds hierarchical information by being attached to the action segment combining unit segments or by being attached to the unit segments.
  • hierarchical information is attached to an action segment combining unit segments, it is assumed that, for example, an action segment SG 17 of the action B in FIG. 13 is focused on.
  • the action segment SG 17 is considered to be an action segment newly generated by combining the unit segments SG 01 and SG 02 .
  • the facts that the action segment has the grain size 1 - 1 and the action content is “got on a train on some line”, and the start time and the end time of the action are attached to the action segment SG 17 as hierarchical information.
  • the fact that the action segment SG 17 includes the unit segments SG 01 and SG 02 may be attached to the action segment SG 17 as hierarchical information.
  • Hierarchical information may be attached in any form and can also be attached in other forms.
  • the case of attaching hierarchical information to an action segment combining unit segments is superior in terms of the amount of data and the case of attaching hierarchical information to unit segments is superior in terms of a database search.
  • the living action recognition unit 321 outputs each action segment to the data management unit 330 (S 140 ).
  • the data management unit 330 records the acquired action segment in a storage unit (the unit data storage DB 334 or the hierarchical information attached data storage DB 335 ).
  • the analysis server 300 can accumulate an action log by action segments in real time and at the same time, can generate an action representation based on the meaning and content of an operation action.
  • the analysis server 300 can also generate an action representation from a past action history.
  • a detailed configuration of the action representation generation unit 320 and the data management unit 330 of the analysis server 300 is shown in FIG. 14 .
  • FIG. 14 is functional block diagram showing the functional configuration of the analysis server 300 .
  • the action representation generation unit 320 can include, in addition to the living action recognition unit 321 and the hierarchical structure judgment unit 322 described above, a hierarchical processing unit 323 , a registration processing unit 324 , a comment creation unit 325 , and an acquisition unit 326 .
  • the hierarchical processing unit 323 performs subsequent processing of a judgment result of the hierarchical structure judgment unit 322 .
  • the hierarchical processing unit 323 functions based on a hierarchical structure when only a portion of data to be attached to action segments is recorded in the storage unit for slimming down or speedup of data or hierarchical information of the specified action segment is delivered to an application.
  • hierarchical information may be attached to an action segment combining unit segments or to the unit segments.
  • the hierarchical processing unit 323 processes the action segment of the hierarchical information selected by the user via the input unit 160 .
  • the hierarchical processing unit 323 generates an action segment by combining unit segments based on the hierarchical information selected by the user via the input unit 160 .
  • the hierarchical processing unit 323 a processing result of information to the registration processing unit 324 and the comment creation unit 325 .
  • the registration processing unit 324 records the action segment generated by the hierarchical processing unit 323 in the data management unit 330 .
  • the registration processing unit 324 outputs an action segment to the data acquisition unit 331 to record the action segment in the hierarchical information attached data storage DB 335 .
  • the comment creation unit 325 creates and attaches a comment such as the meaning and content of an action to a generated action segment.
  • a comment created by the comment creation unit 325 is output to the data acquisition unit 331 .
  • the data acquisition unit 331 associates the comment with the corresponding action segment and records the comment in, for example, the hierarchical information attached data storage DB 335 .
  • the acquisition unit 326 acquires a predetermined action segment from the unit data storage DB 334 or the hierarchical information attached data storage DB 335 .
  • the acquisition unit 326 past data recorded in the unit data storage DB 334 or the hierarchical information attached data storage DB 335 .
  • Data to be acquired is decided based on instructions from the user.
  • an action segment to which the meaning and content of an operation action is attached is generated.
  • An application function to represent an action log of the user using action segments will be described below.
  • FIG. 15 An example of the action log is shown in FIG. 15 .
  • the action log in FIG. 15 can be displayed, for example, in the display unit 150 of the action recording device 100 .
  • An action log includes action segments arranged in chronological order. For each action segment, for example, the start time and the end time of the operation action and operation content are displayed.
  • the operation content is a movement action like “movement by train”
  • a position representation like, for example, from the start location to the goal for example, “from Gotanda to Ohsaki”
  • the location for example, “in Ohsaki” where the operation is performed is added to operation content.
  • an object of operation content may be displayed or an object showing the feeling of the user when performing the operation may also be displayed.
  • the number of steps (step value) for the user to perform the operation or a value (exercise value) indicating energy consumption by the operation may be displayed.
  • the content displayed in each action segment constituting an action log is not limited to the example of FIG. 15 and information obtained from sensor information acquired by sensors may further be displayed in each action segment.
  • the action log shown in FIG. 15 there are locations where times if adjacent action segments are not continuous. For example, there is a gap between the action of “meal at Ohsaki” started at 12:30 and the action of “was in Shibuya” started at 14:30. In this manner, the action log may be prevented from displaying all action segments.
  • the non-display of an action segment may be caused by, for example, editing/deletion of the action segment by the user or by setting a non-display filter to prevent a portion of action segments from being displayed.
  • non-display filter for example, a filter that prevents the display when the action time is short or a filter that prevents the display of an action segment judged to be unimportant to the use can be considered. Also, a filter that prevents the display when the precision of recognition is low or a filter that allows the display of an action or location specified by the user may be set.
  • FIG. 16 shows a display example of the action log when the action log display application is activated.
  • FIG. 16 shows a state in which an action log 410 in units of days is displayed in the display unit 150 of the action recording device 100 .
  • the action log 410 includes action segments 412 arranged in chronological order, for example, from the upper end of the screen toward the lower end.
  • the location where an operation is performed as described in FIG. 15 , the location where an operation is performed, description of the location, type of the action, number of steps by the user for the operation, exercise value and the like are displayed. If the action log for one day cannot be displayed in the screen of display, the display unit 150 can be caused to display an action segment that is not displayed by performing a screen scroll operation.
  • the display unit 150 can be caused to display an action log of another day. If the action log of the previous day is not present when the Prev button 422 is pressed, the display unit 150 may be caused to display an action log of the day when an action log is acquired next by further going back to the past. Similarly, if the action log of the next day is not present when the Next button 424 is pressed, the display unit 150 may be caused to display an action log of the day when an action log is acquired next by further moving to the present. Operation buttons 430 for browsing, editing and other operations of the action log 410 are displayed in the display unit 150 and the user can touch the button corresponding to desired processing to perform the processing.
  • a calendar 440 is displayed in the display unit 150 .
  • the calendar 440 displayed first in the display unit 150 after the screen transition may be the current month or the month corresponding to the action log 410 displayed before the screen transition.
  • a day for which the action log 410 can be displayed that is, a day for which the action log 410 is accumulated as data can be selected by, for example, a touch operation.
  • the action log 410 of the day is displayed in the display unit 150 .
  • the display of the calendar 440 is changed by a previous ( ⁇ ) button 442 or a next (>) button 444 . If the previous ( ⁇ ) button 442 is operated, the calendar of the previous month is displayed and if the next (>) button 444 is operated, the calendar of the next month is displayed. If no action log of the previous month is present when the previous ( ⁇ ) button 442 is pressed, the calendar 440 of a month when any action log is acquired next may be displayed by further going back to the past. Similarly, if no action log of the next month is present when the next (>) button 444 is pressed, the calendar 440 of a month when any action log is acquired next may be displayed by further moving to the present.
  • FIG. 17 shows an example in which the calendar 440 in units of months is displayed, but the present technology is not limited to such an example and can display a calendar, for example, in units of weeks, two weeks, or years.
  • an icon indicating a representative operation action for the day may be displayed for all days for which an action log can be displayed in the calendar 440 . Accordingly, the user can recognize the action of the day at a glance.
  • the action log 410 desired to browse can easily be searched for and the display unit 150 can be caused to display the action log.
  • a Back button 446 may be operated.
  • the action log display application activates a map 450 to display position information corresponding to the action log 410 in the map 450 .
  • FIG. 18 shows a display example when the map 450 is caused to display position information corresponding to the action log 410 by operating the map button 431 .
  • the map button 431 is touched while none of the action segments 412 constituting the action log 410 is selected, for example, a history of all position information of the action log 410 of the day is displayed on the map 450 . If the user is on the move, a movement locus thereof is displayed on the screen.
  • the map button 431 is touched while one action segment 412 a is selected from the action log 410 , a history of position information of the action segment 412 a is displayed on the map 450 . If, for example, as shown on the left of FIG. 18 , the action segment 412 a of “movement from Saginomiya to Ohsaki by train” is selected and the map button 431 is touched, as shown on the right of FIG. 18 , a movement locus by train from the start location (Saginomiya) to the goal (Ohsaki) is displayed on the map 450 . Accordingly, on which line the user moved can be presented in an easy-to-understand manner. While position information is displayed in the map 450 , time information 452 of the position information may be displayed in the display unit 150 .
  • an icon or the like may be displayed in a location where the operation is performed.
  • FIG. 18 a case when the one action segment 412 is selected is described, but the present technology is not limited to such an example and a plurality of the action segments 412 can be selected and position information of all the selected action segments 412 can be displayed on the map 450 .
  • the position information on the map 450 may be distinguished by color-coding or the like for each action segment and displayed.
  • a Back button 454 may be operated.
  • FIG. 19 is an explanatory view showing a state in which a correction screen to correct the action segment to be corrected is displayed.
  • FIG. 20 is an explanatory view showing an example of the correction screen to correct an operation action.
  • the user selects the action segment 412 a to be corrected from the action log 410 displayed in the display unit 150 and touches an edit button 435 . Then, as shown on the right of FIG. 19 , a correction screen 460 to correct the action segment 412 a to be corrected is displayed. If, for example, the correction of the action segment 412 a of “movement from Saginomiya to Ohsaki by train” is selected, a screen transition occurs from the display screen of the action log 410 to the correction screen 460 in which the operation content, start location and end location of the operation, and feeling during operation can be edited.
  • the operation content can be corrected in an operation content correction area 461 of the correction screen 460 . If, for example, the operation content correction area 461 is selected, as shown in FIG. 20 , an operation content candidate list 461 a in which operation content candidates for correction are listed is displayed. The user can correct the operation content by selecting the correct operation content from the operation content candidate list 461 a . When “Free input” is selected from the operation content candidate list 461 a , an input field into which the user can freely input operation content is displayed and the correct operation content can be input.
  • the user After selecting operation content from the operation content candidate list 461 a , the user continues to correct the start location and end location of the operation. At this point, if the selected operation content is a movement action like, for example, “movement by bus”, corrections of a start location correction area 462 , a start location description correction area 463 , an end location correction area 464 , or an end location description correction area 465 can be made.
  • a movement action like, for example, “movement by bus”
  • a location name list may be displayed for the start location correction area 462 and the end location correction area 464 so that the user can select and input the location name or the user may be enabled to directly input the location name.
  • location names to be a landmark such as a building name, station name, or shop name may be displayed. If there is no location to be a landmark, place names (addresses) may be displayed in the location name list.
  • the display unit 150 may be caused to display a start location map 462 a and an end location map 464 a that display a map by being linked to input content of the start location correction area 462 and the end location correction area 464 .
  • the start location map 462 a and the end location map 464 a can be caused to display a map of any location by a scroll operation on the map.
  • the location name corresponding to the position where the touch operation is performed may automatically be input into the start location correction area 462 or the end location correction area 464 .
  • the start location description correction area 463 and the end location description correction area 465 are areas where what kind of location the location input into the start location correction area 462 and the end location correction area 464 is for the user is input respectively.
  • a description candidate list 463 a or 465 a is displayed. The user can input a location description by selecting the correct description from the description candidate list 463 a or 465 a.
  • a location for example, “location to go back to” like the home, “location to work” like a company, and “location to learn” like a school can be cited.
  • location to go back to like the home
  • location to work like a company
  • location to learn like a school
  • a description may directly be input into the start location description correction area 463 or the end location description correction area 465 .
  • the end location correction area 464 and the end location description correction area 465 can be corrected.
  • the correction method is as described above.
  • the feeling correction area 466 includes a “no feeling setting” button and a feeling selection button that stepwise selects good or bad of feelings.
  • the feeling selection button can be configured so that, for example, five levels of feeling of “very bad (irritated or depressed)”, “bad (somewhat irritated)”, “not bad (no problem)”, “good (a bit fine)”, and “very good (refreshing, bracing, full)” are selectable.
  • the user can select the feeling when an operation is performed from the feeling correction area 466 .
  • correction content can be reflected in the action segment 412 a by pressing a save button 468 a at the bottom or a save button 468 b at the top of the correction screen 460 .
  • the save button 468 a or 468 b is pressed, a transition to the screen before the transition to the correction screen 460 occurs.
  • a cancel button 467 a at the bottom or a cancel button 467 b at the top of the correction screen 460 may be pressed.
  • the display of the action log 410 can easily be changed not only by correcting content of each of the action segments 412 , but also by correcting relationships between the action segments 412 .
  • a plurality of the action segments 412 may be combined to display the resultant segment as the one action segment 412 .
  • the combination of the action segments 412 is a function to combine the plurality of the action segments 412 that are temporally consecutive into the one action segment 412 .
  • the time range of the combined action segment 412 extends from the start time of the oldest action segment 412 to the end time of the newest action segment 412 .
  • FIG. 21 shows an example of the method of combining the action segments 412 .
  • a merge button 432 is pressed while the action log 410 is displayed, a state (action combination mode) in which the action segments 412 can be combined is entered.
  • the selected action segments 412 can be combined by selecting the action segment 412 to be combined in action combination mode.
  • the action of an action segment 412 b selected first among the action segments 412 to be combined can be set as operation content after the combination.
  • the operation content of “work in Ohsaki” becomes the operation content after the combination.
  • the other action segments 412 to be combined are selected.
  • the selection of the action segment 412 can be made by a touch operation or a drag operation.
  • action segments surrounded by a reference sign 412 c are selected for combination.
  • the save button 468 a is pressed after the action segments 412 to be combined are selected, the action segments 412 to be combined are displayed by being combined.
  • the combined action segment 412 represents the action of “work in Ohsaki” is done between 9:55 and 22:42.
  • the cancel button 467 a may be pressed.
  • FIG. 22 An example of another method of deciding operation content of the action segment 412 after combination is shown in FIG. 22 .
  • an action segment to be combined is selected by a drag operation. That is, the first action segment (start segment) to be combined is first selected by contact with a finger and the finger is moved while in contact until the last action segment (end segment) to be combined is selected. In this manner, the action segments (action segments surrounded by the reference sign 412 c ) to be combined are decided. Then, when the save button 468 a (see FIG. 21 ) is pressed, the action segment after the combination is displayed.
  • the operation content of any action segment to be combined may be set as the operation content of the action segment after the combination. If, for example, the operation content of “walked in Ohsaki” is selected, the operation content of an action segment 412 d 1 after the combination becomes “walked in Ohsaki”.
  • the operation content of the action segment after the combination may be decided, for example, by majority of action segments to be combined.
  • the operation content of the two action segments 412 is “work in Ohsaki”
  • the operation content of the one action segment is “meal in Ohsaki”
  • the operation content of the other one action segment is “walked in Ohsaki”. Therefore, the operation content of “work in Ohsaki” of the most action segments is decided as the operation content of the action segment 412 d 2 after the combination.
  • the operation content of the action segment after the combination may be decided, for example, by reanalyzing action segments to be combined.
  • the operation content of the four action segments 412 to be combined includes “work in Ohsaki”, “meal in Ohsaki”, and “walked in Ohsaki”. Accordingly, the meaning and content of user's actions are reanalyzed and, for example, the operation content of “was in Ohsaki” can be decided as the operation content of an action segment 412 d 3 after the combination.
  • the action segments 412 can easily be combined by selecting the action segments to be combined.
  • a plurality of the action segments 412 can be divided and displayed as a plurality of the action segments 412 as a correction of the relationship between the action segments 412 .
  • the division of the action segment 412 is a function to segment the one action segment 412 into a plurality of the action segments 412 .
  • the division method of the action segment 412 for example, a method of setting the time to divide the action segment 412 and a division method using hierarchical information are known.
  • FIG. 23 shows an example of a division method by time settings of the action segment 412 .
  • a division 433 is pressed while the action log 410 is displayed, a state (action division mode) in which the action segment 412 can be divided is entered.
  • the selected action segment 412 can be divided at a specified time by selecting the action segment 412 to be divided in action division mode.
  • an action segment 412 e is selected for division in FIG. 23 .
  • a division time setting screen to input the division time of the action segment 412 e to be divided is displayed.
  • ant time between the start time and the end time of the action segment to be divided can be set.
  • pressing the save button 468 a divides action segment 412 to be divided and displays the divided action segments 412 . If, for example, the division time is set to 11:50, the action segment 412 is divided into a first action segment of “work in Ohsaki” between 9:58 and 11:50 and a second action segment of “work in Ohsaki” between 11:50 and 12:29. To cancel the division mode, the cancel button 467 a may be pressed.
  • FIG. 24 shows an example of a division method based on hierarchical information of the action segment 412 . It is assumed that the action division mode is already set in FIG. 24 . If the user selects the action segment 412 e to be divided while the action log 410 is displayed, as shown in FIG. 24 , a hierarchical list 414 representing the action segment 412 e to be divided by more detailed action segments is displayed. When the hierarchical list 414 is displayed, the user selects the action segment to be the division point.
  • an action segment 414 a is selected in FIG. 24 , for example, the end time of the action segment 414 a is decided as the division time and, as shown on the right of FIG. 24 , the action segment is divided into a first action segment 412 e 1 of “work in Ohsaki” between 9:30 and 11:59 and a second action segment 412 e 2 of “work in Ohsaki” between 11:59 and 12:30.
  • the action segments 412 constituting the action log 410 hold a hierarchical relationship based on the meaning and content thereof as hierarchical information.
  • the display roughness of the displayed action log 410 can be changed by changing the segmentation grain size using the hierarchical information.
  • the display roughness can be changed by using, for example, a slider or a zoom button.
  • FIG. 25 shows a case when display roughness is changed by using a slider 471 .
  • a knob 472 to set the display roughness (grain size) is displayed in the slider 471 and the display roughness can be changed by changing the position of the knob 472 .
  • the knob 472 is positioned on the side of small grain size of the slider 471 , the action log 410 is displayed by the detailed action segments 412 .
  • a plurality of the action segments 412 are displayed by being combined based on the segmentation grain size attached to the action segments 412 in advance.
  • FIG. 26 shows a case when the display roughness is changed by using a zoom button 473 .
  • a plurality of buttons 474 to set the display roughness (grain size) is displayed in a row in the zoom button 473 .
  • the action log 410 can be made to be displayed in the display roughness corresponding to the button 474 .
  • the button 474 on the side of small grain size in the zoom button 473 is checked, the action log 410 is displayed by the detailed action segments 412 .
  • a plurality of the action segments 412 are displayed by being combined based on the segmentation grain size attached to the action segments 412 in advance.
  • the display roughness of the action log 410 can easily be changed based on the segmentation grain size attached to the action segment 412 so that the user can view the action log 410 in the desired display roughness.
  • the display roughness of the action log 410 is changed in FIGS. 25 and 26 , but in the present technology, the display roughness of the action segment 412 can also be changed depending on the purpose. That is, the display roughness of the action segment 412 is changed independently of the type of action. Work, shopping, movement and the like can be considered as the types of action and, for example, by checking a button 476 of the corresponding action from an action type selection list 475 as shown in FIG. 27 , the display roughness of only the action can be changed.
  • action segments 412 f 1 , 412 g 1 related to work are displayed in detail.
  • the operation content of “work in Ohsaki” for the action segment 412 f 1 is displayed by five action segments 412 f 2 of “desk work in Ohsaki”, “movement on foot”, “meeting in Ohsaki”, “movement on foot”, and “desk work in Ohsaki”.
  • the operation content of the action segment 412 g 1 is the most detailed and thus, the same content is displayed after the display roughness is changed (action segment 412 g 2 ).
  • an action segment 412 h 1 related to shopping is displayed in detail.
  • the operation content of “shopping in Shibuya” for the action segment 412 h 1 is displayed by seven action segments 412 h 2 of “shopping in Shibuya”, “movement on floor”, “shopping in Shibuya”, “movement between shops in Shibuya”, “shopping in Shibuya”, “moving between shops in Shibuya”, and “shopping in Shibuya”.
  • action segments 412 i 1 , 412 j 1 , 412 k 1 related to movement are displayed in detail.
  • the operation content of “movement from Saginomiya to Ohsaki by train” for the action segment 412 i 1 is displayed by five action segments 412 i 2 of “waiting for train in Saginomiya”, “movement from Saginomiya to Takadanobaba by train”, “trains changed in Takadanobaba”, “movement from Takadanobaba to Ohsaki by train”, and “movement in Ohsaki on foot”.
  • the operation content of “moving from Ohsaki to Shibuya by train” for the action segment 412 j 1 is displayed by four action segments 412 j 2 of “movement to Ohsaki station”, “waiting for train in Ohsaki station”, “movement from Ohsaki to Shibuya by train”, and “movement to Shibuya”.
  • the action segment 412 k 1 similarly detailed content is displayed.
  • a button 476 a of “uniform detail” may be checked in the action type selection list 475 . Accordingly, all the action segments 412 of the action log 410 are displayed in detail in the same grain size.
  • FIG. 31 is shows a display grain size setting unit 477 provided with a slider 478 a to set the display roughness of an action segment related to “work”, a slider 478 b to set the display roughness of an action segment related to “shopping”, and a slider 478 c to set the display roughness of an action segment related to “movement”.
  • the display roughness can be adjusted for each type of action.
  • the action segment 412 can be deleted from the action log 410 . If, for example, as shown in FIG. 32 , an action segment 412 l to be deleted is selected and a deletion button 436 is pressed while the action log 410 is displayed, as shown on the right of FIG. 32 , a deletion confirmation screen 480 is displayed. In the deletion confirmation screen 480 , the user can be caused to enter the reason for deleting the action segment 412 l to be deleted. When the user presses a button on which the reason for deletion is written, the action segment 412 l to be deleted is deleted from the action log 410 . Depending on the reason selected by the user, the deletion of the action segment may be fed back as an action correction.
  • content of the action segment 412 of the action log 410 can be posted. If, for example, as shown in FIG. 33 , an action segment 412 m to be posted is selected and a post button 437 is pressed while the action log 410 is displayed, as shown on the right of FIG. 33 , a posting screen 482 is displayed. In the posting screen 482 , the operation content of the action segment 412 m to be posted is automatically pasted to a posting content input area 482 a . When the user presses a posting button 482 b , the description content in the posting content input area 482 a is posted to a posting site.
  • a settings screen 490 is made to display by pressing a settings button 438 .
  • various settings about the action log display application can be made.
  • “stop” of an acquisition function setting unit 491 that sets the operation of the action log acquisition function is selected.
  • the action log display application stops the action log acquisition function.
  • “restart” of the acquisition function setting unit 491 may be selected.
  • the action log display application in the present technology automatically uploads operation action data acquired by the action recording device 100 to the action log server 200 in predetermined timing (for example, twice per day). Also, the analysis server 300 automatically generates an action segment in predetermined timing (for example, twice per day). While an action log is displayed based on generated action segments, an action log displayed in accordance with the system function or circumstances may not correspond to the latest results. Thus, by pressing an update button 493 that updates the action log displayed in the settings screen 490 of FIG. 34 to the latest results, the action log can be updated to the latest results. When a transition to the action log display screen occurs after the update button 493 being pressed, the display unit 150 can be caused to display the latest results.
  • an action log display system in the present technology, the meaning and content of an action is analyzed by the analysis server 300 and an action log is displayed by action segments.
  • content of the displayed action log may not all correct.
  • the user can make corrections to correct content by using the action log display application.
  • correction feedback of the user is reflected in the next analysis processing by the analysis server 300 and used to improve the precision of the next and subsequent analysis results. The reflection processing of correction feedback will be described below based on FIGS. 35 to 42 .
  • the precision of analysis results is improved by reflecting correction feedback of the user in analysis processing, but the user may not correct all errors of analysis results by the analysis server 300 . That is, content of an action log that is not corrected may not necessarily be correct. Thus, in the present technology, it is necessary to assume a system capable of collecting substantially biased information only. In addition, analysis results before corrections by the user do not necessarily match the latest analysis results. Thus, by reflecting information showing which action segment is corrected in what way in analysis processing for each user, the action specific to each user can be learned, which is considered to effectively work to improve the precision of analysis results.
  • an action pattern is decided based on characteristic amount analysis results in recognition processing of an operation action and acquires a plurality of probability distributions corresponding to the action pattern, time, and position information (location).
  • a weight of a histogram is assigned and an operation action is recognized based on results of assigning weights depending on the location. If position information cannot be acquired or there is no need to acquire position information, uniform weights may be assigned or specific weights like “no location can be acquired” or “there is no need for location” may be assigned.
  • FIG. 35 shows an example of action recognition processing by the living action recognition unit 321 .
  • FIG. 36 shows operation action estimation information that decides an operation action.
  • the operation action estimation information is, for example, information showing the relation between a weighting factor depending on the location and the probability distribution of each action and, as shown in FIG. 36 , a plurality (four in FIG. 36 ) of probability distributions of actions of, for example, “shopping”, “work”, “meal”, and “others” is acquired. Then, the weighting factor depending on the location is set to each probability distribution.
  • the operation action estimation information is preset and recorded in, for example, the analysis parameter DB 333 .
  • the living action recognition unit 321 After an action log in the unit time is acquired, the living action recognition unit 321 starts processing to recognize the action of the action log. First, as shown in FIG. 35 , at least one pair of the probability distribution and the weighting factor depending on the location is acquired based on the action pattern, time information, and position information (S 300 ).
  • the living action recognition unit 321 performs processing to decide operation content of the action log in the unit time.
  • the first pair is the pair of the probability distribution and the weighting factor in the first row
  • the probability of 50 is multiplied by the weighting factor of 1 for “shopping” and the probability of 10 is multiplied by the weighting factor of 1 for “work”.
  • the probability of 10 is multiplied by the weighting factor of 1 for “meal” and the probability of 30 is multiplied by the weighting factor of 1 for “others”. Accordingly, the integrated values (“shopping”: 50, “work”: 10, “meal”: 10, “others”: 30) of actions are acquired.
  • the living action recognition unit 321 adds 1 to the parameter i (S 306 ) and repeats the processing in step S 302 and thereafter.
  • the probability of 10 is multiplied by the weighting factor of 6 for “meal” and the probability of 30 is multiplied by the weighting factor of 6 for “others”. Accordingly, the integrated values (“shopping”: 60, “work”: 300, “meal”: 60, “others”: 180) of actions are acquired.
  • the living action recognition unit 321 adds the integrated value in the second row to the integrated value in the first row for each action.
  • integrated values are calculated for the pairs of the probability distribution and the weighting factor in the third and fourth rows and these integrated values are added to the above integrated values of each action to finally obtain added values of “shopping”: 260, “work”: 460, “meal”: 420, “others”: 460.
  • the living action recognition unit 321 decides the action of the maximum final added value as the operation content of the action log.
  • both “work” and “others” have the maximum added value of 460, but in this case, the operation content is decided by giving priority to actions other than “others”. Therefore, in the example of FIG. 36 , “work” is recognized as the operation content.
  • recognition processing results of operation content changes considerably depending on operation action estimation information showing the relation between the weighting factor depending on the location and the probability distribution.
  • the action log is corrected by the user, correction content is reflected in operation action estimation information recorded in the analysis parameter DB 333 by the feedback adjustment unit 332 of the analysis server 300 . Accordingly, the precision of recognition processing of operation content can be enhanced.
  • correction feedback is converted into data in a format that can be transmitted to the analysis server 300 by the input information processing unit 144 before being transmitted to the analysis server 300 via the client interface unit 130 .
  • the analysis server 300 having received the correction feedback from the action recording device 100 through the analysis server interface unit 310 reflects content of the correction feedback in the operation action estimation information through the feedback adjustment unit 332 .
  • the feedback adjustment unit 332 corrects the probability distribution of the operation action estimation information if the content of the correction feedback concerns an action and corrects the weighting factor depending on the location if the content of the correction feedback concerns position information (location).
  • the feedback adjustment unit 332 corrects, among a plurality of probability distributions, the probability distribution having the maximum probability of “work”. For example, the feedback adjustment unit 332 makes a correction to set the probability of “work” as an analysis result and the probability of “shopping” as a correction result to the average value of these two probabilities for the probability distribution in the second row with the maximum probability of “work”.
  • the feedback adjustment unit 332 corrects, among a plurality of probability distributions, the weighting factor of the probability distribution having the maximum probability of “shopping”. For example, the feedback adjustment unit 332 makes a correction of increasing the weighting factor in the first row with the maximum probability of “shopping” by a factor of a predetermined number (for example, 10).
  • correction feedback is reflected in analysis results of action segments so that the precision of analysis results of the operation content can be expected.
  • the reflection processing of correction feedback will be described in more detail below based on FIGS. 38 to 42 .
  • FIG. 38 is a flow chart showing the reflection processing of correction feedback of an action.
  • FIG. 39 is an explanatory view illustrating corrections of the operation action estimation information based on the processing in FIG. 38 .
  • FIG. 40 is a flow chart showing other reflection processing of correction feedback of the action.
  • “001”, “002”, and “003” shown in the top row of each column in FIG. 39 are IDs representing respective actions. It is assumed in the description below that the analysis result by the living action recognition unit 321 is “action 002” and the correct action by user's correction feedback is “action 003”.
  • the feedback adjustment unit 332 When correction feedback is received from the action recording device 100 , the feedback adjustment unit 332 first recognizes correction content. It is assumed here that operation content of an action segment is corrected. The feedback adjustment unit 332 acquires the action segment to be corrected from the unit data storage DB 334 or the hierarchical information attached data storage DB 335 and starts processing shown in FIG. 38 .
  • the feedback adjustment unit 332 first acquires the probability distribution (partial probability distribution) used to recognize the operation content of the action segment to be corrected from operation action estimation information stored in the analysis parameter DB 333 (S 310 ). Next, the feedback adjustment unit 332 calculates a value M(i) obtained by multiplying the maximum probability of each probability distribution by the weighting factor of the row for the partial probability distribution and sorts these probability distributions (S 311 ).
  • step S 312 the feedback adjustment unit 332 acquires a correction ratio C(i) from a loss ratio calculation function using an action having the maximum value of probability distribution of each action of target rows from operation action estimation information and the correct action obtained from correction feedback (S 313 ).
  • the loss ratio calculation function is assumed to be a single comprehensive measure representing losses caused when some available decision is made.
  • the loss ratio calculation function is used to set, for example, a correction ratio table representing a correction ratio C between the action of analysis results and the correct action as shown in the lower portion of FIG. 39 .
  • the correction ratio table can be preset and can be stored in the analysis parameter DB 333 .
  • the operation action estimation information after the correction feedback being reflected as shown on the right of FIG. 39 can be obtained.
  • the reflection processing shown in FIG. 38 is effective in being able to control the reflection speed or divergence. That is, a more impermissible error can be reflected earlier and a value can be made to converge if the value is handled as a ratio to the maximum value.
  • the correction ratio C(i) is added to or subtracted from the value of the probability distribution, but the present technology is not limited to such an example and, for example, correction feedback may be reflected in operation action estimation information by multiplying the value of the probability distribution by the correction ratio.
  • correction feedback In the reflection processing of correction feedback shown in FIG. 38 , content of the correction feedback is reflected in operation action estimation information by using a correction ratio table, but the present technology is not limited to such an example.
  • a feedback system using a neural network technique may be configured. It is assumed that operation content of an action segment is corrected also in FIG. 40 .
  • the feedback adjustment unit 332 first acquires the probability distribution (partial probability distribution) used to recognize the operation content of the action segment to be corrected from operation action estimation information stored in the analysis parameter DB 333 (S 320 ). Next, the feedback adjustment unit 332 calculates a value M(i) obtained by multiplying the maximum probability of each probability distribution by the weighting factor of the row for the partial probability distribution and sorts these probability distributions (S 321 ).
  • the processing in steps S 320 , S 321 can be made the same as the processing in steps S 310 , S 311 in FIG. 38 .
  • the feedback adjustment unit 332 determines whether the parameter i is smaller than n (S 322 ). If the condition in step S 322 is not satisfied, the processing shown in FIG. 40 is terminated. On the other hand, if the condition in step S 322 is satisfied, the feedback adjustment unit 332 uses the neural network technique to reflect correction content of correction feedback in each probability distribution of operation action estimation information based on the weighting factor (S 323 ). Then, the feedback adjustment unit 332 adds 1 to the parameter i (S 324 ) and repeats the processing in step S 322 and thereafter.
  • FIG. 41 is a flow chart showing the reflection processing of correction feedback of the action and position information.
  • the feedback adjustment unit 332 acquires the action segment to be corrected from the unit data storage DB 334 or the hierarchical information attached data storage DB 335 and starts processing shown in FIG. 41 .
  • the feedback adjustment unit 332 first determines whether correction feedback contains position information (S 330 ). If the correction feedback contains position information in step S 330 , processing in step S 331 and thereafter is performed to reflect corrected position information in operation action estimation information.
  • step S 331 whether any correction related to position information is made on action segments accompanied by movement is determined. If a correction related to position information is made on action segments accompanied by movement, representative coordinates of end points (two representative coordinates like position X to position Y) are calculated (S 332 ). On the other hand, if no correction related to position information is made on action segments accompanied by movement, representative coordinates of the movement are calculated (S 333 ). Incidentally, representative coordinates can be calculated by using the center, center of gravity, most frequent point and the like.
  • the feedback adjustment unit 332 records representative coordinates calculated in step S 332 or S 333 , the precision, and attached attributes in a feedback DB (not shown) (S 334 ).
  • the feedback DB is a storage unit provided in the analysis server 300 .
  • the feedback adjustment unit 332 analyses operation content using new position information recorded in the feedback DV in step S 334 and determines whether the analysis result matches the correct action input by the correction feedback (S 335 ). If it is determined in step S 335 that the operation content analyzed by using new position information matches the correct action, a judgment can be made that correction feedback about position information is correctly reflected and also there is no error in the action content. Therefore, the feedback adjustment unit 332 judges that the reflection processing of correction feedback is completed and terminates the processing in FIG. 41 .
  • step S 335 if it is determined in step S 335 that the operation content analyzed by using new position information does not match the correct action, a judgment can be made that with corrections of position information alone, correction feedback is not correctly determined.
  • processing in steps S 336 to S 341 is performed to reflect operation content of the correction feedback in operation action estimation information.
  • the processing in steps S 336 to S 341 can be made the same as the processing in FIG. 38 .
  • the feedback adjustment unit 332 first acquires the probability distribution (partial probability distribution) used to recognize the operation content of the action segment to be corrected from operation action estimation information stored in the analysis parameter DB 333 (S 336 ). Next, the feedback adjustment unit 332 calculates a value M(i) obtained by multiplying the maximum probability of each probability distribution by the weighting factor of the row for the partial probability distribution and sorts these probability distributions (S 337 ).
  • the feedback adjustment unit 332 subtracts the correction ratio C(i) acquired in step S 339 from the value of the probability distribution of the action of the maximum value of probability distribution, adds the correction ratio C(i) to the value of the probability distribution of the correct action, and reflects these corrections in the operation action estimation information (S 340 ). Then, the feedback adjustment unit 332 adds 1 to the parameter i (S 341 ) and repeats the processing in step S 338 and thereafter. By performing the above processing, the operation action estimation information after the correction feedback being reflected can be obtained.
  • the processing shown in FIG. 40 may be performed instead of the processing in steps S 336 to S 341 . Also in this case, the operation action estimation information after the correction feedback being reflected can similarly be obtained.
  • the correction feedback of position information may be reflected by, as shown on the lower left of FIG. 37 , changing the weighting factor of the probability distribution of operation action estimation information.
  • a corrected attribute dependence section, a specified attribute dependence section, and an ID dependence section may be set to reflect the correction feedback of position information in each weighting factor.
  • the specified attribute dependence section is strengthened for the position information of “house” and a peripheral attribute dependence section is strengthened for the position information of “location to do shopping”.
  • the position information of, for example, “company” a plurality of pieces of position information like different offices may be present. In this case, the position information of the same meaning can correctly be selected by giving differences like business content and scale to each piece of position information as respective features.
  • the added amount of weighting factor for position information may be decided based on, for example, original position information or changed for each attribute type of position information. Further, a probability distribution specific to position information may randomly be generated and added to operation action estimation information. Accordingly, over-learning can be prevented.
  • an action log is displayed by using action segments to which the meaning and content is attached.
  • action segments to which the meaning and content is attached.
  • temporal or action errors can be absorbed.
  • a user's typical action pattern can be extracted from a small amount of data.
  • a typical action pattern generation unit 336 is provided in the analysis server 300 .
  • an action log for a predetermined period (for example, for one day) acquired by the living action recognition unit 321 from the action log server 200 is first smoothed and then output to the typical action pattern generation unit 336 .
  • the typical action pattern generation unit 336 generates a typical action pattern using a statistical technique, for example, cross correlation processing on action segments of the smoothed action log.
  • the typical action pattern generation unit 336 acquires a typical action pattern of the user as a result of analyzing, based on action logs of seven days, correlations of action segments of these action logs.
  • a user's probable typical action pattern can be generated even from a small amount of data.
  • the living action recognition unit 321 identifies position information of the user based on which medium of transport the user uses to move or which means of transport the user uses to move ( FIG. 43 ). More specifically, when a movement action is recognized from operation action data, the living action recognition unit 321 analyzes how to display position information thereof. In this case, the living action recognition unit 321 first acquires nearest station candidates as an easy-to-use landmark, the last position information and information related thereto. Nearest station candidates can be identified by using the line name, station name, distance to the station and the like. As the last position information and information related thereto, the medium of transport or means of transport, time difference, distance from the last latitude/longitude and the like are acquired.
  • the living action recognition unit 321 assigns weights to the lines and stations using the above information to identify the nearest station. Weights may be assigned to lines and stations by, for example, increasing weights of nearest station candidates with a decreasing distance or assigning weights preferentially to lines and stations that are continuously acquired in action logs. Alternatively, weights may be assigned in consideration of distance differences or time differences that can be acquired from information up to the last time. Accordingly, if the fact of being a predetermined distance apart or that a predetermined time has passed is recognized from the information up to the last time and information this time, the possibility of having changed trains to another line can be considered.
  • the estimation of line can be determined from, for example, the number of passed stations recognized from an action log.
  • the movement locus of the user can be estimated by considering the possibility of changing trains at a station identified from position information or whether a direct service between a plurality of lines is available. If a plurality of lines runs between the same stations, which line is used can be identified by estimating a more likely line from the user's past movement locus or acquiring more detailed position information from a position information acquisition sensor.
  • the line may be displayed on a map only when the precision of line estimation processing of a predetermined value or more is secured.
  • the station name is selected by, as described above, identifying the nearest station.
  • the expression may be changed in accordance with the position precision of the station name of the nearest station. If, for example, the Ohsaki station is identified as the nearest station, the expression is changed like “work in the Ohsaki station”, “work near the Ohsaki station”, or “work in Ohsaki” based on the distance difference between the position of the nearest station and the position information. Accordingly, the location where work is done can be expressed more appropriately.
  • priority may be given to the identified nearest station as a landmark to represent the location of operation by the station name (excluding “station”).
  • station For example, it is assumed that, as a result of analyzing an action log, movement by car is recognized and the “Higash-Koganei station” and the “Shin-Koganei station” are identified as landmarks. In this case, it is not natural to move between stations by car and thus, action content can naturally be expressed by representing the start location and the goal as the “Higash-Koganei station” and the “Shin-Koganei station”.
  • a process of the action recording device 100 in accordance with this embodiment can be executed either by hardware or software.
  • the action recording device 100 can be configured as shown in FIG. 46 .
  • an exemplary hardware configuration of the action recording device 100 in accordance with this embodiment will be described with reference to FIG. 46 .
  • the action recording device 100 in accordance with this embodiment can be implemented by a processing device such as a computer as described above.
  • the action recording device 100 includes a CPU (Central Processing Unit) 101 , ROM (Read Only Memory) 102 , RAM (Random Access Memory) 103 , and a host bus 104 a .
  • the action recording device 100 also includes a bridge 104 , an external bus 104 b , an interface 105 , an input device 106 , an output device 107 , a storage device (HDD) 108 , a drive 109 , a connection port 111 , and a communication device 113 .
  • a processing device such as a computer as described above.
  • the action recording device 100 includes a CPU (Central Processing Unit) 101 , ROM (Read Only Memory) 102 , RAM (Random Access Memory) 103 , and a host bus 104 a .
  • the action recording device 100 also includes a bridge 104 , an external bus 104 b
  • the CPU 101 functions as an arithmetic processing unit and a control unit, and controls the entire operation within the action recording device 100 in accordance with various programs.
  • the CPU 101 may also be a microprocessor.
  • the ROM 102 stores programs, operation parameters, and the like used by the CPU 101 .
  • the RAM 103 temporarily stores programs used in the execution of the CPU 101 , parameters that change as appropriate during the execution, and the like. These units are mutually connected via the host bus 104 a including a CPU bus or the like.
  • the host bus 104 a is connected to the external bus 104 b such as a PCI (Peripheral Component Interconnect/Interface) bus via the bridge 104 .
  • PCI Peripheral Component Interconnect/Interface
  • the host bus 104 a , the bridge 104 , and the external bus 104 b need not necessarily be arranged separately, and the functions of such components may be integrated into a single bus.
  • the input device 106 includes an input means for a user to input information, such as a mouse, a keyboard, a touch panel, a button, a microphone, a switch, or a lever; an input control circuit that generates an input signal on the basis of a user input and outputs the signal to the CPU 101 ; and the like.
  • the output device 107 includes a display device such as, for example, a liquid crystal display (LCD) device, an OLED (Organic Light Emitting Diode) device, or a lamp; and an audio output device such as a speaker.
  • LCD liquid crystal display
  • OLED Organic Light Emitting Diode
  • the storage device 108 is a device for storing data, constructed as an example of a storage unit of the action recording device 100 .
  • the storage device 108 can include a storage medium, a recording device that records data on the storage medium, a reading device that reads data from the storage medium, a deletion device that deletes data recorded on the storage medium, and the like.
  • the storage device 108 includes, for example, a HDD (Hard Disk Drive).
  • the storage device 108 stores programs and various data for driving the hard disk and executed by the CPU 101 .
  • the drive 109 is a reader/writer for a storage medium, and is incorporated in or externally attached to the action recording device 100 .
  • the drive 109 reads information recorded on a removable storage medium such as a magnetic disk, an optical disc, a magnetooptical disk, or semiconductor memory that is mounted, and outputs the information to the RAM 103 .
  • the connection port 111 is an interface for connection to an external device, and is, for example, a connection port for connection to an external device that can transmit data via a USB (Universal Serial Bus).
  • the communication device 113 is, for example, a communication interface including a communication device and the like for connection to the communication network 10 .
  • the communication device 113 may be any of a communication device supporting a wireless LAN (Local Area Network), a communication device supporting a wireless USB, or a wire communication device that performs wire communication.
  • the action representation generation unit 320 and the data management unit 330 are provided in the analysis server 300 and the action representation generation unit 140 is provided in the action recording device 100 , but the present disclosure is not limited to such an example.
  • these functional units may all be provided in the analysis server 300 or in the action recording device 100 .
  • present technology may also be configured as below.
  • An information processing device including:
  • an action recognition unit that recognizes an operation action of a user based on sensor information
  • an action representation generation unit that analyzes operation action data showing the operation action of the user recognized by the action recognition unit to generate an action segment represented by a meaning and content of the operation action from the operation action data.
  • the action representation generation unit estimates the meaning and content of the operation action from relations before and after the operation action data arranged chronologically based on the dictionary data to generate the action segment.
  • the action representation generation unit estimates the meaning and content of the operation action in accordance with a time period and a time of the operation action data to be analyzed to generate the action segment.
  • the action representation generation unit estimates the meaning and content of the operation action in accordance with positions of the operation action data before and after the operation action data to be analyzed to generate the action segment.
  • the action representation generation unit displays the action segments based on a segmentation grain size deciding roughness of segmentation of the action segments and the hierarchical information.
  • the action representation generation unit combines or divides the action segments based on a size of the segmentation grain size and displays the combined or divided action segments.
  • a typical action pattern generation unit that extracts one action pattern from a plurality of action segment groups including the action segments of a predetermined unit based on a correlation between the action segments.
  • the action representation generation unit displays, in a display unit, the action segments represented at least by a start time, an end time, position information, and operation content of the operation action by chronologically arranging the action segments.
  • a feedback adjustment unit that corrects operation action estimation information that decides the operation action based on correction feedback from the user to the action segment generated by the action representation generation unit
  • the action representation generation unit generates the action segment constituting an action log from the operation action data based on the operation action estimation data
  • the feedback adjustment unit corrects the operation action estimation information based on the correction feedback.
  • An information processing device including:
  • an action recognition unit that recognizes an operation action of a user based on sensor information
  • an action representation generation unit that generates an action segment constituting an action log from operation action data showing the operation action of the user recognized by the action recognition unit based on operation action estimation information that decides the operation action;
  • a feedback adjustment unit that corrects the operation action estimation information based on correction feedback from the user to the action segment generated by the action representation generation unit.
  • the operation action estimation information includes a plurality of combinations of a probability distribution and a weighting factor depending on a location for a plurality of the operation actions
  • the feedback adjustment unit corrects the probability distribution or the weighting factor for each of the operation actions based on the correction feedback.
  • the feedback adjustment unit corrects the probability distribution of the operation action estimation information in accordance with content of the correction feedback.
  • the feedback adjustment unit corrects the weighting factor of the operation action estimation information in accordance with content of the correction feedback.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • Databases & Information Systems (AREA)
  • Human Resources & Organizations (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Strategic Management (AREA)
  • Human Computer Interaction (AREA)
  • Data Mining & Analysis (AREA)
  • Economics (AREA)
  • Tourism & Hospitality (AREA)
  • General Business, Economics & Management (AREA)
  • Quality & Reliability (AREA)
  • Operations Research (AREA)
  • Marketing (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Development Economics (AREA)
  • Educational Administration (AREA)
  • Game Theory and Decision Science (AREA)
  • User Interface Of Digital Computer (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Navigation (AREA)

Abstract

An information processing device according to the present technology includes an action recognition unit that recognizes an operation action of a user based on sensor information, and an action representation generation unit that analyzes operation action data showing the operation action of the user recognized by the action recognition unit to generate an action segment represented by a meaning and content of the operation action from the operation action data.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • The present application is a continuation of U.S. patent application Ser. No. 14/123,886 filed on Dec. 4, 2013, which is a National Stage Filing of PCT Application No. PCT/JP2012/064564 filed on Jun. 6, 2012, which claims benefit to Japanese Patent Application No. 2011-131130 filed on Jun. 13, 2011, the disclosure of which are incorporated herein by reference.
  • TECHNICAL FIELD
  • The present disclosure relates to an information processing device that processes a user's action records, an information processing method, and a computer program.
  • BACKGROUND ART
  • A technology to recognize a user's operation action from sensor information acquired by using various sensing technologies is proposed. The recognized user's operation action is automatically recorded as an action log and can be represented by various techniques, for example, reproducing the operation action by animation such as an avatar, showing a user's movement locus on a map, or using an index abstracting various operation actions for representation.
  • CITATION LIST Patent Literature
  • Patent Literature 1: JP 2008-3655A
  • SUMMARY OF INVENTION Technical Problem
  • However, when an action log is reproduced by animation such as an avatar using an action recording device like, for example, a motion capture, a very large-scale device will be needed. On the other hand, an action log generated by using a small sensor containing recording device such as a smartphone limits the types of action that can be recorded/recognized and thus, it is difficult to present an action record that is valuable to the user. Therefore, an action log is generally shown as a user's movement locus on a map or displayed as an action record converted to the amount of activity like a health index.
  • Therefore, a proposal of the representation technique to present an action log recorded by a small sensor containing recording device to the user in a manner that is easy to understand has been sought.
  • Solution to Problem
  • According to the present disclosure, there is provided an information processing device including an action recognition unit that recognizes an operation action of a user based on sensor information, and an action representation generation unit that analyzes operation action data showing the operation action of the user recognized by the action recognition unit to generate an action segment represented by a meaning and content of the operation action from the operation action data.
  • According to the present disclosure, there is provided an information processing device including an action recognition unit that recognizes an operation action of a user based on sensor information, an action representation generation unit that generates an action segment constituting an action log from operation action data showing the operation action of the user recognized by the action recognition unit based on operation action estimation information that decides the operation action, and a feedback adjustment unit that corrects the operation action estimation information based on correction feedback from the user to the action segment generated by the action representation generation unit.
  • According to the present disclosure, there is provided an information processing method including a step for recognizing an operation action of a user based on sensor information, and a step for analyzing operation action data showing the recognized operation action of the user to generate an action segment represented by a meaning and content of the operation action from the operation action data.
  • According to the present disclosure, there is provided an information processing method including a step for recognizing an operation action of a user based on sensor information, a step for generating an action segment constituting an action log from operation action data showing the recognized operation action of the user based on operation action estimation information that decides the operation action, and a step for correcting the operation action estimation information based on correction feedback from the user to the action segment.
  • According to the present disclosure, there is provided a computer program for causing a computer to function as an information processing device including an action recognition unit that recognizes an operation action of a user based on sensor information, and an action representation generation unit that analyzes operation action data showing the operation action of the user recognized by the action recognition unit to generate an action segment represented by a meaning and content of the operation action from the operation action data.
  • According to the present disclosure, there is provided a computer program for causing a computer to function as an information processing device including an action recognition unit that recognizes an operation action of a user based on sensor information, an action representation generation unit that generates an action segment constituting an action log from operation action data showing the operation action of the user recognized by the action recognition unit based on operation action estimation information that decides the operation action, and a feedback adjustment unit that corrects the operation action estimation information based on correction feedback from the user to the action segment generated by the action representation generation unit.
  • According to the present disclosure, operation action data showing a user's operation action recognized by an action recognition unit based on sensor information is analyzed by an action representation generation unit to generate an action segment represented by the meaning and content of the operation action from the operation action data. By displaying an action log with the action segment represented by the meaning and content of the operation action, information can be presented to the user in a manner that is easy to understand.
  • Advantageous Effects of Invention
  • According to the present disclosure, as described above, a recorded action log can be presented to the user in a manner that is easy to understand.
  • BRIEF DESCRIPTION OF DRAWINGS
  • FIG. 1 is an explanatory view showing the configuration of an action log display system according to an embodiment of the present disclosure.
  • FIG. 2 is a functional block diagram showing a functional configuration of the action log display system according to the embodiment.
  • FIG. 3 is an explanatory view showing an example of a context level dictionary.
  • FIG. 4 is an explanatory view showing a case when an action segment is generated from operation action data by contextual analysis and the display of the action log is changed by changing a segmentation grain size of the action segment.
  • FIG. 5 is an explanatory view showing another example of the case when the action segment is generated from operation action data by the contextual analysis and the display of the action log is changed by changing the segmentation grain size of the action segment.
  • FIG. 6 is an explanatory view showing a case when an action segment is generated from operation action data by combining the contextual analysis and a time width and the display of the action log is changed by changing the segmentation grain size of the action segment.
  • FIG. 7 is an explanatory view showing a case when an action segment is generated from operation action data by combining the contextual analysis, the time width, and position changes and the display of the action log is changed by changing the segmentation grain size of the action segment.
  • FIG. 8 is a flow chart showing the flow of overall processing of action recognition.
  • FIG. 9 is a flow chart showing processing by a living action recognition unit.
  • FIG. 10 is a flow chart showing processing by a hierarchical structure judgment unit.
  • FIG. 11 is a flow chart showing action segment generation processing.
  • FIG. 12 is a flow chart showing the action segment generation processing.
  • FIG. 13 is an explanatory view illustrating a method of attaching hierarchical information to the action segment.
  • FIG. 14 is functional block diagram showing the functional configuration of an analysis server.
  • FIG. 15 is an explanatory view showing an example of the representation of the action log.
  • FIG. 16 is an explanatory view showing a display example of the action log when an action log display application is activated.
  • FIG. 17 is an explanatory view showing a display example of a calendar.
  • FIG. 18 is an explanatory view showing a display example when a map is caused to display position information corresponding to the action log by operating a map button.
  • FIG. 19 is an explanatory view showing a state in which a correction screen to correct the action segment to be corrected is displayed.
  • FIG. 20 is an explanatory view showing an example of the correction screen to correct an operation action.
  • FIG. 21 is an explanatory view showing an example of a method of combining action segments.
  • FIG. 22 is an explanatory view showing an example of another method of deciding operation content of the action segment after combination.
  • FIG. 23 is an explanatory view showing an example of a division method by time settings of the action segment.
  • FIG. 24 is an explanatory view showing an example of the division method based on hierarchical information of the action segment.
  • FIG. 25 is an explanatory view showing a case when display roughness is changed by using a slider.
  • FIG. 26 is an explanatory view showing a case when the display roughness is changed by using a zoom button.
  • FIG. 27 is an explanatory view showing a display change of the action segment when a button of “work details” is checked in an action type selection list.
  • FIG. 28 is an explanatory view showing a display change of the action segment when a button of “shopping details” is checked in the action type selection list.
  • FIG. 29 is an explanatory view showing a display change of the action segment when a button of “movement details” is checked in the action type selection list.
  • FIG. 30 is an explanatory view showing a display change of the action segment when a button of “uniform details” is checked in the action type selection list.
  • FIG. 31 is an explanatory view showing a positional configuration example of a display grain size setting unit provided with a slider that sets the display roughness of the action segment for each type of action.
  • FIG. 32 is an explanatory view showing a method of deleting an action segment from the action log.
  • FIG. 33 is an explanatory view showing a method of posting content of the action segment of the action log to a posting site.
  • FIG. 34 is an explanatory view showing a positional configuration example of a setting screen to make various settings about the action log display application.
  • FIG. 35 is a flow chart showing an example of action recognition processing by the living action recognition unit.
  • FIG. 36 is an explanatory view showing operation action estimation information showing a relationship between a weighting factor depending on the location and a probability distribution of each action.
  • FIG. 37 is an explanatory view providing an overview of reflection processing of correction feedback.
  • FIG. 38 is a flow chart showing the reflection processing of correction feedback of an action.
  • FIG. 39 is an explanatory view illustrating corrections of the operation action estimation information based on the processing in FIG. 38.
  • FIG. 40 is a flow chart showing other reflection processing of correction feedback of the action.
  • FIG. 41 is a flow chart showing the reflection processing of correction feedback of the action and position information.
  • FIG. 42 is an explanatory view illustrating personal modeling of an action pattern by a typical action pattern generation unit.
  • FIG. 43 is an explanatory view illustrating a position display technique by determining a medium/means of transport.
  • FIG. 44 is an explanatory view illustrating line estimation processing.
  • FIG. 45 is an explanatory view illustrating station name selection processing.
  • FIG. 46 is a block diagram showing a hardware configuration example of an action recording device according to the embodiment.
  • DESCRIPTION OF EMBODIMENT
  • Hereinafter, preferred embodiments of the present disclosure will be described in detail with reference to the appended drawings. Note that, in this specification and the drawings, elements that have substantially the same function and structure are denoted with the same reference signs, and repeated explanation is omitted.
  • The description will be provided in the order shown below:
      • <1. Overview of Action Log Display System>
      • <2. Functional Configuration of Action Log Display System>
      • [2-1. Action Recording Device]
      • [2-2. Action Log Server]
      • [2-3. Analysis Server]
      • <3. Action Segment Generation Processing>
      • [3-1. Relationship between Operation Action and Meaning/Content thereof]
      • [3-2. Action Segment Generation Processing]
      • (Example 1: Generation of an action segment by the contextual analysis)
      • (Example 2: Generation of an action segment by combining the contextual analysis and the time width)
      • (Example 3: Generation of an action segment by combining the contextual analysis, the time width, and position changes)
      • [3-3. Action Recognition Processing]
      • [3-4. Processing Content of Action Representation Generation Unit]
      • <4. Action Recording and Display Application>
      • [4-1. Representation of Action Log Based on Action Segment]
      • [4-2. Browsing Action]
      • [4-3. Correcting Action]
      • [4-4. Combining Actions]
      • [4-5. Dividing Action]
      • [4-6. Representation of Action Segment Based on Segmentation Grain Size]
      • [4-7. Deleting Action]
      • [4-8. Posting Action]
      • [4-9. Action Log Acquisition Stop Processing]
      • [4-10. Updating Display Content]
      • <5. Reflection Processing of Correction Feedback>
      • [5-1. Properties of Correction Feedback]
      • [5-2. Action Recognition Processing]
      • [5-3. Reflection Processing of Correction Feedback]
      • (5-3-1. Overview of reflection processing of correction feedback)
      • (5-3-2. Reflection processing of correction feedback of an action)
      • (5-3-3. Reflection processing of correction feedback of an action and position information)
      • <6. Others>
      • [6-1. Personal Modeling of Action Pattern]
      • [6-2 Position Display Technique by Moving Medium/Means Determination]
      • (6-2-1. Line estimation processing)
      • (6-2-2. Station name selection processing)
      • <7. Exemplary Hardware Configuration>
    1. Overview of Action Log Display System
  • First, an overview of an action log display system according to an embodiment of the present disclosure will be provided with reference to FIG. 1. FIG. 1 is an explanatory view showing an outline configuration of the action log display system according to the present embodiment.
  • The action log display system according to the present embodiment realizes a representation technique that presents an action log recorded by a recording device 100 containing a small sensor (hereinafter, referred to as an “action recording device”) to the user in a manner that is easy to understand. As the action recording device 100, for example, a mobile terminal such as a mobile phone, PDA (Personal Digital Assistant), and smartphone can be used. The action recording device 100 is provided with at least one sensor to sense conditions or an action of a user holding the device. The action recording device 100 estimates an operation action of the user based on sensor information acquired by the sensor and transmits the operation action to an action log server 200 as an action log. In this manner, the action log of the user is accumulated in the action log server 200.
  • An action log analyzed by the action recording device 100 and stored in the action log server 200 records an operation like, for example, “meal”, “movement”, and “sleep” together with the action time, position information and the like. An action log display system according to the present embodiment further analyzes an action log representing the operation content by an analysis server 300 to recognize the meaning of action and generates information (action segment) to which the meaning of action is added. The action segment is unit information as an easy-to-understand representation for the user of an action log. Instead of simply presenting an action log to the user, the action segment can present an action log in a manner that conveys the meaning of action.
  • An action log analyzed by the analysis server 300 and presented to the user can be corrected by the user. In addition, data of the presented action log can be combined, divided, or deleted by generating an action segment. The presented action log can also be posted to a posting site. Thus, by using an action log display system according to the present embodiment, an action log acquired as an operation can be analyzed and presented to the user in an easy-to-understand manner. The configuration and function of an action log display system according to the present embodiment will be described in detail below.
  • 2. Functional Configuration of Action Log Display System
  • FIG. 2 shows the functional configuration of the action log display system according to the present embodiment. The action log display system includes, as described above, the action recording device 100 that records a user's operation action, the action log server 200 that manages an action log recorded by the action recording device 100, and the analysis server 300 that analyzes an action log to generate an action segment.
  • 2-1. Action Recording Device
  • The action recording device 100 includes sensors 110, an action recognition unit 120, a client interface unit 130, an action representation processing unit 140, a display unit 150, and an input unit 160.
  • The sensors 110 are devices that sense a user's action or conditions and are installed in the action recording device 100. As the sensors 110, for example, an acceleration sensor, gyro sensor, magnetic field sensor, atmospheric pressure sensor, illuminance sensor, temperature sensor, microphone and the like can be used. As the sensors 110, a latitude/longitude acquisition sensor that acquires the latitude/longitude can also be installed. As the latitude/longitude acquisition sensor, for example, not only GPS (Global Positioning System) or WiFi, but also base station information of other communication networks or information such as RFID and images may be used. The sensors 110 output detected information to the action recognition unit 120 as sensor information.
  • The action recognition unit 120 estimates a user's action based on sensor information. The action recognition unit 120 includes a sensor controller 122 and an operation action recognition unit 124. The sensor controller 122 controls the sensor 110, the CPU or the overall system to cause sensing by the sensor 110 to operate effectively. The sensor controller 122 controls the above devices based on recognition results by the sensor 110 or the operation action recognition unit 124.
  • The operation action recognition unit 124 recognizes a user's action or conditions by performing signal processing or statistical processing of sensor information. The action recording device 100 holds a correspondence between an action model as information about a user's action obtained as a result of processing sensor information and an operation action in advance. When action parameters are obtained by processing sensor information, the operation action recognition unit 124 identifies an operation action corresponding to the parameters. Then, the operation action recognition unit 124 associates the identified operation action and the action time period, action time, position information and the like and outputs the associated information to the client interface unit 130 as operation action data. The operation action data is uploaded from the client interface unit 130 to the action log server 200.
  • The client interface unit 130 transmits/receives information between the action recording device 100, and the action log server 200 and the analysis server 300. For example, the client interface unit 130 transmits operation action data input from the action recognition unit 120 to the action log server 200 or outputs an analysis result received from the analysis server 300 to the action representation processing unit 140. Also, the client interface unit 130 transmits feedback information from the user input through the input unit 160 to the analysis server 300.
  • The action representation processing unit 140 is a functional unit that displays an action log or processes feedback information from the user and includes a display processing unit 142 and an input information processing unit 144. The display processing unit 142 performs processing to display an analysis result by the analysis server 300 input from the client interface unit 130 in the display unit 150. The input information processing unit 144 performs processing to transmit feedback information from the user for an action log input from the input unit 160 to the analysis server 300 via the client interface unit 130.
  • The display unit 150 is an output device that displays information and can be configured by, for example, a liquid crystal display, organic EL display or the like. For example, an action log processed for display by the display processing unit 142 is displayed in the display unit 150.
  • The input unit 160 is an input device to input information and, for example, a touch panel, keyboard, hardware button or the like can be used. In the present embodiment, it is assumed that the display surface of the display unit 150 is provided with a touch panel as the input unit 160. In this case, the user can input information by, for example, bringing an operation body such as a finger or touch pen into contact with the display surface of the display unit 150 or moving the operation body brought into contact with the display surface. Information input from the input unit 160 is output to the input information processing unit 144.
  • 2-2. Action Log Server
  • The action log server 200 includes a log server interface unit 210 and an action log DB 220.
  • The log server interface unit 210 transmits/receives information between the action log server 200, and the action recording device 100 and the analysis server 300. For example, the log server interface unit 210 records operation action data received from the action recording device 100 in the action log DB 220 or acquires operation action data in accordance with a transmission request from the analysis server 300 from the action log DB 220 and transmits the operation action data to the analysis server 300.
  • The action log DB 220 is a storage unit that stores operation action data of the user acquired by the action recording device 100. In operation action data stored in the action log DB 220, as described above, the operation action identified by the operation action recognition unit 124 and the action time period, action time, position information and the like are associated and stored in the action log DB 220 in, for example, chronological order.
  • 2-3. Analysis Server
  • The analysis server 300 includes an analysis server interface unit 310, an action representation generation unit 320, and a data management unit 330.
  • The analysis server interface unit 310 transmits/receives information between the analysis server 300, and the action recording device 100 and the action log server 200. For example, the analysis server interface unit 310 receives an analysis instruction (analysis request) of an action log from the action recording device 100 or transmits a transmission request of necessary operation action data in accordance with an analysis request. The analysis server interface unit 310 also receives feedback information from the user of an action log from the action recording device 100.
  • The action representation generation unit 320 analyzes operation action data to understand the meaning thereof and generates an action segment to which the meaning and content is added. The action representation generation unit 320 includes a living action recognition unit 321 and a hierarchical structure judgment unit 322. The living action recognition unit 321 generates an action segment from an action log including operation action data. The living action recognition unit 321 analyses the meaning and content of operation action data arranged in chronological order based on relationships between data and the time period, time and the like of data. Then, the living action recognition unit 321 selects data classified as the most detailed meaning and content of analyzed meaning and content as the action segment. The generated action segment is output to the data management unit 330 and held there.
  • The hierarchical structure judgment unit 322 judges a hierarchical structure about the meaning and content of an action segment generated by the living action recognition unit 321 and attaches hierarchical information representing a hierarchical relationship of the meaning and content to the action segment. Hierarchical information is hierarchical meaning information attached to an action segment by processing described later. Hierarchical information may be, for example, information using a normalized value as a key or information using ID identifying the level of meaning information as a direct key. An action segment to which hierarchical information is attached is also called a hierarchical information attached action segment. Hierarchical relationships of the meaning and content of action are stored in the data management unit 330. The hierarchical structure judgment unit 322 outputs a hierarchical information attached action segment to the data management unit 330 via the living action recognition unit 321. The function of the action representation generation unit 320 and details of processing content thereby will be described later.
  • The data management unit 330 manages an action segment generated by the action representation generation unit 320. The data management unit 330 includes a data acquisition unit 331, a feedback adjustment unit 332, an analysis parameter DB 333, a unit data storage DB 334, and a hierarchical information attached data storage DB 335.
  • The data acquisition unit 331 transmits/receives data to/from the action representation generation unit 320. The data acquisition unit 331 records an action segment transmitted from the action representation generation unit 320 in the unit data storage DB 334 or records a hierarchical information attached action segment in the hierarchical information attached data storage DB 335. The data acquisition unit 331 acquires the specified action segment in accordance with a request from the action representation generation unit 320 from the unit data storage DB 334 or the hierarchical information attached data storage DB 335 and outputs the action segment to the action representation generation unit 320.
  • The feedback adjustment unit 332 reflects feedback information received from the action recording device 100 in analysis parameters used for analyzing the meaning and content of operation action data. The feedback information represents content of processing such as corrections made by the user on an action log displayed in the display unit 150 of the action recording device 100. The feedback adjustment unit 332 corrects analysis parameters using feedback information so that the meaning and content of a user's action can be recognized more correctly.
  • The analysis parameter DB 333 is a storage unit that holds analysis parameters used for analyzing the meaning and content of operation action data. In the analysis parameter DB 333, for example, a correspondence between an operation action and the meaning and content is stored as analysis parameters. Information stored in the analysis parameter DB 333 can be referenced by both of the living action recognition unit 321 and the hierarchical structure judgment unit 322. Analysis parameters are updated when necessary based on feedback information from the user.
  • The unit data storage DB 334 stores an action segment generated by the action representation generation unit 320. The action segment stored in the unit data storage DB 334 is a segment (unit segment) of the minimum unit necessary for recognition.
  • The hierarchical information attached data storage DB 335 stores an action segment in which hierarchical information is attached to an action segment generated by the action representation generation unit 320. The action segment stored in the hierarchical information attached data storage DB 335 is a hierarchical information attached action segment to which hierarchical information representing a hierarchical structure of the meaning and content of action is attached by the hierarchical structure judgment unit 322. The recording timing of an action segment to which hierarchical information is attached may be, for example, when requested by an application or analysis results of a plurality of segmentation grain sizes may be recorded in advance by the action representation generation unit 320.
  • That is, the action representation generation unit 320 and the data management unit 330 function as information processing devices that analyze the meaning and content of operation action data generated by the action recording device 100 to present information that is easy for the user to understand.
  • 3. Action Segment Generation Processing
  • In an action log display system according to the present embodiment, the meaning and content of an operation action generated by the action recording device 100 is analyzed by the analysis server 300 to generate an action segment based on the meaning and content of action. Hierarchical information about the meaning and content of action can also be attached to an action segment and the display form of an action log can also be changed easily based on the hierarchical information. First, generation processing of an action segment will be described based on FIGS. 3 to 13.
  • 3-1. Relationship Between Operation Action and Meaning/Content Thereof
  • The action recording device 100 analyzes an operation action, for example, “meal”, “movement”, or “sleep”. The analysis server 300 analyzes content of each operation action more deeply using operation action data containing the operation action. The analysis of the meaning and content of the operation action is conducted by using, for example, as shown in FIG. 3, a context level dictionary.
  • If, among the operation actions of “meal”, “movement”, and “sleep”, the operation action of “movement” is taken up, As shown in FIG. 3, the meaning and content thereof changes depending on what kind of movement. For example, when “moving on foot”, the user can take action of continuing to “walk” or then change to an action of “halting”. When, for example, “waiting for means of transport” on the move, action of continuing to “wait for means of transport” can be taken. Further, when, for example, “moving by train”, the movement can be made an action of “movement by train on a single line”. Further, a state in which a state of “movement by rain” continues or a state of “train stopped” in which the train on which the user moves stops can be assumed. Alternatively, action of “trains changed” after “movement by train” can be taken or a state transition to “waiting for means of transport” can also take place.
  • Then, a further action of “walking” can be associated with a “walking” action or “changing trains” action and a further action of “stopping” can be associated with a “halting” action, a “waiting for means of transport” state, or a “train stop” state. A “train” as a means of transport can further be associated with a state of “moving by train”.
  • Thus, action meta information at an operation action level could change to, as shown in FIG. 3, a higher level of action meta information depending on the context. An action log display system in the present embodiment can analyze the meaning and content of an action based on the relationship between an operation action and the meaning and content and so can present an action log that is easy for the user to understand. In addition, by attaching the hierarchical relationship as hierarchical information to an action segment regarding the meaning and content of the operation action, the segmentation grain size of an action log described later can easily be changed.
  • In the present embodiment, the ontology/semantic technology is applied to the recognition of the meaning and content of an operation action to segment the action recognition that judges “context” in an “action” and operation action data. Ontology systematically represents the concept of relations between words and in the present embodiment, for example, as shown in FIG. 3, the concept of relations between actions is systematized. Then, using the systematized concept of relations between actions, the meaning of an action or the relationship of actions is understood by applying the semantic technology and recognition processing of the meaning and content of the operation action is performed based on the understanding. For example, by judging the context using properties that a transition from some action (for example, a meal) to some action (for example, work) is likely to occur accompanying, for example, an operation action of “movement on foot”, an action segment fitting to the feeling of the user can be generated.
  • 3-2. Action Segment Generation Processing Example 1 Generation of an Action Segment by the Contextual Analysis
  • As a concrete example of generation processing of an action segment using the relationship between the operation action and the meaning and content, generation processing of an action segment by the contextual analysis will be described based on FIGS. 4 and 5. FIG. 4 is an explanatory view showing a case when an action segment is generated from operation action data by the contextual analysis and the display of the action log is changed by changing a segmentation grain size of the action segment. FIG. 5 is an explanatory view showing another example of the case when the action segment is generated from operation action data by the contextual analysis and the display of the action log is changed by changing the segmentation grain size of the action segment.
  • As shown in FIG. 4, it is assumed that an action log including operation action data is acquired by the operation action recognition unit 124. The operation action data is arranged from left to right in chronological order. The operation action data is an operation action recognized based on sensor information of the sensors 110 and an action log is represented by a simple operation action like “walked”, “stopped”, and “got on a train”. The living action recognition unit 321 recognizes the meaning of each piece of operation action data or the relationship between operation action data using a dictionary as shown in FIG. 3 from such an action log.
  • For example, a state of short “stopped” of a predetermined time or shorter between operation action data of “got on a train” is estimated to be a “train stopped (at a station)” state. A state of short “walked” of a predetermined time or shorter between operation action data of “got on a train” is estimated to be a “changing trains” action. Further, “stopped” of operation action data immediately before operation action data of “got on a train” is estimated to be a state of “waited for a train”.
  • By using the action time of operation action data, the action can be estimated more appropriately. Regarding an action of “movement by train”, for example, the meaning of action of “going to office” or “going to school” can be estimated if the action time is a morning hour (for example, from 6 am to 10 am) and the meaning of action of “going home” can be estimated if the action time is an evening hour (for example, from 5 pm to 8 pm). Similarly, regarding an action of “meal”, the meaning of action of “breakfast” can be estimated if the action time is a morning hour, “lunch” can be estimated if the action time is an hour around noon, “supper” can be estimated if the action time is an evening hour.
  • Thus, an action log including action segments as shown in FIG. 4 is generated by an action log including operation action data being analyzed by the living action recognition unit 321. An action segment is data representing an operation action to which operation content is added and is a chunk of consecutive operation actions having the same meaning and content. An action segment generated to match operation action data includes unit segments to which detailed operation content is added. Therefore, if an action log is represented by action segments, as shown in FIG. 4, the moving state while riding on a train can also be known.
  • Then, by acquiring hierarchical action meta information at an operation action level from the context level dictionary shown in FIG. 3 to hierarchically change the segmentation grain size as a parameter to determine the roughness of segmentation of the action segment, the display of the action segment can be changed. With an increasing segmentation grain size, a plurality of action segments that can be considered to be one action is combined to produce an action segment representing rough operation content. On the other hand, with a decreasing segmentation grain size, an action segment approaches the unit segment.
  • For example, as shown in FIG. 4, the segmentation grain size of an action log represented by unit segments is increased. In a state of a grain size 1-1 reached by increasing the segmentation grain size from the segmentation grain size of the unit segment, a series of operations of “got on a train”, “train stopped”, and “got on a train” are represented by one action segment of “got on a train on some line”. In a state of a grain size 1-2 reached by further increasing the segmentation grain size, a series of operations of “got on a train on some line”, “trains changed”, and “got on a train on some line” are represented by one action segment of “movement by train”. In a state of a grain size 1-3 reached by further increasing the segmentation grain size, a series of operations of “walked”, “waited for a train”, “movement by train” and “walked” are represented by one action segment of “moved”.
  • By changing the segmentation grain size based on hierarchical action meta information at the operation action level in this manner, the action log can be displayed at an operation action level that is easy for the user to view.
  • An action segment concerning movement is described with reference to FIG. 4, but an action log can also be displayed similarly for other actions. Assume that, for example, as shown in FIG. 5, “action model X” representing a conspicuously unsteady motion appearing in a shopping action or the like and an operation action of “had a meal” are recognized by the operation action recognition unit 124 as operation action data. “Action model X” normally means the class of shopping, but the meaning thereof changes in accordance with a prior or subsequent determination result.
  • In the example shown in FIG. 5, “action model X” is present before and after “had a meal”. In this case, the living action recognition unit 321 recognizes the action model X as “accepted” and “paid the bill” as operations performed before and after an operation of “have a meal” from operation action data before or after the action model X. Thus, an action log that is easy for the user to understand can be presented by the meaning and content being added to the user's unsteady motion from the prior or subsequent operation. Also in this case, a series of actions of “accepted”, “had a meal”, and “paid the bill” can be represented as one action segment of “had a meal” by increasing the segmentation grain size.
  • Example 2 Generation of an Action Segment by Combining the Contextual Analysis and the Time Width
  • Next, a concrete example when an action segment is generated by considering, in addition to the contextual analysis, the time width will be described based on FIG. 6. FIG. 6 is an explanatory view showing a case when an action segment is generated from operation action data by combining the contextual analysis and the time width and the display of the action log is changed by changing the segmentation grain size of the action segment.
  • As shown in FIG. 6, it is assumed that an action log including operation action data is acquired by the operation action recognition unit 124. In the present example, an action log is represented by simple operation actions like “desk work”, “walked”, “advance arrangements”, and “meal” as operation action data. The living action recognition unit 321 recognizes the meaning of each piece of operation action data or the relationship between operation action data using a dictionary as shown in FIG. 3 and also recognizes the meaning and content thereof by considering the time width of the operation action data.
  • FIG. 6 shows an example of the action log in a company. What kind of walking the operation action data of “walked” is about can be recognized from prior or subsequent operation action data, but can also be estimated based on a walking time. An operation action of short “walked” in a company normally does not have any special meaning. However, if the walking continues for a predetermined time or longer, the walking is estimated not to be simple movement on the floor, but to be movement between premises. Thus, action segments generated by combining the contextual analysis using a dictionary and the time width of operation action data are as shown in FIG. 6. The operation action data of “walked” is divided into “short walking” and “movement between premises” depending on the time width thereof.
  • When action segments are generated, like in the above case, the display of the action log including the action segments can easily be changed by changing the segmentation grain size. For example, in a state of a grain size 2-1 reached by increasing the segmentation grain size from the segmentation grain size of the unit segment, a series of operations of “desk work”, “short walking”, “advance arrangements”, “short walking”, and “desk work” are represented as one action segment of “worked”. In this case, “short walking” is combined into one action segment of “worked” and thus, the action segment of “movement between premises” may be displayed simply as “movement”.
  • In a state of a grain size 2-2 reached by further increasing the segmentation grain size, a series of operations of “worked”, “meal”, “worked”, “movement”, and “worked” are represented as one action segment of “was in company”. By changing the segmentation grain size based on hierarchical action meta information at the operation action level in this manner, the action log can be displayed at an operation action level that is easy for the user to view.
  • Example 3 Generation of an Action Segment by Combining the Contextual Analysis, the Time Width, and Position Changes
  • Next, a concrete example when an action segment is generated by considering, in addition to the contextual analysis and time width, position changes will be described based on FIG. 7. FIG. 7 is an explanatory view showing a case when an action segment is generated from operation action data by combining the contextual analysis, the time width, and position changes and the display of the action log is changed by changing the segmentation grain size of the action segment.
  • As shown in FIG. 7, it is assumed that an action log including operation action data is acquired by the operation action recognition unit 124. In the present example, an action log is represented by simple operation actions like “did shopping”, and “walked” as operation action data. The living action recognition unit 321 recognizes the meaning of each piece of operation action data or the relationship between operation action data using a dictionary as shown in FIG. 3 and also recognizes the meaning and content thereof by considering the time width of the operation action data and position changes of the action recording device 100 (that is, the user).
  • FIG. 7 shows an example of the action log in shopping. What kind of walking the operation action data of “walked” is about can be recognized from prior or subsequent operation action data, but can also be estimated in detail based on a walking time and position changes.
  • If, for example, the operation actions before and after the operation action data of “walked” are “did shopping” and a movement time t is t1 (for example, 35 s) or more and position changes of the action recording device 100 carried by the user are measured, the user is estimated to “move between shops”. Also, if, for example, the operation actions before and after the operation action data of “walked” are “did shopping” and the movement time t is t2 (for example, 20 s) or more and shorter than t1, and no position change of the action recording device 100 is measured, the user is estimated to “move between floors” during shopping. Further, if, for example, the operation actions before and after the operation action data of “walked” are “did shopping” and the movement time t is t3 (for example, 5 s) or more and shorter than t2, and no position change of the action recording device 100 is measured, the user is estimated to “move in a shop” during shopping.
  • Thus, if the meaning and content of operation action data is recognized by combining the contextual analysis, time width, and position changes, as shown in FIG. 7, action segments including three action segments of “movement in a shop (SG1)”, “movement on floors (SG2)”, and “movement between shops (SG3)” are generated from operation action data of “walked”.
  • When action segments are generated, like in the above case, the display of the action log including the action segments can easily be changed by changing the segmentation grain size. For example, in a state of a grain size 3-1 reached by increasing the segmentation grain size from the segmentation grain size of the unit segment, among action segments concerning walking, the action segment SG1 with the shortest walking time is combined with the action segments of “did shopping” preceding and succeeding the action segment SG1. These action segments are represented by an action segment as a series of operations of “did shopping”. In this case, “movement in a shop” is combined into one action segment of “did shopping”, other action segments concerning walking may be displayed simply as “movement”.
  • In a state of a grain size 3-2 reached by further increasing the segmentation grain size, among action segments concerning walking, the action segment SG2 with the shortest walking time next to the action segment SG1 is combined with the action segments of “did shopping” preceding and succeeding the action segment SG2. Then, in a state of a grain size 3-3 reached by further increasing the segmentation grain size, among action segments concerning walking, the action segment SG3 with the longest walking time is also combined with the action segments of “did shopping” preceding and succeeding the action segment SG3. Accordingly, a series of operations of “did shopping” and “walked” are represented as one action segment. By changing the segmentation grain size based on hierarchical action meta information at the operation action level in this manner, the action log can be displayed at an operation action level that is easy for the user to view.
  • 3-3. Action Recognition Processing
  • Processing to generate an action segment from operation action data will be described in detail based on FIGS. 8 to 10. FIG. 8 is a flow chart showing the flow of overall processing of action recognition. FIG. 9 is a flow chart showing processing by the living action recognition unit 321. FIG. 10 is a flow chart showing processing by the hierarchical structure judgment unit 322.
  • Action recognition processing includes, as shown in FIG. 8, operation action data creation processing (S100, S110) performed by the action recognition unit 120 of the action recording device 100 and action segment generation processing (S120 to S140) performed by the action representation generation unit 320 and the data management unit 330 of the analysis server 300.
  • The operation action recognition unit 124 of the action recognition unit 120 having acquired sensor information from the sensors 110 starts creation processing of operation action data (S100). The operation action data creation processing can be performed by using an existing technique. After creating operation action data, the operation action recognition unit 124 outputs the operation action data to the action log server 200 (S110). In this manner, operation action data constituting an action log of the user is accumulated in the action log server 200. Incidentally, the action recognition unit 120 may generate, as operation action data, not only action information at the operation level, but also information including, for example, time information, location information, an operation history of devices and the like.
  • When operation action data is created, the action representation generation unit 320 of the analysis server 300 analyzes the meaning and content of the operation action data through the living action recognition unit 321 (S120). The living action recognition unit 321 segments the operation action data into data of a preset unit length and attaches living action meta information to each piece of segmented data. The unit length of the operation action data is defined by a predetermined time T (for example, T=1 min) The segmentation order of the operation action data is set as i (i=1 to N).
  • After segmenting the operation action data into the unit time T in chronological order, the living action recognition unit 321 first determines whether an integrated value of the unit length (T) and the parameter i is smaller than the length (time) of the operation action data (S121). If it is determined in step S121 that the integrated value of the unit length (T) and the parameter i is smaller than the length (time) of the operation action data, the living action recognition unit 321 attaches living action meta information to the segmented data between time T*i and time T*(i+1) (step S122). The symbol “*” indicates integration processing. In step S122, the meaning and content (living action meta information) applicable to the segmented data at the time can be attached by using, for example, ruled-based branching processing. Alternatively, living action meta information can also be attached the segmented data using machine learning such as the Hidden Markov Model (HMM) or Neural Network. The number of pieces of living action meta information attached to the segmented data is not limited to one and a plurality of pieces of living action meta information may be attached.
  • When living action meta information is attached to the segmented data of the operation action data in step S122, the living action recognition unit 321 adds 1 to the parameter i (S123) to repeat the processing from step S121. If it is determined in step S121 that the integrated value of the unit length (T) and the parameter i is equal to or greater than the length (time) of the operation action data, the living action recognition unit 321 outputs each piece of segmented data to which living action meta information is attached by the processing in step S122 as living action data (S124). An output result of the living action recognition unit 321 may be recorded in a predetermined storage unit (not shown) or may be output directly to the functional unit (in this case, the hierarchical structure judgment unit 322) that performs the next processing.
  • To return to the description of FIG. 8, when living action data as information in which living action meta information is attached to segmented data constituting operation action data by the living action recognition unit 321 is generated, the hierarchical structure judgment unit 322 attaches hierarchical information representing a hierarchical relationship about the meaning and content to the living action data. The processing will be described based on FIG. 10. The hierarchical structure judgment unit 322 first determines whether the input living action data satisfies at least one of conditions that no hierarchical information is attached and it is possible to further shift to a higher level from the attached hierarchical information (S131).
  • If none of these conditions is satisfied in step S131, that is, the highest hierarchical information is already attached, the hierarchical structure judgment unit 322 terminates the processing shown in FIG. 10. On the other hand, if one of the above conditions is satisfied in step S131, the hierarchical structure judgment unit 322 combines adjacent segmented data of the same action of each piece of segmented data arranged in chronological order (S132). In step S132, processing to regard data discretized in step S120 as one operation action when the same action continues is performed. A chunk (segment) generated by combining segmented data may be recorded in a predetermined storage unit (not shown) (S133).
  • Next, if the chronological arrangement order of segments generated in step S132 is set as j (j=1 to M), the hierarchical structure judgment unit 322 determines whether j is smaller than the number of segments generated in step S131 and dictionary information about the action of the j-th segment can be acquired (S134). If both of these conditions are satisfied in step S134, the hierarchical structure judgment unit 322 selects and attaches dictionary information optimum to the relevant segment of the acquired dictionary information (S135). Then, the hierarchical structure judgment unit 322 determines whether still higher hierarchical information can be attached to the segment (j) based on the selected dictionary information and temporal context (S136). If it is determined in step S136 that higher hierarchical information can be attached, the hierarchical structure judgment unit 322 attaches higher hierarchical information to the segment (j) and adds 1 to the parameter j (S137). Then, the processing in step S134 and thereafter is repeated.
  • On the other hand, if it is determined in step S134 that j is equal to or greater than the number of segments generated in step S132 or dictionary information about the action of the j-th segment cannot be acquired, the hierarchical structure judgment unit 322 repeats the processing in step S131 and thereafter. Also when it is determined in step S136 that higher hierarchical information cannot be attached to the segment (j), the hierarchical structure judgment unit 322 repeats the processing in step S131 and thereafter.
  • As shown in FIG. 10, the hierarchical structure judgment unit 322 segments segmented data generated in step S120 by piecing together segmented data as the same action that is temporally consecutive. Then, related dictionary information is acquired and attached to each segment and whether still higher hierarchical information can be added is judged. Thus, by performing the processing in steps S131 to S137, an action segment (unit segment) in the minimum unit to which hierarchical information is attached is generated.
  • FIGS. 11 and 12 show a flow chart showing an example of action segment generation processing. As shown in FIG. 11, the hierarchical structure judgment unit 322 determines whether “another action” sandwiched between the same action of segmented data is present (S200). If “another action” sandwiched between the same action is present, “another action” is merged into the action before and after “another action” (S202). However, merge processing of “another action” sandwiched between operation actions of “walking” whose precision is sufficiently high and which is likely to be a point of change is not performed. After “another action” is merged in step S202 or there is no “another action” sandwiched between the same action in step S200, the hierarchical structure judgment unit 322 merges the same actions that are consecutive (S204).
  • Then, the hierarchical structure judgment unit 322 sets the parameter k (k=1 to K) representing the chronological order of segments generated by merge processing to the initial value 1 (S206) and determines whether the action time of the segment (k) is shorter than a predetermined time T1 (for example, T1=3 min) (S208). If the action time of the segment (k) is shorter than the predetermined time T1 in step S208, the hierarchical structure judgment unit 322 accumulates the segment in a buffer (S210). On the other hand, if the action time of the segment (k) is determined to be equal to or longer than the predetermined time T1 in step S208, the hierarchical structure judgment unit 322 further determines whether the action time of the segment (k) is shorter a predetermined time T2 (T2>T1; for example, T2=10 min) (S212).
  • If the action time of the segment (k) is determined to be shorter than the predetermined time T2 in step S212, the hierarchical structure judgment unit 322 merges the segment (k) into the action immediately before (S214). On the other hand, if the action time of the segment (k) is equal to or longer than the predetermined time T2 in step S212, the hierarchical structure judgment unit 322 decides the operation action of the segment as “another action” (S216). Then, the hierarchical structure judgment unit 322 determines whether the processing of steps S208 to S216 has been performed for all segments (S218) and if there is any unprocessed segment, the hierarchical structure judgment unit 322 adds 1 to k (S220) and then repeats the processing in step S208 and thereafter,
  • On the other hand, if the processing of steps S208 to S216 has been performed for all segments, as shown in FIG. 12, processing to merge consecutive segments of the same action is performed (S222). Then, if a vehicle action of only the data unit time (for example, only 1 min) is present, the action of the segment is set as “another action” (S224).
  • Next, the hierarchical structure judgment unit 322 determines whether action content of the segment is “walking” (S226) and, if the action content is other than “walking”, accumulates the segment in the buffer (S228). On the other hand, if the action content of the segment is “walking”, the hierarchical structure judgment unit 322 determines whether any vehicle action is accumulated in the buffer (S230). If a vehicle action is accumulated in the buffer, the hierarchical structure judgment unit 322 sets the operation action of the segment as an action of vehicle with the maximum share from “walking” (S323). On the other hand, if no vehicle action is accumulated in the buffer, If a vehicle action is accumulated in the buffer, the hierarchical structure judgment unit 322 sets the operation action of the segment as “another action” (S324).
  • Here, hierarchical information added to an action segment will be described based on FIG. 13. As shown in FIG. 13, it is assumed that the operation content of six unit segments (SG01 to SG06) is “train”, “train”, “walking”, “walking”, “train”, and “train”. The unit segment SG01 and the unit segment SG02 are merged into an action segment SG07, the unit segment SG03 and the unit segment SG04 are merged into an action segment SG08, and the unit segment SG05 and the unit segment SG06 are merged into an action segment SG09 by the processing shown in FIG. 10. The state is set as the grain size 1-0. The action segments SG07 to SG09 in the grain size 1-0 becomes an action segment SG17 of “got on a train on some line (action B)”, an action segment SG18 of “trains changed (action C)”, and an action segment SG19 of “got on a train on some line (action C)” respectively. The grain size at this point is set as 1-1. If the action segments SG17 to SG19 are further merged from the grain size 1-1, one action segment SG10 of “got on a train (action A)” is obtained. The grain size at this point is set as 1-2.
  • An action segment holds hierarchical information by being attached to the action segment combining unit segments or by being attached to the unit segments. When hierarchical information is attached to an action segment combining unit segments, it is assumed that, for example, an action segment SG17 of the action B in FIG. 13 is focused on. The action segment SG17 is considered to be an action segment newly generated by combining the unit segments SG01 and SG02. In this case, the facts that the action segment has the grain size 1-1 and the action content is “got on a train on some line”, and the start time and the end time of the action are attached to the action segment SG17 as hierarchical information. Alternatively, the fact that the action segment SG17 includes the unit segments SG01 and SG02 may be attached to the action segment SG17 as hierarchical information.
  • On the other hand, when hierarchical information is attached to unit segments as action segments of the minimum unit, it is assumed that, for example, an action segment SG01 in FIG. 13 is focused on. In this case, the facts that the action segment is a unit segment and is contained in the action segment SG07 in the grain size 1-0, in the action segment SG17 in the grain size 1-1, and in the action segment SG10 in the grain size 1-2 is attached to the action segment SG01 as hierarchical information. More specifically, information in which operation content is associated in the order of hierarchy like, for example, [train, action B, action A] can be represented as hierarchical information.
  • Hierarchical information may be attached in any form and can also be attached in other forms. The case of attaching hierarchical information to an action segment combining unit segments is superior in terms of the amount of data and the case of attaching hierarchical information to unit segments is superior in terms of a database search.
  • Returning to the description of FIG. 8, when action segments to which hierarchical information is attached from living action data are generated in step S130, the living action recognition unit 321 outputs each action segment to the data management unit 330 (S140). The data management unit 330 records the acquired action segment in a storage unit (the unit data storage DB 334 or the hierarchical information attached data storage DB 335).
  • 3-4. Processing Content of Action Representation Generation Unit
  • The analysis server 300 can accumulate an action log by action segments in real time and at the same time, can generate an action representation based on the meaning and content of an operation action. The analysis server 300 can also generate an action representation from a past action history. A detailed configuration of the action representation generation unit 320 and the data management unit 330 of the analysis server 300 is shown in FIG. 14. FIG. 14 is functional block diagram showing the functional configuration of the analysis server 300.
  • As shown in FIG. 14, the action representation generation unit 320 can include, in addition to the living action recognition unit 321 and the hierarchical structure judgment unit 322 described above, a hierarchical processing unit 323, a registration processing unit 324, a comment creation unit 325, and an acquisition unit 326.
  • The hierarchical processing unit 323 performs subsequent processing of a judgment result of the hierarchical structure judgment unit 322. The hierarchical processing unit 323 functions based on a hierarchical structure when only a portion of data to be attached to action segments is recorded in the storage unit for slimming down or speedup of data or hierarchical information of the specified action segment is delivered to an application.
  • As described above, hierarchical information may be attached to an action segment combining unit segments or to the unit segments. When hierarchical information is attached to a combined action segment, the hierarchical processing unit 323 processes the action segment of the hierarchical information selected by the user via the input unit 160. On the other hand, when hierarchical information is attached to unit segments, the hierarchical processing unit 323 generates an action segment by combining unit segments based on the hierarchical information selected by the user via the input unit 160. The hierarchical processing unit 323 a processing result of information to the registration processing unit 324 and the comment creation unit 325.
  • The registration processing unit 324 records the action segment generated by the hierarchical processing unit 323 in the data management unit 330. The registration processing unit 324 outputs an action segment to the data acquisition unit 331 to record the action segment in the hierarchical information attached data storage DB 335.
  • The comment creation unit 325 creates and attaches a comment such as the meaning and content of an action to a generated action segment. A comment created by the comment creation unit 325 is output to the data acquisition unit 331. The data acquisition unit 331 associates the comment with the corresponding action segment and records the comment in, for example, the hierarchical information attached data storage DB 335.
  • The acquisition unit 326 acquires a predetermined action segment from the unit data storage DB 334 or the hierarchical information attached data storage DB 335. When, for example, processing that needs to use a past action log is performed by the action representation generation unit 320, the acquisition unit 326 past data recorded in the unit data storage DB 334 or the hierarchical information attached data storage DB 335. Data to be acquired is decided based on instructions from the user.
  • 4. Action Recording and Display Application
  • As described above, by analyzing operation action data acquired by the action recording device 100 by the analysis server 300, an action segment to which the meaning and content of an operation action is attached is generated. An application function to represent an action log of the user using action segments will be described below.
  • 4-1. Representation of Action Log Based on Action Segment
  • First, the representation of an action log using action segments will be described. An example of the action log is shown in FIG. 15. The action log in FIG. 15 can be displayed, for example, in the display unit 150 of the action recording device 100.
  • An action log includes action segments arranged in chronological order. For each action segment, for example, the start time and the end time of the operation action and operation content are displayed. When the operation content is a movement action like “movement by train”, a position representation like, for example, from the start location to the goal (for example, “from Gotanda to Ohsaki”) is added to the operation content. When the operation content is other than a movement action like “work” and “meal”, the location (for example, “in Ohsaki”) where the operation is performed is added to operation content.
  • Further, to notify the user of the operation of such an action segment in an easy-to-understand manner, an object of operation content may be displayed or an object showing the feeling of the user when performing the operation may also be displayed. Also, the number of steps (step value) for the user to perform the operation or a value (exercise value) indicating energy consumption by the operation may be displayed. The content displayed in each action segment constituting an action log is not limited to the example of FIG. 15 and information obtained from sensor information acquired by sensors may further be displayed in each action segment.
  • In the action log shown in FIG. 15, there are locations where times if adjacent action segments are not continuous. For example, there is a gap between the action of “meal at Ohsaki” started at 12:30 and the action of “was in Shibuya” started at 14:30. In this manner, the action log may be prevented from displaying all action segments. The non-display of an action segment may be caused by, for example, editing/deletion of the action segment by the user or by setting a non-display filter to prevent a portion of action segments from being displayed.
  • As the non-display filter, for example, a filter that prevents the display when the action time is short or a filter that prevents the display of an action segment judged to be unimportant to the use can be considered. Also, a filter that prevents the display when the precision of recognition is low or a filter that allows the display of an action or location specified by the user may be set.
  • 4-2. Browsing Action
  • When an action log display application is activated in a browsing terminal (for example, the action recording device 100), for example, the user can browse the action log of the user in predetermined units, for example, in units of days. FIG. 16 shows a display example of the action log when the action log display application is activated. FIG. 16 shows a state in which an action log 410 in units of days is displayed in the display unit 150 of the action recording device 100.
  • The action log 410 includes action segments 412 arranged in chronological order, for example, from the upper end of the screen toward the lower end. In each of the action segments 412, as described in FIG. 15, the location where an operation is performed, description of the location, type of the action, number of steps by the user for the operation, exercise value and the like are displayed. If the action log for one day cannot be displayed in the screen of display, the display unit 150 can be caused to display an action segment that is not displayed by performing a screen scroll operation.
  • By touching a Prev button 422 to display the action log of the previous day of the action log currently displayed or a Next button 424 to display the action log of the next day of the action log currently displayed, the display unit 150 can be caused to display an action log of another day. If the action log of the previous day is not present when the Prev button 422 is pressed, the display unit 150 may be caused to display an action log of the day when an action log is acquired next by further going back to the past. Similarly, if the action log of the next day is not present when the Next button 424 is pressed, the display unit 150 may be caused to display an action log of the day when an action log is acquired next by further moving to the present. Operation buttons 430 for browsing, editing and other operations of the action log 410 are displayed in the display unit 150 and the user can touch the button corresponding to desired processing to perform the processing.
  • If, for example, a calendar button 434 is touched, as shown in FIG. 17, a calendar 440 is displayed in the display unit 150. The calendar 440 displayed first in the display unit 150 after the screen transition may be the current month or the month corresponding to the action log 410 displayed before the screen transition. In the calendar 440, a day for which the action log 410 can be displayed, that is, a day for which the action log 410 is accumulated as data can be selected by, for example, a touch operation. When some day is selected from the calendar 440, the action log 410 of the day is displayed in the display unit 150.
  • The display of the calendar 440 is changed by a previous (<) button 442 or a next (>) button 444. If the previous (<) button 442 is operated, the calendar of the previous month is displayed and if the next (>) button 444 is operated, the calendar of the next month is displayed. If no action log of the previous month is present when the previous (<) button 442 is pressed, the calendar 440 of a month when any action log is acquired next may be displayed by further going back to the past. Similarly, if no action log of the next month is present when the next (>) button 444 is pressed, the calendar 440 of a month when any action log is acquired next may be displayed by further moving to the present.
  • FIG. 17 shows an example in which the calendar 440 in units of months is displayed, but the present technology is not limited to such an example and can display a calendar, for example, in units of weeks, two weeks, or years. In addition, an icon indicating a representative operation action for the day may be displayed for all days for which an action log can be displayed in the calendar 440. Accordingly, the user can recognize the action of the day at a glance. By using such a calendar function, the action log 410 desired to browse can easily be searched for and the display unit 150 can be caused to display the action log. To cause a transition from the display screen of the calendar 440 to the previous screen, a Back button 446 may be operated.
  • If, for example, a map button 431 of the operation buttons 430 is touched, the action log display application activates a map 450 to display position information corresponding to the action log 410 in the map 450. FIG. 18 shows a display example when the map 450 is caused to display position information corresponding to the action log 410 by operating the map button 431.
  • If, when the action log 410 is displayed, the map button 431 is touched while none of the action segments 412 constituting the action log 410 is selected, for example, a history of all position information of the action log 410 of the day is displayed on the map 450. If the user is on the move, a movement locus thereof is displayed on the screen.
  • On the other hand, if, when the action log 410 is displayed, the map button 431 is touched while one action segment 412 a is selected from the action log 410, a history of position information of the action segment 412 a is displayed on the map 450. If, for example, as shown on the left of FIG. 18, the action segment 412 a of “movement from Saginomiya to Ohsaki by train” is selected and the map button 431 is touched, as shown on the right of FIG. 18, a movement locus by train from the start location (Saginomiya) to the goal (Ohsaki) is displayed on the map 450. Accordingly, on which line the user moved can be presented in an easy-to-understand manner. While position information is displayed in the map 450, time information 452 of the position information may be displayed in the display unit 150.
  • If the user is not on the move, an icon or the like may be displayed in a location where the operation is performed. In FIG. 18, a case when the one action segment 412 is selected is described, but the present technology is not limited to such an example and a plurality of the action segments 412 can be selected and position information of all the selected action segments 412 can be displayed on the map 450. In this case, the position information on the map 450 may be distinguished by color-coding or the like for each action segment and displayed. To cause a transition from the display screen of the map 450 to the previous screen, a Back button 454 may be operated.
  • 4-3. Correcting Action
  • In the foregoing, the method of displaying the action log 410 in the display unit 150 using the action segments 412 analyzed and generated by the analysis server 300 has been described. However, display content of the generated action log 410 may be erroneous. In such a case, the user can correct content of the action log 410. The correction feedback is reflected in action recognition determination processing. First, the method of correcting the action log 410 will be described based on FIGS. 19 and 20. FIG. 19 is an explanatory view showing a state in which a correction screen to correct the action segment to be corrected is displayed. FIG. 20 is an explanatory view showing an example of the correction screen to correct an operation action.
  • To correct content of the action segment 412, the user selects the action segment 412 a to be corrected from the action log 410 displayed in the display unit 150 and touches an edit button 435. Then, as shown on the right of FIG. 19, a correction screen 460 to correct the action segment 412 a to be corrected is displayed. If, for example, the correction of the action segment 412 a of “movement from Saginomiya to Ohsaki by train” is selected, a screen transition occurs from the display screen of the action log 410 to the correction screen 460 in which the operation content, start location and end location of the operation, and feeling during operation can be edited.
  • The operation content can be corrected in an operation content correction area 461 of the correction screen 460. If, for example, the operation content correction area 461 is selected, as shown in FIG. 20, an operation content candidate list 461 a in which operation content candidates for correction are listed is displayed. The user can correct the operation content by selecting the correct operation content from the operation content candidate list 461 a. When “Free input” is selected from the operation content candidate list 461 a, an input field into which the user can freely input operation content is displayed and the correct operation content can be input.
  • After selecting operation content from the operation content candidate list 461 a, the user continues to correct the start location and end location of the operation. At this point, if the selected operation content is a movement action like, for example, “movement by bus”, corrections of a start location correction area 462, a start location description correction area 463, an end location correction area 464, or an end location description correction area 465 can be made.
  • A location name list may be displayed for the start location correction area 462 and the end location correction area 464 so that the user can select and input the location name or the user may be enabled to directly input the location name. In the location name list, for example, location names to be a landmark such as a building name, station name, or shop name may be displayed. If there is no location to be a landmark, place names (addresses) may be displayed in the location name list.
  • In addition the display unit 150 may be caused to display a start location map 462 a and an end location map 464 a that display a map by being linked to input content of the start location correction area 462 and the end location correction area 464. The start location map 462 a and the end location map 464 a can be caused to display a map of any location by a scroll operation on the map. When a touch operation is performed on a map displayed on the start location map 462 a or the end location map 464 a, the location name corresponding to the position where the touch operation is performed may automatically be input into the start location correction area 462 or the end location correction area 464.
  • The start location description correction area 463 and the end location description correction area 465 are areas where what kind of location the location input into the start location correction area 462 and the end location correction area 464 is for the user is input respectively. When the start location description correction area 463 or the end location description correction area 465 is touched, for example, as shown in FIG. 20, a description candidate list 463 a or 465 a is displayed. The user can input a location description by selecting the correct description from the description candidate list 463 a or 465 a.
  • As the description content of a location, for example, “location to go back to” like the home, “location to work” like a company, and “location to learn” like a school can be cited. By inputting the description of such a location, what king of meaning the location has for the user can be grasped and a contribution can be made to improve the precision of action recognition for the user. If no correct description is found in the description candidate list 463 a or 465 a, a description may directly be input into the start location description correction area 463 or the end location description correction area 465.
  • When the operation content selected from the operation content candidate list 461 a in FIG. 20 is an action other than a movement action like, for example, “shopping” or “work”, the end location correction area 464 and the end location description correction area 465 can be corrected. The correction method is as described above.
  • An object indicating the feeling of the user when an operation is performed can be corrected by, for example, as shown in FIG. 19, a feeling correction area 466. The feeling correction area 466 includes a “no feeling setting” button and a feeling selection button that stepwise selects good or bad of feelings. The feeling selection button can be configured so that, for example, five levels of feeling of “very bad (irritated or depressed)”, “bad (somewhat irritated)”, “not bad (no problem)”, “good (a bit fine)”, and “very good (refreshing, bracing, full)” are selectable. The user can select the feeling when an operation is performed from the feeling correction area 466.
  • When all corrections are completed, correction content can be reflected in the action segment 412 a by pressing a save button 468 a at the bottom or a save button 468 b at the top of the correction screen 460. When the save button 468 a or 468 b is pressed, a transition to the screen before the transition to the correction screen 460 occurs. When a transition to the screen before the transition to the correction screen 460 should occur without reflecting input content in the correction screen 460, a cancel button 467 a at the bottom or a cancel button 467 b at the top of the correction screen 460 may be pressed.
  • 4-4. Combining Actions
  • According to the present technology, the display of the action log 410 can easily be changed not only by correcting content of each of the action segments 412, but also by correcting relationships between the action segments 412. For example, a plurality of the action segments 412 may be combined to display the resultant segment as the one action segment 412. The combination of the action segments 412 is a function to combine the plurality of the action segments 412 that are temporally consecutive into the one action segment 412. The time range of the combined action segment 412 extends from the start time of the oldest action segment 412 to the end time of the newest action segment 412.
  • FIG. 21 shows an example of the method of combining the action segments 412. First, if a merge button 432 is pressed while the action log 410 is displayed, a state (action combination mode) in which the action segments 412 can be combined is entered. The selected action segments 412 can be combined by selecting the action segment 412 to be combined in action combination mode.
  • The action of an action segment 412 b selected first among the action segments 412 to be combined can be set as operation content after the combination. In FIG. 21, the operation content of “work in Ohsaki” becomes the operation content after the combination. Then, the other action segments 412 to be combined are selected. The selection of the action segment 412 can be made by a touch operation or a drag operation. In FIG. 21, action segments surrounded by a reference sign 412 c are selected for combination. Then, if the save button 468 a is pressed after the action segments 412 to be combined are selected, the action segments 412 to be combined are displayed by being combined. The combined action segment 412 represents the action of “work in Ohsaki” is done between 9:55 and 22:42. To cancel the combination mode, the cancel button 467 a may be pressed.
  • In the example of FIG. 21, the operation content of the action segment 412 after the combination is decided in favor of the action segment 412 selected first, but the present technology is not limited to such an example. An example of another method of deciding operation content of the action segment 412 after combination is shown in FIG. 22. In FIG. 22, an action segment to be combined is selected by a drag operation. That is, the first action segment (start segment) to be combined is first selected by contact with a finger and the finger is moved while in contact until the last action segment (end segment) to be combined is selected. In this manner, the action segments (action segments surrounded by the reference sign 412 c) to be combined are decided. Then, when the save button 468 a (see FIG. 21) is pressed, the action segment after the combination is displayed.
  • For example, the operation content of any action segment to be combined may be set as the operation content of the action segment after the combination. If, for example, the operation content of “walked in Ohsaki” is selected, the operation content of an action segment 412 d 1 after the combination becomes “walked in Ohsaki”.
  • The operation content of the action segment after the combination may be decided, for example, by majority of action segments to be combined. In the example of FIG. 22, for example, among the four action segments 412 to be combined, the operation content of the two action segments 412 is “work in Ohsaki”, the operation content of the one action segment is “meal in Ohsaki”, and the operation content of the other one action segment is “walked in Ohsaki”. Therefore, the operation content of “work in Ohsaki” of the most action segments is decided as the operation content of the action segment 412 d 2 after the combination.
  • Alternatively, the operation content of the action segment after the combination may be decided, for example, by reanalyzing action segments to be combined. In the example of FIG. 22, for example, the operation content of the four action segments 412 to be combined includes “work in Ohsaki”, “meal in Ohsaki”, and “walked in Ohsaki”. Accordingly, the meaning and content of user's actions are reanalyzed and, for example, the operation content of “was in Ohsaki” can be decided as the operation content of an action segment 412 d 3 after the combination.
  • Therefore, the action segments 412 can easily be combined by selecting the action segments to be combined.
  • 4-5. Dividing Action
  • Also according to the present technology, for example, a plurality of the action segments 412 can be divided and displayed as a plurality of the action segments 412 as a correction of the relationship between the action segments 412. The division of the action segment 412 is a function to segment the one action segment 412 into a plurality of the action segments 412. As the division method of the action segment 412, for example, a method of setting the time to divide the action segment 412 and a division method using hierarchical information are known.
  • For example, FIG. 23 shows an example of a division method by time settings of the action segment 412. First, if a division 433 is pressed while the action log 410 is displayed, a state (action division mode) in which the action segment 412 can be divided is entered. The selected action segment 412 can be divided at a specified time by selecting the action segment 412 to be divided in action division mode.
  • For example, it is assumed that an action segment 412 e is selected for division in FIG. 23. Then, a division time setting screen to input the division time of the action segment 412 e to be divided is displayed. In the division time setting screen, ant time between the start time and the end time of the action segment to be divided can be set. Then, pressing the save button 468 a divides action segment 412 to be divided and displays the divided action segments 412. If, for example, the division time is set to 11:50, the action segment 412 is divided into a first action segment of “work in Ohsaki” between 9:58 and 11:50 and a second action segment of “work in Ohsaki” between 11:50 and 12:29. To cancel the division mode, the cancel button 467 a may be pressed.
  • Also, for example, FIG. 24 shows an example of a division method based on hierarchical information of the action segment 412. It is assumed that the action division mode is already set in FIG. 24. If the user selects the action segment 412 e to be divided while the action log 410 is displayed, as shown in FIG. 24, a hierarchical list 414 representing the action segment 412 e to be divided by more detailed action segments is displayed. When the hierarchical list 414 is displayed, the user selects the action segment to be the division point.
  • If, for example, an action segment 414 a is selected in FIG. 24, for example, the end time of the action segment 414 a is decided as the division time and, as shown on the right of FIG. 24, the action segment is divided into a first action segment 412 e 1 of “work in Ohsaki” between 9:30 and 11:59 and a second action segment 412 e 2 of “work in Ohsaki” between 11:59 and 12:30.
  • 4-6. Representation of Action Segment Based on Segmentation Grain Size
  • In the present technology, the action segments 412 constituting the action log 410 hold a hierarchical relationship based on the meaning and content thereof as hierarchical information. The display roughness of the displayed action log 410 can be changed by changing the segmentation grain size using the hierarchical information. The display roughness can be changed by using, for example, a slider or a zoom button.
  • FIG. 25 shows a case when display roughness is changed by using a slider 471. A knob 472 to set the display roughness (grain size) is displayed in the slider 471 and the display roughness can be changed by changing the position of the knob 472. When, for example, as shown on the left of FIG. 25, the knob 472 is positioned on the side of small grain size of the slider 471, the action log 410 is displayed by the detailed action segments 412. As the knob 472 is moved toward the side of large grain size of the slider 471 from the above state, a plurality of the action segments 412 are displayed by being combined based on the segmentation grain size attached to the action segments 412 in advance.
  • FIG. 26 shows a case when the display roughness is changed by using a zoom button 473. A plurality of buttons 474 to set the display roughness (grain size) is displayed in a row in the zoom button 473. By checking one of the plurality of buttons 474, the action log 410 can be made to be displayed in the display roughness corresponding to the button 474. When, for example, as shown on the left of FIG. 26, the button 474 on the side of small grain size in the zoom button 473 is checked, the action log 410 is displayed by the detailed action segments 412. As the button 474 on the side of large grain size is checked from the above state, a plurality of the action segments 412 are displayed by being combined based on the segmentation grain size attached to the action segments 412 in advance.
  • Thus, the display roughness of the action log 410 can easily be changed based on the segmentation grain size attached to the action segment 412 so that the user can view the action log 410 in the desired display roughness.
  • The display roughness of the action log 410 is changed in FIGS. 25 and 26, but in the present technology, the display roughness of the action segment 412 can also be changed depending on the purpose. That is, the display roughness of the action segment 412 is changed independently of the type of action. Work, shopping, movement and the like can be considered as the types of action and, for example, by checking a button 476 of the corresponding action from an action type selection list 475 as shown in FIG. 27, the display roughness of only the action can be changed.
  • When, for example, as shown in FIG. 27, a button 476 b of “work detail” is checked in the action type selection list 475, action segments 412 f 1, 412 g 1 related to work are displayed in detail. For example, the operation content of “work in Ohsaki” for the action segment 412 f 1 is displayed by five action segments 412 f 2 of “desk work in Ohsaki”, “movement on foot”, “meeting in Ohsaki”, “movement on foot”, and “desk work in Ohsaki”. Incidentally, the operation content of the action segment 412 g 1 is the most detailed and thus, the same content is displayed after the display roughness is changed (action segment 412 g 2).
  • When, for example, as shown in FIG. 28, a button 476 c of “shopping detail” is checked in the action type selection list 475, an action segment 412 h 1 related to shopping is displayed in detail. For example, the operation content of “shopping in Shibuya” for the action segment 412 h 1 is displayed by seven action segments 412 h 2 of “shopping in Shibuya”, “movement on floor”, “shopping in Shibuya”, “movement between shops in Shibuya”, “shopping in Shibuya”, “moving between shops in Shibuya”, and “shopping in Shibuya”.
  • Further, when, for example, as shown in FIG. 29, a button 476 d of “movement detail” is checked in the action type selection list 475, action segments 412 i 1, 412 j 1, 412 k 1 related to movement are displayed in detail. For example, the operation content of “movement from Saginomiya to Ohsaki by train” for the action segment 412 i 1 is displayed by five action segments 412 i 2 of “waiting for train in Saginomiya”, “movement from Saginomiya to Takadanobaba by train”, “trains changed in Takadanobaba”, “movement from Takadanobaba to Ohsaki by train”, and “movement in Ohsaki on foot”. Similarly, the operation content of “moving from Ohsaki to Shibuya by train” for the action segment 412 j 1 is displayed by four action segments 412 j 2 of “movement to Ohsaki station”, “waiting for train in Ohsaki station”, “movement from Ohsaki to Shibuya by train”, and “movement to Shibuya”. For the action segment 412 k 1, similarly detailed content is displayed.
  • When the action segments 412 should be displayed in the same display grain size regardless of the action, for example, as shown in FIG. 30, a button 476 a of “uniform detail” may be checked in the action type selection list 475. Accordingly, all the action segments 412 of the action log 410 are displayed in detail in the same grain size.
  • Thus, because the display roughness of the action segments 412 can be changed independently in accordance with the type of action, only the action the user wants to check in detail can be displayed in detail.
  • Incidentally, the method of changing the display roughness shown in FIGS. 25 and 26 and the method of changing the display roughness in accordance with the type of action shown in FIGS. 27 to 30 may be combined. For example, as shown in FIG. 31, a slider may be provided for each type of action so that the display roughness of each type of action can be adjusted. FIG. 31 is shows a display grain size setting unit 477 provided with a slider 478 a to set the display roughness of an action segment related to “work”, a slider 478 b to set the display roughness of an action segment related to “shopping”, and a slider 478 c to set the display roughness of an action segment related to “movement”. By moving respective knobs 479 a, 479 b, 479 c of the sliders 478 a, 478 b, 478 c of the display grain size setting unit 477, the display roughness can be adjusted for each type of action.
  • 4-7. Deleting Action
  • According to the present technology, the action segment 412 can be deleted from the action log 410. If, for example, as shown in FIG. 32, an action segment 412 l to be deleted is selected and a deletion button 436 is pressed while the action log 410 is displayed, as shown on the right of FIG. 32, a deletion confirmation screen 480 is displayed. In the deletion confirmation screen 480, the user can be caused to enter the reason for deleting the action segment 412 l to be deleted. When the user presses a button on which the reason for deletion is written, the action segment 412 l to be deleted is deleted from the action log 410. Depending on the reason selected by the user, the deletion of the action segment may be fed back as an action correction.
  • 4-8. Posting Action
  • According to the present technology, content of the action segment 412 of the action log 410 can be posted. If, for example, as shown in FIG. 33, an action segment 412 m to be posted is selected and a post button 437 is pressed while the action log 410 is displayed, as shown on the right of FIG. 33, a posting screen 482 is displayed. In the posting screen 482, the operation content of the action segment 412 m to be posted is automatically pasted to a posting content input area 482 a. When the user presses a posting button 482 b, the description content in the posting content input area 482 a is posted to a posting site.
  • 4-9. Action Log Acquisition Stop Processing
  • According to the present technology, when the acquisition of an action log should be stopped for some reason, for example, as shown in FIG. 34, a settings screen 490 is made to display by pressing a settings button 438. In the settings screen 490, various settings about the action log display application can be made. When, for example, the acquisition of the action log 410 should be stopped, “stop” of an acquisition function setting unit 491 that sets the operation of the action log acquisition function is selected. Accordingly, the action log display application stops the action log acquisition function. To restart the stopped action log acquisition function, “restart” of the acquisition function setting unit 491 may be selected.
  • 4-10. Updating Display Content
  • The action log display application in the present technology automatically uploads operation action data acquired by the action recording device 100 to the action log server 200 in predetermined timing (for example, twice per day). Also, the analysis server 300 automatically generates an action segment in predetermined timing (for example, twice per day). While an action log is displayed based on generated action segments, an action log displayed in accordance with the system function or circumstances may not correspond to the latest results. Thus, by pressing an update button 493 that updates the action log displayed in the settings screen 490 of FIG. 34 to the latest results, the action log can be updated to the latest results. When a transition to the action log display screen occurs after the update button 493 being pressed, the display unit 150 can be caused to display the latest results.
  • 5. Reflection Processing of Correction Feedback
  • In an action log display system in the present technology, the meaning and content of an action is analyzed by the analysis server 300 and an action log is displayed by action segments. However, as described above, content of the displayed action log may not all correct. Thus, according to the present technology, the user can make corrections to correct content by using the action log display application. In the present technology, correction feedback of the user is reflected in the next analysis processing by the analysis server 300 and used to improve the precision of the next and subsequent analysis results. The reflection processing of correction feedback will be described below based on FIGS. 35 to 42.
  • 5-1. Properties of Correction Feedback
  • In the present technology, the precision of analysis results is improved by reflecting correction feedback of the user in analysis processing, but the user may not correct all errors of analysis results by the analysis server 300. That is, content of an action log that is not corrected may not necessarily be correct. Thus, in the present technology, it is necessary to assume a system capable of collecting substantially biased information only. In addition, analysis results before corrections by the user do not necessarily match the latest analysis results. Thus, by reflecting information showing which action segment is corrected in what way in analysis processing for each user, the action specific to each user can be learned, which is considered to effectively work to improve the precision of analysis results.
  • 5-2. Action Recognition Processing
  • In consideration of the above points, according to the present embodiment, an action pattern is decided based on characteristic amount analysis results in recognition processing of an operation action and acquires a plurality of probability distributions corresponding to the action pattern, time, and position information (location). In this case, a weight of a histogram is assigned and an operation action is recognized based on results of assigning weights depending on the location. If position information cannot be acquired or there is no need to acquire position information, uniform weights may be assigned or specific weights like “no location can be acquired” or “there is no need for location” may be assigned.
  • FIG. 35 shows an example of action recognition processing by the living action recognition unit 321. FIG. 36 shows operation action estimation information that decides an operation action. The operation action estimation information is, for example, information showing the relation between a weighting factor depending on the location and the probability distribution of each action and, as shown in FIG. 36, a plurality (four in FIG. 36) of probability distributions of actions of, for example, “shopping”, “work”, “meal”, and “others” is acquired. Then, the weighting factor depending on the location is set to each probability distribution. The operation action estimation information is preset and recorded in, for example, the analysis parameter DB 333.
  • After an action log in the unit time is acquired, the living action recognition unit 321 starts processing to recognize the action of the action log. First, as shown in FIG. 35, at least one pair of the probability distribution and the weighting factor depending on the location is acquired based on the action pattern, time information, and position information (S300).
  • Next, in steps S302 to S306, the living action recognition unit 321 performs processing to decide operation content of the action log in the unit time. First, it is assumed that the number of pairs of the probability distribution and the weighting factor acquired in step S300 is n and the parameter representing the processing number is i (i=0 to n) (S302). Then, the living action recognition unit 321 multiplies the probability distribution by the weighting factor of each action for the first (i=0) pair of the probability distribution and the weighting factor (S304). If, for example, in FIG. 36, the first pair is the pair of the probability distribution and the weighting factor in the first row, the probability of 50 is multiplied by the weighting factor of 1 for “shopping” and the probability of 10 is multiplied by the weighting factor of 1 for “work”. Then, the probability of 10 is multiplied by the weighting factor of 1 for “meal” and the probability of 30 is multiplied by the weighting factor of 1 for “others”. Accordingly, the integrated values (“shopping”: 50, “work”: 10, “meal”: 10, “others”: 30) of actions are acquired.
  • When the processing in step S304 is completed, the living action recognition unit 321 adds 1 to the parameter i (S306) and repeats the processing in step S302 and thereafter. In the example of FIG. 36, the living action recognition unit 321 multiplies the probability distribution by the weighting factor of each action for the next (i=1) pair of the probability distribution and the weighting factor, that is the pair of the probability distribution and the weighting factor in the second row (S304). For the pair in the second row, first the probability of 10 is multiplied by the weighting factor of 6 for “shopping” and the probability of 50 is multiplied by the weighting factor of 6 for “work”. Then, the probability of 10 is multiplied by the weighting factor of 6 for “meal” and the probability of 30 is multiplied by the weighting factor of 6 for “others”. Accordingly, the integrated values (“shopping”: 60, “work”: 300, “meal”: 60, “others”: 180) of actions are acquired.
  • Then, the living action recognition unit 321 adds the integrated value in the second row to the integrated value in the first row for each action. This results in integrated values of “shopping”: 110, “work”: 310, “meal”: 70, “others”: 210. Similarly, integrated values are calculated for the pairs of the probability distribution and the weighting factor in the third and fourth rows and these integrated values are added to the above integrated values of each action to finally obtain added values of “shopping”: 260, “work”: 460, “meal”: 420, “others”: 460.
  • The living action recognition unit 321 decides the action of the maximum final added value as the operation content of the action log. In the example of FIG. 36, both “work” and “others” have the maximum added value of 460, but in this case, the operation content is decided by giving priority to actions other than “others”. Therefore, in the example of FIG. 36, “work” is recognized as the operation content.
  • 5-3. Reflection Processing of Correction Feedback 5-3-1. Overview of Reflection Processing of Correction Feedback
  • As described based on FIGS. 35 and 36, recognition processing results of operation content changes considerably depending on operation action estimation information showing the relation between the weighting factor depending on the location and the probability distribution. Thus, the action log is corrected by the user, correction content is reflected in operation action estimation information recorded in the analysis parameter DB 333 by the feedback adjustment unit 332 of the analysis server 300. Accordingly, the precision of recognition processing of operation content can be enhanced.
  • An overview of reflection processing of correction feedback will be provided based on FIG. 37. It is assumed that the relation between the weighting factor depending on the location and the probability distribution of each action shown in FIG. 36 is held as operation action estimation information before correction. As a result of generating an action segment from operation action data based on such operation action estimation information, the user is assumed to input correction information (correction feedback) of the action segment from the input unit 160. The correction feedback is converted into data in a format that can be transmitted to the analysis server 300 by the input information processing unit 144 before being transmitted to the analysis server 300 via the client interface unit 130.
  • The analysis server 300 having received the correction feedback from the action recording device 100 through the analysis server interface unit 310 reflects content of the correction feedback in the operation action estimation information through the feedback adjustment unit 332. At this point, the feedback adjustment unit 332 corrects the probability distribution of the operation action estimation information if the content of the correction feedback concerns an action and corrects the weighting factor depending on the location if the content of the correction feedback concerns position information (location).
  • It is assumed that, for example, as shown in FIG. 37, the operation content of “work” is acquired as an analysis result, but correction feedback to change the operation content to “shopping” by the user is received. In this case, the feedback adjustment unit 332 corrects, among a plurality of probability distributions, the probability distribution having the maximum probability of “work”. For example, the feedback adjustment unit 332 makes a correction to set the probability of “work” as an analysis result and the probability of “shopping” as a correction result to the average value of these two probabilities for the probability distribution in the second row with the maximum probability of “work”.
  • It is assumed, on the other hand, that an analysis result of “location to work” is acquired, but correction feedback to change the location description to “location to do shopping frequently” by the user is received. In this case, the feedback adjustment unit 332 corrects, among a plurality of probability distributions, the weighting factor of the probability distribution having the maximum probability of “shopping”. For example, the feedback adjustment unit 332 makes a correction of increasing the weighting factor in the first row with the maximum probability of “shopping” by a factor of a predetermined number (for example, 10).
  • By correcting the operation action estimation information in this manner, correction feedback is reflected in analysis results of action segments so that the precision of analysis results of the operation content can be expected. The reflection processing of correction feedback will be described in more detail below based on FIGS. 38 to 42.
  • 5-3-2. Reflection Processing of Correction Feedback of an Action
  • First, the reflection processing of correction feedback of an action will be described based on FIGS. 38 to 40. FIG. 38 is a flow chart showing the reflection processing of correction feedback of an action. FIG. 39 is an explanatory view illustrating corrections of the operation action estimation information based on the processing in FIG. 38. FIG. 40 is a flow chart showing other reflection processing of correction feedback of the action. “001”, “002”, and “003” shown in the top row of each column in FIG. 39 are IDs representing respective actions. It is assumed in the description below that the analysis result by the living action recognition unit 321 is “action 002” and the correct action by user's correction feedback is “action 003”.
  • When correction feedback is received from the action recording device 100, the feedback adjustment unit 332 first recognizes correction content. It is assumed here that operation content of an action segment is corrected. The feedback adjustment unit 332 acquires the action segment to be corrected from the unit data storage DB 334 or the hierarchical information attached data storage DB 335 and starts processing shown in FIG. 38.
  • The feedback adjustment unit 332 first acquires the probability distribution (partial probability distribution) used to recognize the operation content of the action segment to be corrected from operation action estimation information stored in the analysis parameter DB 333 (S310). Next, the feedback adjustment unit 332 calculates a value M(i) obtained by multiplying the maximum probability of each probability distribution by the weighting factor of the row for the partial probability distribution and sorts these probability distributions (S311).
  • The parameter indicating the order of sorted probability distributions is set as i (i=0 to n) and the number of probability distributions constituting the partial probability distribution is set as n. Then, the feedback adjustment unit 332 determines whether the parameter i is smaller than n and the multiplied value M(i) is larger than a predetermined threshold th (S312). If the conditions in step S312 are not satisfied, the processing shown in FIG. 38 is terminated. If, for example, the operation action estimation information on the left of FIG. 39 is provided, only probability distributions of the fourth to sixth rows are corrected. On the other hand, if the conditions in step S312 are satisfied, the feedback adjustment unit 332 acquires a correction ratio C(i) from a loss ratio calculation function using an action having the maximum value of probability distribution of each action of target rows from operation action estimation information and the correct action obtained from correction feedback (S313).
  • The loss ratio calculation function is assumed to be a single comprehensive measure representing losses caused when some available decision is made. In the present embodiment, the loss ratio calculation function is used to set, for example, a correction ratio table representing a correction ratio C between the action of analysis results and the correct action as shown in the lower portion of FIG. 39. The correction ratio table can be preset and can be stored in the analysis parameter DB 333. The feedback adjustment unit 332 acquires the correction ratio C(i) between the action having the maximum value of probability distribution and the correct action “action 003” from the correction ratio table. If, for example, the processing in step S313 is performed for the probability distribution in the fourth row of the operation action estimation information on the left of FIG. 39 is performed, the correction ratio C(0)=0 between “action 003” having the maximum value of probability distribution and the correct action “action 003” is obtained.
  • Then, the feedback adjustment unit 332 subtracts the correction ratio C(i) acquired in step S313 from the value of the probability distribution of the action of the maximum value of probability distribution, adds the correction ratio C(i) to the value of the probability distribution of the correct action, and reflects these corrections in the operation action estimation information (S314). If, for example, the processing in step S314 is performed for the probability distribution in the fourth row of the operation action estimation information on the left of FIG. 39 is performed, the probability distribution of the row is not changed because the correction ratio C(0)=0. Then, the feedback adjustment unit 332 adds 1 to the parameter i (S315) and repeats the processing in step S312 and thereafter.
  • If, for example, the processing in step S313 is performed for the probability distribution in the fifth row of the operation action estimation information on the left of FIG. 39 is performed, the correction ratio C(1)=5 between “action 002” having the maximum value of probability distribution and the correct action “action 003” is obtained. Then, if the processing in step S314 is performed, the value of the probability distribution of “action 002” is corrected from 50 to 45 and the value of the probability distribution of “action 003” is corrected from 10 to 15 based on the correction ratio C(1)=5.
  • Similarly, if the processing in step S313 is performed for the probability distribution in the sixth row of the operation action estimation information on the left of FIG. 39 is performed, the correction ratio C(1)=10 between “action 001” having the maximum value of probability distribution and the correct action “action 003” is obtained. Then, if the processing in step S314 is performed, the value of the probability distribution of “action 001” is corrected from 40 to 30 and the value of the probability distribution of “action 003” is corrected from 10 to 20 based on the correction ratio C(1)=10. By performing the above processing, the operation action estimation information after the correction feedback being reflected as shown on the right of FIG. 39 can be obtained.
  • In this manner, the operation content of correction feedback is reflected in the operation action estimation information. The reflection processing shown in FIG. 38 is effective in being able to control the reflection speed or divergence. That is, a more impermissible error can be reflected earlier and a value can be made to converge if the value is handled as a ratio to the maximum value. In the reflection processing shown in FIG. 38, the correction ratio C(i) is added to or subtracted from the value of the probability distribution, but the present technology is not limited to such an example and, for example, correction feedback may be reflected in operation action estimation information by multiplying the value of the probability distribution by the correction ratio.
  • In the reflection processing of correction feedback shown in FIG. 38, content of the correction feedback is reflected in operation action estimation information by using a correction ratio table, but the present technology is not limited to such an example. For example, as shown in FIG. 40, a feedback system using a neural network technique may be configured. It is assumed that operation content of an action segment is corrected also in FIG. 40.
  • The feedback adjustment unit 332 first acquires the probability distribution (partial probability distribution) used to recognize the operation content of the action segment to be corrected from operation action estimation information stored in the analysis parameter DB 333 (S320). Next, the feedback adjustment unit 332 calculates a value M(i) obtained by multiplying the maximum probability of each probability distribution by the weighting factor of the row for the partial probability distribution and sorts these probability distributions (S321). The processing in steps S320, S321 can be made the same as the processing in steps S310, S311 in FIG. 38.
  • If the parameter showing the order of sorted probability distributions is set as i(i=0 to n), the feedback adjustment unit 332 determines whether the parameter i is smaller than n (S322). If the condition in step S322 is not satisfied, the processing shown in FIG. 40 is terminated. On the other hand, if the condition in step S322 is satisfied, the feedback adjustment unit 332 uses the neural network technique to reflect correction content of correction feedback in each probability distribution of operation action estimation information based on the weighting factor (S323). Then, the feedback adjustment unit 332 adds 1 to the parameter i (S324) and repeats the processing in step S322 and thereafter.
  • Thus, by using, instead of the correction ratio table, learning processing such as the neural network technique, content of correction feedback can be reflected in each value of operation action estimation information without the need to set the correction ratio table in advance.
  • 5-3-3. Reflection Processing of Correction Feedback of an Action and Position Information
  • Next, the reflection processing of correction feedback of an action and position information will be described based on FIG. 41. FIG. 41 is a flow chart showing the reflection processing of correction feedback of the action and position information.
  • The feedback adjustment unit 332 acquires the action segment to be corrected from the unit data storage DB 334 or the hierarchical information attached data storage DB 335 and starts processing shown in FIG. 41. The feedback adjustment unit 332 first determines whether correction feedback contains position information (S330). If the correction feedback contains position information in step S330, processing in step S331 and thereafter is performed to reflect corrected position information in operation action estimation information.
  • In step S331, whether any correction related to position information is made on action segments accompanied by movement is determined. If a correction related to position information is made on action segments accompanied by movement, representative coordinates of end points (two representative coordinates like position X to position Y) are calculated (S332). On the other hand, if no correction related to position information is made on action segments accompanied by movement, representative coordinates of the movement are calculated (S333). Incidentally, representative coordinates can be calculated by using the center, center of gravity, most frequent point and the like.
  • Next, the feedback adjustment unit 332 records representative coordinates calculated in step S332 or S333, the precision, and attached attributes in a feedback DB (not shown) (S334). The feedback DB is a storage unit provided in the analysis server 300. Then, the feedback adjustment unit 332 analyses operation content using new position information recorded in the feedback DV in step S334 and determines whether the analysis result matches the correct action input by the correction feedback (S335). If it is determined in step S335 that the operation content analyzed by using new position information matches the correct action, a judgment can be made that correction feedback about position information is correctly reflected and also there is no error in the action content. Therefore, the feedback adjustment unit 332 judges that the reflection processing of correction feedback is completed and terminates the processing in FIG. 41.
  • On the other hand, if it is determined in step S335 that the operation content analyzed by using new position information does not match the correct action, a judgment can be made that with corrections of position information alone, correction feedback is not correctly determined. In this case, processing in steps S336 to S341 is performed to reflect operation content of the correction feedback in operation action estimation information. The processing in steps S336 to S341 can be made the same as the processing in FIG. 38.
  • That is, the feedback adjustment unit 332 first acquires the probability distribution (partial probability distribution) used to recognize the operation content of the action segment to be corrected from operation action estimation information stored in the analysis parameter DB 333 (S336). Next, the feedback adjustment unit 332 calculates a value M(i) obtained by multiplying the maximum probability of each probability distribution by the weighting factor of the row for the partial probability distribution and sorts these probability distributions (S337).
  • The parameter indicating the order of sorted probability distributions is set as i (i=0 to n) and the number of probability distributions constituting the partial probability distribution is set as n. Then, the feedback adjustment unit 332 determines whether the parameter i is smaller than n and the multiplied value M(i) is larger than a predetermined threshold th (S338). If the conditions in step S338 are not satisfied, the processing shown in FIG. 41 is terminated. On the other hand, if the conditions in step S338 are satisfied, the feedback adjustment unit 332 acquires a correction ratio C(i) from a loss ratio calculation function using an action having the maximum value of probability distribution of each action of target rows from operation action estimation information and the correct action obtained from correction feedback (S339).
  • Then, the feedback adjustment unit 332 subtracts the correction ratio C(i) acquired in step S339 from the value of the probability distribution of the action of the maximum value of probability distribution, adds the correction ratio C(i) to the value of the probability distribution of the correct action, and reflects these corrections in the operation action estimation information (S340). Then, the feedback adjustment unit 332 adds 1 to the parameter i (S341) and repeats the processing in step S338 and thereafter. By performing the above processing, the operation action estimation information after the correction feedback being reflected can be obtained.
  • Incidentally, instead of the processing in steps S336 to S341, the processing shown in FIG. 40 may be performed. Also in this case, the operation action estimation information after the correction feedback being reflected can similarly be obtained.
  • The correction feedback of position information may be reflected by, as shown on the lower left of FIG. 37, changing the weighting factor of the probability distribution of operation action estimation information. Alternatively, a corrected attribute dependence section, a specified attribute dependence section, and an ID dependence section may be set to reflect the correction feedback of position information in each weighting factor. For example, the specified attribute dependence section is strengthened for the position information of “house” and a peripheral attribute dependence section is strengthened for the position information of “location to do shopping”. For the position information of, for example, “company”, a plurality of pieces of position information like different offices may be present. In this case, the position information of the same meaning can correctly be selected by giving differences like business content and scale to each piece of position information as respective features.
  • The added amount of weighting factor for position information may be decided based on, for example, original position information or changed for each attribute type of position information. Further, a probability distribution specific to position information may randomly be generated and added to operation action estimation information. Accordingly, over-learning can be prevented.
  • 6. Others 6-1. Personal Modeling of Action Pattern
  • In an action log display system according to the present embodiment, an action log is displayed by using action segments to which the meaning and content is attached. By performing, for example, autocorrelation processing or filter processing using these action segments, temporal or action errors can be absorbed. Then, a user's typical action pattern can be extracted from a small amount of data.
  • As a functional unit to extract a user's typical action pattern, as shown in FIG. 42, a typical action pattern generation unit 336 is provided in the analysis server 300. To extract a typical action pattern, an action log for a predetermined period (for example, for one day) acquired by the living action recognition unit 321 from the action log server 200 is first smoothed and then output to the typical action pattern generation unit 336. The typical action pattern generation unit 336 generates a typical action pattern using a statistical technique, for example, cross correlation processing on action segments of the smoothed action log.
  • In the example of FIG. 42, the typical action pattern generation unit 336 acquires a typical action pattern of the user as a result of analyzing, based on action logs of seven days, correlations of action segments of these action logs. Thus, by analyzing action segments to which hierarchical information represented by the meaning and content of actions and showing relations between action segments is attached, a user's probable typical action pattern can be generated even from a small amount of data.
  • 6-2. Position Display Technique by Moving Medium/Means Determination
  • If an action is recognized as a movement action when an action segment is generated, the living action recognition unit 321 identifies position information of the user based on which medium of transport the user uses to move or which means of transport the user uses to move (FIG. 43). More specifically, when a movement action is recognized from operation action data, the living action recognition unit 321 analyzes how to display position information thereof. In this case, the living action recognition unit 321 first acquires nearest station candidates as an easy-to-use landmark, the last position information and information related thereto. Nearest station candidates can be identified by using the line name, station name, distance to the station and the like. As the last position information and information related thereto, the medium of transport or means of transport, time difference, distance from the last latitude/longitude and the like are acquired.
  • The living action recognition unit 321 assigns weights to the lines and stations using the above information to identify the nearest station. Weights may be assigned to lines and stations by, for example, increasing weights of nearest station candidates with a decreasing distance or assigning weights preferentially to lines and stations that are continuously acquired in action logs. Alternatively, weights may be assigned in consideration of distance differences or time differences that can be acquired from information up to the last time. Accordingly, if the fact of being a predetermined distance apart or that a predetermined time has passed is recognized from the information up to the last time and information this time, the possibility of having changed trains to another line can be considered.
  • 6-2-1. Line Estimation Processing
  • The estimation of line can be determined from, for example, the number of passed stations recognized from an action log. In addition, the movement locus of the user can be estimated by considering the possibility of changing trains at a station identified from position information or whether a direct service between a plurality of lines is available. If a plurality of lines runs between the same stations, which line is used can be identified by estimating a more likely line from the user's past movement locus or acquiring more detailed position information from a position information acquisition sensor.
  • As a result of performing the above line estimation processing, for example, as shown in FIG. 44, movement of the user can be displayed, for example, on a map. Incidentally, the line may be displayed on a map only when the precision of line estimation processing of a predetermined value or more is secured.
  • 6-2-2. Station Name Selection Processing
  • The station name is selected by, as described above, identifying the nearest station. In this case, even if the user does not actually move, changes in latitude/longitude may erroneously be recognized due to an error of radio field intensity of a sensor. Thus, for example, as shown in FIG. 45, the expression may be changed in accordance with the position precision of the station name of the nearest station. If, for example, the Ohsaki station is identified as the nearest station, the expression is changed like “work in the Ohsaki station”, “work near the Ohsaki station”, or “work in Ohsaki” based on the distance difference between the position of the nearest station and the position information. Accordingly, the location where work is done can be expressed more appropriately.
  • If the medium/means of transport is not movement by train, for example, priority may be given to the identified nearest station as a landmark to represent the location of operation by the station name (excluding “station”). For example, it is assumed that, as a result of analyzing an action log, movement by car is recognized and the “Higash-Koganei station” and the “Shin-Koganei station” are identified as landmarks. In this case, it is not natural to move between stations by car and thus, action content can naturally be expressed by representing the start location and the goal as the “Higash-Koganei station” and the “Shin-Koganei station”.
  • 7. Exemplary Hardware Configuration
  • A process of the action recording device 100 in accordance with this embodiment can be executed either by hardware or software. In this case, the action recording device 100 can be configured as shown in FIG. 46. Hereinafter, an exemplary hardware configuration of the action recording device 100 in accordance with this embodiment will be described with reference to FIG. 46.
  • The action recording device 100 in accordance with this embodiment can be implemented by a processing device such as a computer as described above. As shown in FIG. 46, the action recording device 100 includes a CPU (Central Processing Unit) 101, ROM (Read Only Memory) 102, RAM (Random Access Memory) 103, and a host bus 104 a. In addition, the action recording device 100 also includes a bridge 104, an external bus 104 b, an interface 105, an input device 106, an output device 107, a storage device (HDD) 108, a drive 109, a connection port 111, and a communication device 113.
  • The CPU 101 functions as an arithmetic processing unit and a control unit, and controls the entire operation within the action recording device 100 in accordance with various programs. The CPU 101 may also be a microprocessor. The ROM 102 stores programs, operation parameters, and the like used by the CPU 101. The RAM 103 temporarily stores programs used in the execution of the CPU 101, parameters that change as appropriate during the execution, and the like. These units are mutually connected via the host bus 104 a including a CPU bus or the like.
  • The host bus 104 a is connected to the external bus 104 b such as a PCI (Peripheral Component Interconnect/Interface) bus via the bridge 104. Note that the host bus 104 a, the bridge 104, and the external bus 104 b need not necessarily be arranged separately, and the functions of such components may be integrated into a single bus.
  • The input device 106 includes an input means for a user to input information, such as a mouse, a keyboard, a touch panel, a button, a microphone, a switch, or a lever; an input control circuit that generates an input signal on the basis of a user input and outputs the signal to the CPU 101; and the like. The output device 107 includes a display device such as, for example, a liquid crystal display (LCD) device, an OLED (Organic Light Emitting Diode) device, or a lamp; and an audio output device such as a speaker.
  • The storage device 108 is a device for storing data, constructed as an example of a storage unit of the action recording device 100. The storage device 108 can include a storage medium, a recording device that records data on the storage medium, a reading device that reads data from the storage medium, a deletion device that deletes data recorded on the storage medium, and the like. The storage device 108 includes, for example, a HDD (Hard Disk Drive). The storage device 108 stores programs and various data for driving the hard disk and executed by the CPU 101.
  • The drive 109 is a reader/writer for a storage medium, and is incorporated in or externally attached to the action recording device 100. The drive 109 reads information recorded on a removable storage medium such as a magnetic disk, an optical disc, a magnetooptical disk, or semiconductor memory that is mounted, and outputs the information to the RAM 103.
  • The connection port 111 is an interface for connection to an external device, and is, for example, a connection port for connection to an external device that can transmit data via a USB (Universal Serial Bus). The communication device 113 is, for example, a communication interface including a communication device and the like for connection to the communication network 10. The communication device 113 may be any of a communication device supporting a wireless LAN (Local Area Network), a communication device supporting a wireless USB, or a wire communication device that performs wire communication.
  • In the foregoing, a preferred embodiment of the present disclosure has been described in detail with reference to the appended drawings, but the technical scope of the present disclosure is not limited to the above examples. A person skilled in the art may find various alterations and modifications within the scope of the appended claims, and it should be understood that they will naturally come under the technical scope of the present disclosure.
  • In the above embodiment, for example, the action representation generation unit 320 and the data management unit 330 are provided in the analysis server 300 and the action representation generation unit 140 is provided in the action recording device 100, but the present disclosure is not limited to such an example. For example, these functional units may all be provided in the analysis server 300 or in the action recording device 100.
  • Additionally, the present technology may also be configured as below.
  • (1) An information processing device including:
  • an action recognition unit that recognizes an operation action of a user based on sensor information; and
  • an action representation generation unit that analyzes operation action data showing the operation action of the user recognized by the action recognition unit to generate an action segment represented by a meaning and content of the operation action from the operation action data.
  • (2) The information processing device according to (1),
  • wherein dictionary data defining a relation of the higher meaning and content for the operation action is held, and
  • wherein the action representation generation unit estimates the meaning and content of the operation action from relations before and after the operation action data arranged chronologically based on the dictionary data to generate the action segment.
  • (3) The information processing device according to (1) or (2),
  • wherein the action representation generation unit estimates the meaning and content of the operation action in accordance with a time period and a time of the operation action data to be analyzed to generate the action segment.
  • (4) The information processing device according to any one of (1) to (3),
  • wherein the action representation generation unit estimates the meaning and content of the operation action in accordance with positions of the operation action data before and after the operation action data to be analyzed to generate the action segment.
  • (5) The information processing device according to any one of (1) to (4),
  • wherein hierarchical information showing a hierarchical relationship about the meaning and content is attached to the action segment.
  • (6) The information processing device according to (5),
  • wherein the action representation generation unit displays the action segments based on a segmentation grain size deciding roughness of segmentation of the action segments and the hierarchical information.
  • (7) The information processing device according to (6),
  • wherein the action representation generation unit combines or divides the action segments based on a size of the segmentation grain size and displays the combined or divided action segments.
  • (8) The information processing device according to any one of (1) to (7), further including:
  • a typical action pattern generation unit that extracts one action pattern from a plurality of action segment groups including the action segments of a predetermined unit based on a correlation between the action segments.
  • (9) The information processing device according to any one of (1) to (8),
  • wherein the action representation generation unit displays, in a display unit, the action segments represented at least by a start time, an end time, position information, and operation content of the operation action by chronologically arranging the action segments.
  • (10) The information processing device according to any one of (1) to (9), further including:
  • a feedback adjustment unit that corrects operation action estimation information that decides the operation action based on correction feedback from the user to the action segment generated by the action representation generation unit,
  • wherein the action representation generation unit generates the action segment constituting an action log from the operation action data based on the operation action estimation data, and
  • wherein the feedback adjustment unit corrects the operation action estimation information based on the correction feedback.
  • (11) An information processing device including:
  • an action recognition unit that recognizes an operation action of a user based on sensor information;
  • an action representation generation unit that generates an action segment constituting an action log from operation action data showing the operation action of the user recognized by the action recognition unit based on operation action estimation information that decides the operation action; and
  • a feedback adjustment unit that corrects the operation action estimation information based on correction feedback from the user to the action segment generated by the action representation generation unit.
  • (12) The information processing device according to (11),
  • wherein the operation action estimation information includes a plurality of combinations of a probability distribution and a weighting factor depending on a location for a plurality of the operation actions, and
  • wherein the feedback adjustment unit corrects the probability distribution or the weighting factor for each of the operation actions based on the correction feedback.
  • (13) The information processing device according to (12),
  • wherein, when the correction feedback concerns action content, the feedback adjustment unit corrects the probability distribution of the operation action estimation information in accordance with content of the correction feedback.
  • (14) The information processing device according to (11) or (12),
  • wherein, when the correction feedback concerns the location, the feedback adjustment unit corrects the weighting factor of the operation action estimation information in accordance with content of the correction feedback.
  • REFERENCE SIGN LIST
    • 100 action recording device
    • 110 sensors
    • 120 action recognition unit
    • 122 sensor controller
    • 124 operation action recognition unit
    • 130 client interface unit
    • 140 action representation processing unit
    • 142 display processing unit
    • 144 input information processing unit
    • 150 display unit
    • 160 input unit
    • 200 action log server
    • 210 log server interface unit
    • 220 action log db
    • 300 analysis server
    • 310 analysis server interface unit
    • 320 action representation generation unit
    • 321 living action recognition unit
    • 322 hierarchical structure judgment unit
    • 330 data management unit
    • 331 data acquisition unit
    • 332 feedback adjustment unit
    • 333 analysis parameter db
    • 334 unit data storage db
    • 335 hierarchical information attached data storage db

Claims (16)

1. (canceled)
2. A non-transitory storage medium configured to store a computer program for causing a computer to function as an information processing device, the program comprising:
managing an action segment based on an operation action of a user by referring to sensor information and an action log including at least the action segment;
relating time information, location information and a feeling for the action segment; and
causing a display to display a button to store, related respectively, the time information, the location information and the feeling for the action segment.
3. The non-transitory storage medium according to claim 2, wherein the sensor information includes at least one of the time information or position information.
4. The non-transitory storage medium according to claim 3, wherein the position information is obtained from at least one of a Global Positioning System (GPS), WiFi or another communication network.
5. The non-transitory storage medium according to claim 2, wherein the location information is selectable from a location name list.
6. The non-transitory storage medium according to claim 5, wherein the location name list includes at least one location name to be a landmark including a building name, station name or shop name.
7. The non-transitory storage medium according to claim 2, wherein the feeling is selectable from a feeling selection area.
8. The non-transitory storage medium according to claim 2, wherein the time information, the location information and the feeling for the action segment related respectively are at a time caused to be displayed for the action segment.
9. The non-transitory storage medium according to claim 2, the program further comprising:
causing the display to display a map for the action segment upon user action.
10. The non-transitory storage medium according to claim 2, the program further comprising:
non-display filtering to prevent a portion of the action segment from being displayed by user setting.
11. The non-transitory storage medium according to claim 2, the program further comprising:
non-display filtering to prevent a portion of the action segment from being displayed when a specific action or location is set by the user.
12. The non-transitory storage medium according to claim 2, the program further comprising:
causing the display to display to delete the action segment by the user.
13. The non-transitory storage medium according to claim 2, the program further comprising:
causing the display to display an action icon indicating action taken by the user.
14. The non-transitory storage medium according to claim 13, wherein the action icon is displayed based on user selection from an action candidate list.
15. An information processing apparatus comprising:
circuitry configured to control:
managing an action segment based on an operation action of a user by referring to sensor information and an action log including at least the action segment;
relating time information, location information and a feeling for the action segment; and
causing a display to display a button to store, related respectively, the time information, the location information and the feeling for the action segment.
16. An information processing method comprising:
managing an action segment based on an operation action of a user by referring to sensor information and an action log including at least the action segment;
relating time information, location information and a feeling for the action segment; and
causing a display to display a button to store, related respectively, the time information, the location information and the feeling for the action segment.
US15/173,793 2011-06-13 2016-06-06 Information processing device, information processing method, and computer program Abandoned US20160283579A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US15/173,793 US20160283579A1 (en) 2011-06-13 2016-06-06 Information processing device, information processing method, and computer program
US15/254,482 US10740057B2 (en) 2011-06-13 2016-09-01 Information processing device, information processing method, and computer program

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
JP2011131130A JP6031735B2 (en) 2011-06-13 2011-06-13 Information processing apparatus, information processing method, and computer program
JP2011-131130 2011-06-13
PCT/JP2012/064564 WO2012173027A1 (en) 2011-06-13 2012-06-06 Information processing device, information processing method, and computer program
US201314123886A 2013-12-04 2013-12-04
US15/173,793 US20160283579A1 (en) 2011-06-13 2016-06-06 Information processing device, information processing method, and computer program

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
PCT/JP2012/064564 Continuation WO2012173027A1 (en) 2011-06-13 2012-06-06 Information processing device, information processing method, and computer program
US14/123,886 Continuation US20140101169A1 (en) 2011-06-13 2012-06-06 Information processing device, information processing method, and computer program

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/254,482 Continuation US10740057B2 (en) 2011-06-13 2016-09-01 Information processing device, information processing method, and computer program

Publications (1)

Publication Number Publication Date
US20160283579A1 true US20160283579A1 (en) 2016-09-29

Family

ID=47357016

Family Applications (5)

Application Number Title Priority Date Filing Date
US14/123,886 Abandoned US20140101169A1 (en) 2011-06-13 2012-06-06 Information processing device, information processing method, and computer program
US14/878,065 Abandoned US20160026349A1 (en) 2011-06-13 2015-10-08 Information processing device, information processing method, and computer program
US14/992,142 Abandoned US20160170572A1 (en) 2011-06-13 2016-01-11 Information processing device, information processing method, and computer program
US15/173,793 Abandoned US20160283579A1 (en) 2011-06-13 2016-06-06 Information processing device, information processing method, and computer program
US15/254,482 Active US10740057B2 (en) 2011-06-13 2016-09-01 Information processing device, information processing method, and computer program

Family Applications Before (3)

Application Number Title Priority Date Filing Date
US14/123,886 Abandoned US20140101169A1 (en) 2011-06-13 2012-06-06 Information processing device, information processing method, and computer program
US14/878,065 Abandoned US20160026349A1 (en) 2011-06-13 2015-10-08 Information processing device, information processing method, and computer program
US14/992,142 Abandoned US20160170572A1 (en) 2011-06-13 2016-01-11 Information processing device, information processing method, and computer program

Family Applications After (1)

Application Number Title Priority Date Filing Date
US15/254,482 Active US10740057B2 (en) 2011-06-13 2016-09-01 Information processing device, information processing method, and computer program

Country Status (6)

Country Link
US (5) US20140101169A1 (en)
EP (1) EP2720176A4 (en)
JP (1) JP6031735B2 (en)
CN (5) CN106126556B (en)
BR (1) BR112013031580A2 (en)
WO (1) WO2012173027A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170279907A1 (en) * 2016-03-24 2017-09-28 Casio Computer Co., Ltd. Behavior Analysis Apparatus for Analysis of User's Behavior, Behavior Analysis Method, and Storage Medium

Families Citing this family (58)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5953673B2 (en) * 2011-08-11 2016-07-20 日本電気株式会社 Action identification device, action identification method, and program
KR101887422B1 (en) * 2012-11-19 2018-09-10 삼성전자주식회사 Apparatas and method for displaying a location information of established device in an electronic device
US9336295B2 (en) * 2012-12-03 2016-05-10 Qualcomm Incorporated Fusing contextual inferences semantically
CN104516660B (en) * 2013-09-27 2019-03-08 联想(北京)有限公司 A kind of information processing method, system and electronic equipment
WO2015068440A1 (en) 2013-11-08 2015-05-14 ソニー株式会社 Information processing apparatus, control method, and program
WO2015178066A1 (en) * 2014-05-22 2015-11-26 ソニー株式会社 Information processing device and information processing method
JP6572886B2 (en) 2014-05-22 2019-09-11 ソニー株式会社 Information processing apparatus and information processing method
US20150347912A1 (en) * 2014-05-27 2015-12-03 Sony Corporation Activity tracking based recommendation
WO2015194270A1 (en) * 2014-06-20 2015-12-23 ソニー株式会社 Information-processing device, information processing method, and program
EP3147747A1 (en) 2014-06-27 2017-03-29 Apple Inc. Manipulation of calendar application in device with touch screen
EP3195098B1 (en) 2014-07-21 2024-10-23 Apple Inc. Remote user interface
KR102319896B1 (en) 2014-08-02 2021-11-02 애플 인크. Context-specific user interfaces
US10452253B2 (en) 2014-08-15 2019-10-22 Apple Inc. Weather user interface
CN115695632B (en) 2014-09-02 2024-10-01 苹果公司 Electronic device, computer storage medium, and method of operating electronic device
JPWO2016092924A1 (en) * 2014-12-09 2017-09-14 ソニー株式会社 Information processing apparatus, control method, and program
US10571289B2 (en) 2015-01-05 2020-02-25 Sony Corporation Information processing device, information processing method, and program
JP6603023B2 (en) * 2015-02-09 2019-11-06 東芝ライフスタイル株式会社 Information provision system
WO2016144385A1 (en) 2015-03-08 2016-09-15 Apple Inc. Sharing user-configurable graphical constructs
JP6791120B2 (en) 2015-03-12 2020-11-25 ソニー株式会社 Information processing equipment, information processing methods, and programs
US9774693B2 (en) 2015-04-29 2017-09-26 Facebook, Inc. Methods and systems for viewing user feedback
CN104853249A (en) * 2015-05-12 2015-08-19 天脉聚源(北京)传媒科技有限公司 Program recommendation method and terminal
JP6302954B2 (en) * 2015-05-14 2018-03-28 株式会社アドダイス Management system and management method
US9565521B1 (en) 2015-08-14 2017-02-07 Samsung Electronics Co., Ltd. Automatic semantic labeling based on activity recognition
EP3337583B1 (en) 2015-08-20 2024-01-17 Apple Inc. Exercise-based watch face
JP6577324B2 (en) * 2015-10-09 2019-09-18 株式会社東芝 Action determination device and action determination method
WO2017081829A1 (en) * 2015-11-13 2017-05-18 富士通株式会社 Behavior detection device, behavior detection method, and behavior detection program
WO2017126320A1 (en) 2016-01-19 2017-07-27 ソニー株式会社 Information processing device, information processing method, and program
JP6199421B2 (en) * 2016-02-17 2017-09-20 ソフトバンク株式会社 Display control apparatus and program
DK201770423A1 (en) 2016-06-11 2018-01-15 Apple Inc Activity and workout updates
JP6749164B2 (en) * 2016-07-11 2020-09-02 株式会社エヌ・ティ・ティ・データ Service providing system and program
CN106776385B (en) * 2016-12-06 2019-09-24 Oppo广东移动通信有限公司 A kind of transmission method, device and the terminal of log log information
JP2018132923A (en) * 2017-02-15 2018-08-23 コニカミノルタ株式会社 Information processing program and information processing method
JP6317000B1 (en) * 2017-02-20 2018-04-25 三菱電機インフォメーションシステムズ株式会社 Labeling apparatus and program
JP7143839B2 (en) * 2017-03-08 2022-09-29 ソニーグループ株式会社 Information processing device, information processing method, and program
EP3606055A4 (en) * 2017-03-31 2020-02-26 Nec Corporation Video processing device, video analysis system, method, and program
DK179412B1 (en) 2017-05-12 2018-06-06 Apple Inc Context-Specific User Interfaces
KR102056696B1 (en) 2017-11-09 2019-12-17 숭실대학교 산학협력단 Terminal device for generating user behavior data, Method for generating user behavior data and recording medium
CN108151732A (en) * 2017-12-22 2018-06-12 浙江西湖高等研究院 A kind of long-range position and behavior method of estimation
US11327650B2 (en) 2018-05-07 2022-05-10 Apple Inc. User interfaces having a collection of complications
CN109189218B (en) * 2018-08-20 2019-05-10 广州市三川田文化科技股份有限公司 A kind of method, apparatus of gesture identification, equipment and computer readable storage medium
CN109241900B (en) * 2018-08-30 2021-04-09 Oppo广东移动通信有限公司 Wearable device control method and device, storage medium and wearable device
AU2020239670B2 (en) 2019-05-06 2021-07-15 Apple Inc. Restricted operation of an electronic device
US11960701B2 (en) 2019-05-06 2024-04-16 Apple Inc. Using an illustration to show the passing of time
US11131967B2 (en) 2019-05-06 2021-09-28 Apple Inc. Clock faces for an electronic device
US10852905B1 (en) 2019-09-09 2020-12-01 Apple Inc. Techniques for managing display usage
DK202070624A1 (en) 2020-05-11 2022-01-04 Apple Inc User interfaces related to time
CN115552375A (en) 2020-05-11 2022-12-30 苹果公司 User interface for managing user interface sharing
US11372659B2 (en) 2020-05-11 2022-06-28 Apple Inc. User interfaces for managing user interface sharing
JP2022011533A (en) * 2020-06-30 2022-01-17 株式会社日立製作所 Unique expression classifier and unique expression classification method
US11989254B2 (en) * 2020-09-10 2024-05-21 Taboola.Com Ltd. Semantic meaning association to components of digital content
CN112202632A (en) * 2020-09-24 2021-01-08 建信金融科技有限责任公司 Message extraction method and device, electronic equipment and computer readable storage medium
US11694590B2 (en) 2020-12-21 2023-07-04 Apple Inc. Dynamic user interface with time indicator
US11720239B2 (en) 2021-01-07 2023-08-08 Apple Inc. Techniques for user interfaces related to an event
JP2022109646A (en) * 2021-01-15 2022-07-28 オムロン株式会社 Operation level conversion device, operation level conversion method, and operation level conversion program
JP2022151232A (en) * 2021-03-26 2022-10-07 富士フイルムビジネスイノベーション株式会社 Information processing apparatus and program
US11921992B2 (en) 2021-05-14 2024-03-05 Apple Inc. User interfaces related to time
US20230236547A1 (en) 2022-01-24 2023-07-27 Apple Inc. User interfaces for indicating time
JP7481060B1 (en) 2023-12-27 2024-05-10 エグゼヴィータ株式会社 BEHAVIOR ACQUISITION DEVICE, BEHAVIOR ACQUISITION METHOD, AND PROGRAM

Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS468858Y1 (en) * 1970-06-19 1971-03-29
US20040098462A1 (en) * 2000-03-16 2004-05-20 Horvitz Eric J. Positioning and rendering notification heralds based on user's focus of attention and activity
US20070239610A1 (en) * 2007-05-17 2007-10-11 Floga Holdings, Llc Methods, systems and apparatus for displaying user generated tracking information
US7340393B2 (en) * 2000-09-13 2008-03-04 Advanced Generation Interface, Inc. Emotion recognizing method, sensibility creating method, device, and software
JP2008117152A (en) * 2006-11-02 2008-05-22 Sharp Corp History information display device and method
US20090300525A1 (en) * 2008-05-27 2009-12-03 Jolliff Maria Elena Romera Method and system for automatically updating avatar to indicate user's status
US20100198812A1 (en) * 2009-02-02 2010-08-05 Yahoo! Inc. Automated search
US20100281156A1 (en) * 2009-05-04 2010-11-04 Kies Jonathan K System and method of recording and sharing mobile application activities
US20100318293A1 (en) * 2009-06-12 2010-12-16 Microsoft Corporation Retracing steps
KR20110003849A (en) * 2009-07-06 2011-01-13 엘지전자 주식회사 Mobile terminal and method for visualizing life log
US20110137836A1 (en) * 2008-09-19 2011-06-09 Hiroyuki Kuriyama Method and system for generating history of behavior
US20110246926A1 (en) * 2010-04-02 2011-10-06 Irunurun, Llc Comparative action management
US20110264691A1 (en) * 2010-04-26 2011-10-27 Migita Takahito Information processing apparatus, text selection method, and program
US20120054675A1 (en) * 2010-08-26 2012-03-01 Unisys Corporation Graphical user interface system for a log analyzer
US20120052880A1 (en) * 2010-08-27 2012-03-01 Research In Motion Limited System and method for determining action spot locations relative to the location of a mobile device
US20120150777A1 (en) * 2010-12-13 2012-06-14 Kabushiki Kaisha Toshiba Action history search device
US20120229505A1 (en) * 2009-11-26 2012-09-13 JVC Kenwood Corporation Information display apparatus, and method and program for information display control
US20120251079A1 (en) * 2010-11-10 2012-10-04 Nike, Inc. Systems and Methods for Time-Based Athletic Activity Measurement and Display
US20120253942A1 (en) * 2011-04-04 2012-10-04 Democracyontheweb, Llc Providing content to users

Family Cites Families (112)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0369525A (en) 1989-08-08 1991-03-25 Nippon Electric Glass Co Ltd Method for cutting glass thin tube
JP3472026B2 (en) * 1996-03-26 2003-12-02 富士通株式会社 Log information collection analyzer
JP3570163B2 (en) 1996-07-03 2004-09-29 株式会社日立製作所 Method and apparatus and system for recognizing actions and actions
JP3252721B2 (en) * 1996-10-07 2002-02-04 カシオ計算機株式会社 Behavior analysis device
US6266295B1 (en) * 1998-01-07 2001-07-24 Microsoft Corporation System and method of displaying times corresponding to events on a calendar
CA2356846A1 (en) * 1998-10-06 2000-04-13 Kun Yu Generalized multi-interfaced extensible content management and delivery system, and on-line calendar
JP2001357309A (en) 2000-06-14 2001-12-26 Nec Corp Diary type taste information collection method
US7149970B1 (en) * 2000-06-23 2006-12-12 Microsoft Corporation Method and system for filtering and selecting from a candidate list generated by a stochastic input method
US7895268B2 (en) * 2000-11-01 2011-02-22 International Business Machines Corporation Multiple perspective activity map display
US6751626B2 (en) * 2001-05-03 2004-06-15 International Business Machines Corporation Method, system, and program for mining data in a personal information manager database
JP2003030375A (en) * 2001-07-12 2003-01-31 Casio Comput Co Ltd System, device, and method for action history generation
WO2003019402A1 (en) * 2001-08-22 2003-03-06 Koninklijke Philips Electronics N.V. Timeline display apparatus
EP1422501A1 (en) * 2001-08-31 2004-05-26 Aisin Aw Co., Ltd. Information display system
US20030154116A1 (en) * 2002-01-04 2003-08-14 Lofton Jeffrey C. System and method for scheduling events on an internet based calendar
US6938219B2 (en) * 2002-01-25 2005-08-30 International Business Machines Corporation Method and system for displaying actions and historical content in an output view
US7441194B2 (en) * 2002-02-20 2008-10-21 Microsoft Corporation Calendar-based user interface system
JP2003296782A (en) 2002-03-29 2003-10-17 Casio Comput Co Ltd Device and program for recording action
JP2004078420A (en) 2002-08-13 2004-03-11 Toshiba Tec Corp Information input form management device and information input form management program
US7274375B1 (en) * 2002-11-19 2007-09-25 Peter David Timekeeping system and method for graphically tracking and representing activities
JP3915720B2 (en) 2002-11-20 2007-05-16 ソニー株式会社 Video production system, video production device, video production method
JP2004318503A (en) 2003-04-16 2004-11-11 Toshiba Corp Device, method and program for supporting action management
US20040225969A1 (en) * 2003-05-09 2004-11-11 Nokia Corporation Hand-portable electronic device with symbolic calendar application
JP4608858B2 (en) * 2003-08-12 2011-01-12 ソニー株式会社 Emotion visualization device, emotion visualization method, and emotion visualization output
JP2005124336A (en) * 2003-10-17 2005-05-12 Yaskawa Electric Corp Control method for ac motor and control unit
JP4481663B2 (en) 2004-01-15 2010-06-16 キヤノン株式会社 Motion recognition device, motion recognition method, device control device, and computer program
JP4507243B2 (en) * 2004-03-25 2010-07-21 独立行政法人理化学研究所 Behavior analysis method and system
EP1618844B1 (en) * 2004-07-21 2011-04-27 Panasonic Electric Works Co., Ltd. Physical activity measuring system
US20060026499A1 (en) * 2004-07-28 2006-02-02 Corey Weddle Calendar based graphical user interface for manipulation of spatial operations and information
JP4199713B2 (en) 2004-08-26 2008-12-17 学校法人 芝浦工業大学 Position related information providing apparatus, method and program thereof
JP2006081129A (en) 2004-09-13 2006-03-23 Olympus Corp Image display device, image display program, and recording medium recording image display program and readable by computer
JP5112613B2 (en) * 2004-10-15 2013-01-09 エスアーペー アーゲー Activity management system and method, activity management apparatus, client terminal, and computer program
US7254516B2 (en) * 2004-12-17 2007-08-07 Nike, Inc. Multi-sensor monitoring of athletic performance
JP5028751B2 (en) 2005-06-09 2012-09-19 ソニー株式会社 Action recognition device
CA2615659A1 (en) * 2005-07-22 2007-05-10 Yogesh Chunilal Rathod Universal knowledge management and desktop search system
JP4832842B2 (en) * 2005-09-28 2011-12-07 ヤフー株式会社 Map movement trajectory management method and map movement trajectory management apparatus
US7587252B2 (en) * 2005-10-25 2009-09-08 Fisher-Rosemount Systems, Inc. Non-periodic control communications in wireless and other process control systems
US20070106931A1 (en) * 2005-11-08 2007-05-10 Nokia Corporation Active notes application
US7283927B2 (en) * 2005-12-07 2007-10-16 Katrina Delargy Activity recording module
US8139030B2 (en) * 2006-02-01 2012-03-20 Memsic, Inc. Magnetic sensor for use with hand-held devices
KR100773056B1 (en) * 2006-02-28 2007-11-02 삼성전자주식회사 Method and apparatus for offering user interface in a mobile station
JP4706850B2 (en) 2006-03-23 2011-06-22 富士フイルム株式会社 Nozzle plate manufacturing method, droplet discharge head, and image forming apparatus
KR100714716B1 (en) * 2006-04-10 2007-05-04 삼성전자주식회사 Apparatus and method for displaying events
JP4137952B2 (en) * 2006-05-01 2008-08-20 ソニー・エリクソン・モバイルコミュニケーションズ株式会社 Information processing device, information processing method, information processing program, and portable terminal device
WO2007130039A1 (en) * 2006-05-04 2007-11-15 Schlumberger Technology Corporation Method system and program storage device for synchronizing displays relative to a point in time
EP2021731A4 (en) * 2006-05-08 2010-07-21 Telecomm Systems Inc Location input mistake correction
US8572486B2 (en) * 2006-05-16 2013-10-29 Blackberry Limited System and method for integrated presentation of information
JP5041202B2 (en) 2006-06-20 2012-10-03 ソニー株式会社 Information processing apparatus, information processing method, and program
US20070299631A1 (en) * 2006-06-27 2007-12-27 Microsoft Corporation Logging user actions within activity context
US20070300185A1 (en) * 2006-06-27 2007-12-27 Microsoft Corporation Activity-centric adaptive user interface
DE102006030929A1 (en) * 2006-07-05 2008-01-10 Stabilus Gmbh Piston / Cylinder Unit with Diagnostic Unit
WO2008050711A1 (en) * 2006-10-26 2008-05-02 Panasonic Corporation Mobile history editing device
JP2008146248A (en) 2006-12-07 2008-06-26 Nippon Telegraph & Telephone West Corp Probe data analysis system
US8165405B2 (en) 2006-12-18 2012-04-24 Honda Motor Co., Ltd. Leveraging temporal, contextual and ordering constraints for recognizing complex activities in video
KR100866487B1 (en) * 2007-01-03 2008-11-03 삼성전자주식회사 Apparatus and method for tracking action history of user
US20080165151A1 (en) * 2007-01-07 2008-07-10 Lemay Stephen O System and Method for Viewing and Managing Calendar Entries
JP5110261B2 (en) 2007-03-27 2012-12-26 日本電気株式会社 Homepage providing method and system
US8250474B2 (en) * 2007-03-31 2012-08-21 Topix Llc Chronology display and feature for online presentations and web pages
JP4992043B2 (en) 2007-08-13 2012-08-08 株式会社国際電気通信基礎技術研究所 Action identification device, action identification system, and action identification method
KR101472842B1 (en) * 2007-09-11 2014-12-16 삼성전자 주식회사 Apparatus and method for recognizing moving signal
JP5235370B2 (en) 2007-09-20 2013-07-10 インターナショナル・ビジネス・マシーンズ・コーポレーション Apparatus, method and program for managing and displaying schedule
JP2011503639A (en) * 2007-11-06 2011-01-27 テレ アトラス ノース アメリカ インコーポレイテッド Method and system for using probe data from multiple vehicles to detect real-world changes used in map updates
US8310542B2 (en) * 2007-11-28 2012-11-13 Fuji Xerox Co., Ltd. Segmenting time based on the geographic distribution of activity in sensor data
JP5004777B2 (en) * 2007-12-06 2012-08-22 インターナショナル・ビジネス・マシーンズ・コーポレーション Efficient action history recording and playback method in virtual space
WO2009099128A1 (en) 2008-02-07 2009-08-13 Nec Corporation Information management device, information management method, and information management program
US8055556B2 (en) 2008-03-12 2011-11-08 Intuit Inc. User interface for rapid entry of timesheet data
US9495275B2 (en) * 2008-04-29 2016-11-15 International Business Machines Corporation System and computer program product for deriving intelligence from activity logs
CN101282560A (en) * 2008-05-29 2008-10-08 苏琨 Remote distributed individual dayotday information management service issuance platform based on server-client mode
JP5345157B2 (en) * 2008-06-13 2013-11-20 ジーブイビービー ホールディングス エス.エイ.アール.エル. Log information display method and apparatus
US20110106736A1 (en) * 2008-06-26 2011-05-05 Intuitive User Interfaces Ltd. System and method for intuitive user interaction
JP2010029633A (en) * 2008-06-30 2010-02-12 Tokai Rubber Ind Ltd Method of detecting muscle movement and device of detecting muscle movement
JP2010016444A (en) * 2008-07-01 2010-01-21 Toshiba Corp Situation recognizing apparatus, situation recognizing method, and radio terminal apparatus
EP2098966B1 (en) * 2008-08-04 2011-07-13 Agilent Technologies, Inc. Spatial correlation between display of data and display of applicable processing actions
WO2010029845A1 (en) * 2008-09-11 2010-03-18 日本電気株式会社 Activity data presenting apparatus, activity data presenting method and recording medium
WO2010047337A1 (en) 2008-10-20 2010-04-29 株式会社キャメロット Information processing device operation control system and operation control method
JP5551867B2 (en) 2008-12-05 2014-07-16 ソニー株式会社 Information processing apparatus and information processing method
US8442832B2 (en) * 2008-12-08 2013-05-14 Electronics And Telecommunications Research Institute Apparatus for context awareness and method using the same
EP2197155A1 (en) * 2008-12-12 2010-06-16 Alcatel, Lucent A module and associated method for path information
JP2010146223A (en) * 2008-12-18 2010-07-01 Hitachi Ltd Behavior extraction system, behavior extraction method, and server
JP5372487B2 (en) * 2008-12-18 2013-12-18 株式会社日立製作所 Action record input support system and server
JP5118620B2 (en) * 2008-12-24 2013-01-16 日立Geニュークリア・エナジー株式会社 Dynamic recognition device, dynamic recognition system, and dynamic recognition method
CN102365296A (en) * 2009-01-26 2012-02-29 根马布股份公司 Methods for producing mixtures of antibodies
JP5649808B2 (en) * 2009-01-28 2015-01-07 ソニー株式会社 Information processing apparatus, information processing method, and program
JP2010267207A (en) * 2009-05-18 2010-11-25 Nippon Telegr & Teleph Corp <Ntt> Unusualness evaluation device, unusualness evaluation method, and computer program
US8195393B2 (en) * 2009-06-30 2012-06-05 Apple Inc. Analyzing and consolidating track file data
US20110022443A1 (en) * 2009-07-21 2011-01-27 Palo Alto Research Center Incorporated Employment inference from mobile device data
CN101964047B (en) * 2009-07-22 2012-10-10 深圳泰山在线科技有限公司 Multiple trace point-based human body action recognition method
JP5476833B2 (en) * 2009-07-23 2014-04-23 カシオ計算機株式会社 Animal emotion display system and animal emotion display method
CN101996311A (en) * 2009-08-10 2011-03-30 深圳泰山在线科技有限公司 Yoga stance recognition method and system
JP5440080B2 (en) * 2009-10-02 2014-03-12 ソニー株式会社 Action pattern analysis system, portable terminal, action pattern analysis method, and program
US20110099500A1 (en) * 2009-10-27 2011-04-28 Jared Smith Historical network event viewing
JP5010665B2 (en) 2009-11-02 2012-08-29 株式会社東芝 Sales support system
US20110161085A1 (en) * 2009-12-31 2011-06-30 Nokia Corporation Method and apparatus for audio summary of activity for user
US8180778B1 (en) * 2010-02-05 2012-05-15 Google Inc. Generating action trails from web history
US9386138B2 (en) * 2010-02-15 2016-07-05 Lenovo Innovations Limited (Hong Kong) Mobile terminal device, operation procedure communication system, and operation communication method
CN102792330A (en) * 2010-03-16 2012-11-21 日本电气株式会社 Interest level measurement system, interest level measurement device, interest level measurement method, and interest level measurement program
JP2011252844A (en) * 2010-06-03 2011-12-15 Sony Corp Data processing device, data processing method and program
WO2012008264A1 (en) 2010-07-16 2012-01-19 株式会社日立製作所 Activity visualization system, server, and activity data processing method
US20120022950A1 (en) * 2010-07-26 2012-01-26 At&T Intellectual Property I, L.P. Systems and Methods for Targeted Advertising in Voicemail to Text Systems
US8610766B2 (en) * 2010-09-24 2013-12-17 International Business Machines Corporation Activity determination as function of transaction log
US8738323B2 (en) * 2010-09-30 2014-05-27 Fitbit, Inc. Methods and systems for metrics analysis and interactive rendering, including events having combined activity and location information
CN102088473A (en) * 2010-11-18 2011-06-08 吉林禹硕动漫游戏科技股份有限公司 Implementation method of multi-user mobile interaction
CN102096803B (en) * 2010-11-29 2013-11-13 吉林大学 Safe state recognition system for people on basis of machine vision
KR101511118B1 (en) * 2010-12-20 2015-04-13 에스케이플래닛 주식회사 Apparatus and method for displaying split screen
US20120166285A1 (en) * 2010-12-28 2012-06-28 Scott Shapiro Defining and Verifying the Accuracy of Explicit Target Clusters in a Social Networking System
US20120209839A1 (en) * 2011-02-15 2012-08-16 Microsoft Corporation Providing applications with personalized and contextually relevant content
JP5418520B2 (en) * 2011-02-16 2014-02-19 カシオ計算機株式会社 Location information acquisition device, location information acquisition method, and program
US9262065B2 (en) * 2011-02-24 2016-02-16 Red Hat, Inc. Time based touch screen input recognition
KR20120126161A (en) * 2011-05-11 2012-11-21 삼성전자주식회사 Mobile terminal and method for controlling screen using the same
US9236045B2 (en) * 2011-05-23 2016-01-12 Nuance Communications, Inc. Methods and apparatus for proofing of a text input
US8726142B2 (en) * 2011-09-21 2014-05-13 Facebook, Inc. Selecting social networking system user information for display via a timeline interface
US9773284B2 (en) * 2011-09-21 2017-09-26 Facebook, Inc. Displaying social networking system user information via a map interface
US8749690B2 (en) * 2011-12-13 2014-06-10 Facebook, Inc. In-context content capture

Patent Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS468858Y1 (en) * 1970-06-19 1971-03-29
US20040098462A1 (en) * 2000-03-16 2004-05-20 Horvitz Eric J. Positioning and rendering notification heralds based on user's focus of attention and activity
US7340393B2 (en) * 2000-09-13 2008-03-04 Advanced Generation Interface, Inc. Emotion recognizing method, sensibility creating method, device, and software
JP2008117152A (en) * 2006-11-02 2008-05-22 Sharp Corp History information display device and method
US20070239610A1 (en) * 2007-05-17 2007-10-11 Floga Holdings, Llc Methods, systems and apparatus for displaying user generated tracking information
US20090300525A1 (en) * 2008-05-27 2009-12-03 Jolliff Maria Elena Romera Method and system for automatically updating avatar to indicate user's status
US20110137836A1 (en) * 2008-09-19 2011-06-09 Hiroyuki Kuriyama Method and system for generating history of behavior
US20100198812A1 (en) * 2009-02-02 2010-08-05 Yahoo! Inc. Automated search
US20100281156A1 (en) * 2009-05-04 2010-11-04 Kies Jonathan K System and method of recording and sharing mobile application activities
US20100318293A1 (en) * 2009-06-12 2010-12-16 Microsoft Corporation Retracing steps
KR20110003849A (en) * 2009-07-06 2011-01-13 엘지전자 주식회사 Mobile terminal and method for visualizing life log
US20120229505A1 (en) * 2009-11-26 2012-09-13 JVC Kenwood Corporation Information display apparatus, and method and program for information display control
US20110246926A1 (en) * 2010-04-02 2011-10-06 Irunurun, Llc Comparative action management
US20110264691A1 (en) * 2010-04-26 2011-10-27 Migita Takahito Information processing apparatus, text selection method, and program
US20120054675A1 (en) * 2010-08-26 2012-03-01 Unisys Corporation Graphical user interface system for a log analyzer
US20120052880A1 (en) * 2010-08-27 2012-03-01 Research In Motion Limited System and method for determining action spot locations relative to the location of a mobile device
US20120251079A1 (en) * 2010-11-10 2012-10-04 Nike, Inc. Systems and Methods for Time-Based Athletic Activity Measurement and Display
US20120150777A1 (en) * 2010-12-13 2012-06-14 Kabushiki Kaisha Toshiba Action history search device
US20120253942A1 (en) * 2011-04-04 2012-10-04 Democracyontheweb, Llc Providing content to users

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170279907A1 (en) * 2016-03-24 2017-09-28 Casio Computer Co., Ltd. Behavior Analysis Apparatus for Analysis of User's Behavior, Behavior Analysis Method, and Storage Medium

Also Published As

Publication number Publication date
JP6031735B2 (en) 2016-11-24
EP2720176A1 (en) 2014-04-16
US10740057B2 (en) 2020-08-11
CN106096001A (en) 2016-11-09
JP2013003649A (en) 2013-01-07
WO2012173027A1 (en) 2012-12-20
CN103597476A (en) 2014-02-19
CN106202528B (en) 2021-08-24
CN106406652B (en) 2020-03-17
CN106126556A (en) 2016-11-16
US20140101169A1 (en) 2014-04-10
CN106406652A (en) 2017-02-15
CN103597476B (en) 2017-10-13
US20160170572A1 (en) 2016-06-16
EP2720176A4 (en) 2014-12-10
CN106126556B (en) 2020-09-29
CN106096001B (en) 2020-09-29
BR112013031580A2 (en) 2017-07-04
US20160026349A1 (en) 2016-01-28
US20160371044A1 (en) 2016-12-22
CN106202528A (en) 2016-12-07

Similar Documents

Publication Publication Date Title
US10740057B2 (en) Information processing device, information processing method, and computer program
US11550795B2 (en) Dynamic productivity content rendering based upon user interaction patterns
CN110476176B (en) User objective assistance techniques
US10062415B2 (en) Synchronizing audio and video components of an automatically generated audio/video presentation
CN107077502B (en) Context-based search and relevance generation
CN103403705B (en) Loading a mobile computing device with media files
KR102276766B1 (en) Systems and methods for generating and providing intelligent time to leave reminders
CN103488666B (en) Information processing equipment and method, electronic device and computer readable storage medium
US9877162B2 (en) Systems and methods for generating a user location history
JP2020537198A (en) Identify music as a particular song
CN108700952A (en) Text input is predicted based on user demographic information and contextual information
CN102939515A (en) Device, method, and graphical user interface for mapping directions between search results
KR20170066415A (en) Video editing using contextual data and content discovery using clusters
US20190050814A1 (en) Generation of user profile from source code
CN107430483A (en) Navigation event information
CN105518608A (en) Context-sensitive gesture classification
CN110337660B (en) Context application organizer framework for user life events
US20130035929A1 (en) Information processing apparatus and method
KR20150027885A (en) Operating Method for Electronic Handwriting and Electronic Device supporting the same
JP6160670B2 (en) Information processing apparatus, information processing method, and computer program
JP2011164936A (en) Work time accumulating device, work time accumulating method, and work time accumulation program
CN104272770A (en) Automated retrieval of physical location information
JP2021163238A (en) Recommendation system and recommendation method
JP6265245B2 (en) Information processing apparatus, information processing method, and computer program
JP6420278B2 (en) Computer program, information processing apparatus and information processing method

Legal Events

Date Code Title Description
AS Assignment

Owner name: SONY CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KURATA, MASATOMO;KATSU, MASANORI;OGATA, TAKASHI;AND OTHERS;SIGNING DATES FROM 20130911 TO 20130930;REEL/FRAME:038830/0984

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION