WO2019176398A1 - Information processing device, information processing method, and program - Google Patents

Information processing device, information processing method, and program Download PDF

Info

Publication number
WO2019176398A1
WO2019176398A1 PCT/JP2019/004534 JP2019004534W WO2019176398A1 WO 2019176398 A1 WO2019176398 A1 WO 2019176398A1 JP 2019004534 W JP2019004534 W JP 2019004534W WO 2019176398 A1 WO2019176398 A1 WO 2019176398A1
Authority
WO
WIPO (PCT)
Prior art keywords
feature
modality
search
registration
information processing
Prior art date
Application number
PCT/JP2019/004534
Other languages
French (fr)
Japanese (ja)
Inventor
光平 西村
Original Assignee
ソニー株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニー株式会社 filed Critical ソニー株式会社
Priority to JP2020505682A priority Critical patent/JP7255585B2/en
Publication of WO2019176398A1 publication Critical patent/WO2019176398A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor

Definitions

  • the present disclosure relates to an information processing apparatus, an information processing method, and a program.
  • BoW BoW
  • N the number of vocabulary
  • LBP local binary patterns
  • a space spanned by such vectors is called a feature space and is used as a feature of search technology and machine learning technology.
  • the demand for feature spaces will increase further in the future, and there will be demands for handling multiple feature spaces and using them across and switching.
  • Non-Patent Document 1 discloses a technique for mapping text and images to a semantic space (a technique for mapping multimodal to a single space).
  • the present disclosure proposes an information processing apparatus, an information processing method, and a program that can further improve the convenience of a system that handles a plurality of feature spaces.
  • An information processing apparatus including a control unit is proposed.
  • the processor stores the registration object included in the registration request information in the storage unit in association with the unique first identification information common to the plurality of feature extraction units, and the registration object Control for converting according to the definition of the modality of the object and generating conversion data for registration; and control for outputting the first identification information and the conversion data for registration to a plurality of feature extractors corresponding to the modality
  • An information processing method including performing the above is proposed.
  • the computer controls the registration object included in the registration request information to be stored in the storage unit in association with the unique first identification information common to the plurality of feature extraction units, and the registration object is registered in the registration unit.
  • Control for converting according to the definition of the modality of the object and generating conversion data for registration; and control for outputting the first identification information and the conversion data for registration to a plurality of feature extractors corresponding to the modality A program for functioning as a control unit for performing the above is proposed.
  • FIG. 1 is a diagram illustrating an overview of an information processing system according to an embodiment of the present disclosure. As illustrated in FIG. 1, the information processing system according to the present embodiment includes an information processing apparatus 10 and a feature management server 20.
  • the feature management server 20 replaces various data (hereinafter referred to as “objects”) such as text, images, tabular data, and time-series data with N-dimensional vectors that well represent the features of the data (feature extraction).
  • objects such as text, images, tabular data, and time-series data
  • a feature extraction unit 202 an example of a feature extractor that performs processing
  • the feature extraction unit 202 and the feature space have a one-to-one relationship.
  • a plurality of feature management servers 20 may exist as shown in FIG. 1, each having one feature extraction unit 202 (that is, each feature management server 20 has one feature space. It can be said that it is managing.
  • the handling of a general feature space will be described below.
  • the degree of similarity and relationship between objects can be expressed, and for example, search and recommendation by concept can be performed.
  • the accuracy of other machine learning techniques can be improved, and for example, it can be used as a feature in a recognition system, data analysis, or the like.
  • different modalities can be handled in a unified manner. For example, text, handwriting, and notes can be handled in one feature space.
  • the “modality” is a data input / output format, and for example, various modalities are assumed as described below.
  • first feature space that can process text, notes, and Web pages from a viewpoint of meaning
  • second feature space that can process handwriting and an image from a viewpoint of shape
  • third-party extension extension by plug-in
  • third-party extension is assumed as a method of handling multiple feature spaces across. More specifically, it is possible for another person to associate another modality with the same feature space, or to configure another feature space from the same modality. It is also possible to expand by doing both of these (for example, by associating the modalities of “paper” and “case” and entering sensor data for each case, it is possible to search for cases close to the paper or case. it can). It can also be used in combination with the feature space handled by others (for example, searching for products from text or images).
  • pre-process when performing feature extraction, it is important to process (pre-process) data until it is put into the feature extractor, and it is necessary to follow a predetermined process.
  • pre-process -Accept only JPEG or PNG data-Unify to RGB 3 channel
  • Cleaning process removes noise in the text.
  • Web text an HTML tag or a JavaScript (registered trademark) source code) -Word division of sentences, word normalization, stop word removal, etc.
  • a rule for conversion to a predetermined data format is defined for each modality, and converted data (hereinafter also referred to as “entity”) converted according to the definition of the object modality in the information processing apparatus 10 corresponds to the modality.
  • entity converted data
  • To one or more feature management servers 20 server devices having a feature extraction unit 202 which is an example of a feature extractor).
  • search database (hereinafter referred to as a search DB)
  • data is stored for each feature space, the same data is registered in a plurality of databases, which is redundant. Therefore, it is conceivable to prepare a database capable of storing data and extracting data from the ID separately from the search DB, and storing only the ID in the search DB. At this time, if the user arbitrarily inputs the ID together with the data, there is a possibility that the following problem may occur and the user must take care.
  • a unique (unique) ID (identification information) common to a plurality of feature extractors is given to the conversion data of the registered object.
  • a plurality of feature spaces are created by uniformly converting objects for each modality and managing objects with unique IDs common to a plurality of feature spaces. It is possible to further improve the convenience of the handling system.
  • the information processing apparatus 10 includes a control unit 100, a communication unit 120, an output unit 130, and a storage unit 140.
  • the information processing apparatus 10 is a local terminal such as a smartphone, a tablet terminal, or a PC used by a user, for example.
  • Control unit 100 The control unit 100 functions as an arithmetic processing unit and a control unit, and controls overall operations in the information processing apparatus 10 according to various programs.
  • the control unit 100 is realized by an electronic circuit such as a CPU (Central Processing Unit) or a microprocessor, for example.
  • the control unit 100 may include a ROM (Read Only Memory) that stores programs to be used, calculation parameters, and the like, and a RAM (Random Access Memory) that temporarily stores parameters that change as appropriate.
  • ROM Read Only Memory
  • RAM Random Access Memory
  • the control unit 100 also functions as a feature space management unit 101 (Feature Space Manager) and a modality management unit 102 (Modality Manager).
  • the feature space management unit 101 performs feature space management (acquisition of one or more feature management server 20 IDs), object registration processing, search processing, and the like.
  • FIG. 2 shows an example of processing contents in the main functional blocks (the feature space management unit 101, the modality management unit 102, and the feature management unit 201) of the information processing system according to the present embodiment.
  • the feature space management unit 101 according to the present embodiment can perform the following processing.
  • the feature space management unit 101 when the registration request information is input, the feature space management unit 101 outputs the object and modality included in the registration request information to the modality management unit 102, and the modality management unit 102 follows the definition of the modality.
  • the converted conversion data (entity) and ID are acquired, and the entity and ID are output to the feature management server 20.
  • the feature space management unit 101 can output to one or more feature management servers 20 corresponding to the modality of the object.
  • the “feature management server 20 corresponding to the modality” is the feature management server 20 capable of handling the modality.
  • the feature space management unit 101 can grasp what information each feature space handles from the ID (space ID: string, etc.) of the feature management server 20.
  • the feature space management unit 101 can refer to each space ID and identify the feature management server 20 that manages the feature space that handles “Text”. .
  • the feature space management unit 101 may output the information to all the feature management servers 20 (in this case, the feature management server 20 may determine whether or not processing can be appropriately performed).
  • the feature space management unit 101 when the search request information is input, the feature space management unit 101 outputs the object and modality included in the search request to the modality management unit 102, and the converted data (converted by the modality management unit 102 according to the definition of the modality ( entity), and outputs the entity to the feature management server 20.
  • the feature space management unit 101 can output to one or more feature management servers 20 corresponding to the modality of the object.
  • the feature space management unit 101 may output the feature request to the feature management server 20 corresponding to the designated space ID when the search request includes a space ID as a search condition.
  • the feature space management unit 101 may output the information to all the feature management servers 20 (in this case, the feature management server 20 may determine whether or not processing can be appropriately performed).
  • the feature space management unit 101 extracts a corresponding object (that is, original data) from the storage unit 140 based on one or more IDs searched in the feature management server 20, and outputs them as search results to the search request source.
  • the search condition may further include filter information as an additional condition. Examples of the filter information include designation of a search DB (corresponding to a feature space to be searched), designation of the number of searches, and the like. For example, when the number of searches is specified, the feature space management unit 101 sets the similarity of each search result (similarity indicating the similarity between each search result and the feature amount of the conversion data (entity) of the search object). Based on this, a predetermined upper number of search results may be output to the search request source.
  • the modality management unit 102 manages modalities. For example, as shown in FIG. 2, the modality management unit 102 can perform the following processing. ⁇ Get Modalities (): string [] ⁇ Register Modality (modality: string, definer: Modality Definer) Create (obj: object, modality: string): entity
  • the modality management unit 102 registers the modality definition (data), or the modality definition unit 103 inputs the object (registered object) input from the feature space management unit 101 to the modality of the object.
  • the data is converted into a predetermined data format according to the definition, and converted data (entity) is generated.
  • the entity is shaped data that can be directly passed to the feature extraction unit 202.
  • the modality definition unit 103 assigns a unique ID common to the plurality of feature spaces to the generated entity, and outputs the entity and ID to the feature space management unit 101.
  • the modality definition unit 103 stores the object (registered object) and the ID assigned to the entity of the registered object in the storage unit 140 in association with each other.
  • Such an ID is a unique character string across a plurality of feature spaces that handle the same modality.
  • a hash value or the like may be used so that the same value is returned for the same entity.
  • Modality definition data includes definition data in a format that can be received as input data (obj) (for example, file name (string), OpenCV Mat format) (multiple) and output data (entity) format definition data (Only one) (for example, char [3] [256] [256]).
  • Modality definition data exists, for example, for each data format or for each pattern described above (for example, conversion between formats, unification of formats, reading of special data, etc.), and is stored in the storage unit 140.
  • the modality management unit 102 converts the registration object (input data) using the definition data of the modality, and outputs the conversion data (entity).
  • the modality definition unit 103 saves the data corresponding to the ID as necessary or confirms that the data is saved, the function of retrieving the data corresponding to the ID, and deletes the data corresponding to the ID. Has the function of The modality definition unit 103 exists for each modality.
  • the input unit 110 has a function of receiving user instruction content such as an operation input unit that receives an operation instruction from the user and a voice input unit that receives a voice instruction from the user, and outputs the instruction content to the control unit 100.
  • the operation input unit may be a touch sensor, a pressure sensor, or a proximity sensor.
  • the input unit 110 may be a physical configuration such as a button, a switch, and a lever.
  • the communication unit 120 is connected to an external device by wire or wireless, and transmits / receives data to / from the external device.
  • the communication unit 120 may be a wired / wireless local area network (LAN), Wi-Fi (registered trademark), Bluetooth (registered trademark), a mobile communication network (LTE: Long Term Evolution, 3G (third generation mobile communication)).
  • LAN local area network
  • Wi-Fi registered trademark
  • Bluetooth registered trademark
  • LTE Long Term Evolution
  • 3G third generation mobile communication
  • the output unit 130 has a function of presenting (outputting) information to the user, such as a display unit and an audio output unit.
  • the output unit 130 outputs a search screen or outputs a search result under the control of the control unit 100.
  • the storage unit 140 is realized by a ROM (Read Only Memory) that stores programs used in the processing of the control unit 100, calculation parameters, and the like, and a RAM (Random Access Memory) that temporarily stores parameters that change as appropriate.
  • ROM Read Only Memory
  • RAM Random Access Memory
  • the storage unit 140 stores feature space management information, modality definition information, and an object (substance data) assigned with a unique ID.
  • each process by the control unit 100 of the information processing apparatus 10 may be executed by a plurality of apparatuses or may be executed by a server on the network.
  • the feature management server 20 includes a control unit 200, a communication unit 210, and a feature amount database 220.
  • the feature management server 20 since the feature management server 20 has one feature extraction unit 202, it can be said that each feature management server 20 manages one feature space, but the present disclosure is limited to this. For example, if the feature management server 20 includes a plurality of feature extraction units 202, a plurality of feature spaces can be managed.
  • Control unit 200 The control unit 200 functions as an arithmetic processing unit and a control unit, and controls the overall operation in the feature management server 20 according to various programs.
  • the control unit 200 is realized by an electronic circuit such as a CPU (Central Processing Unit) or a microprocessor, for example.
  • the control unit 200 may include a ROM (Read Only Memory) that stores programs to be used, calculation parameters, and the like, and a RAM (Random Access Memory) that temporarily stores parameters that change as appropriate.
  • ROM Read Only Memory
  • RAM Random Access Memory
  • control unit 200 also functions as the feature management unit 201.
  • the feature management unit 201 performs feature extraction (for example, substitution processing for an N-dimensional vector) on the entity transmitted from the information processing apparatus 10 by using the feature extraction unit 202, and sets the extracted feature amount as the ID of the entity.
  • a process of registering in the feature amount database 220 in association with each other is performed.
  • An existing technique can be used for the feature quantity extraction algorithm, and is not particularly limited here.
  • the feature extraction unit 202 can handle different modalities in a unified manner.
  • the feature quantity extracted by the feature extraction unit 202 is registered in the feature quantity database 220.
  • the feature quantity database 220 exists for each modality.
  • the feature management unit 201 registers the feature amount extracted by the feature extraction unit 202 in the feature amount database 220 corresponding to the modality of the original data (entity transmitted from the information processing apparatus 10) in association with the ID.
  • a feature extraction unit 202a color feature feature space a
  • the feature amount extracted by the feature extraction unit 202a is stored in the feature amount database 220-1 corresponding to the handwritten data when extracted from the handwritten data, and the feature corresponding to the image data when extracted from the image data. It is stored in the quantity database 220-2.
  • the feature management unit 201 can also perform a search process (similarity search) using a feature space in response to a request from the feature space management unit 101. Since the feature quantity database 220 exists for each modality, the feature management unit 201 may perform a similarity search using the feature quantity database 220 corresponding to the target modality (search target modality). For example, the feature management unit 201 can perform the following processing as shown in FIG. ⁇ Get Space ID (): string ⁇ Register Database (modality: string, database: Feature Database) ⁇ Get Vector (entity: object, modality: string): vector Add (id: string, entity: object, modality: string) ⁇ Most Similar (query: object, modality: string, target Modality: string): Search Result []
  • the communication unit 210 is connected to an external device by wire or wireless, and transmits / receives data to / from the external device.
  • the communication unit 210 may be a wired / wireless local area network (LAN), Wi-Fi (registered trademark), Bluetooth (registered trademark), a mobile communication network (LTE: Long Term Evolution, 3G (third generation mobile communication).
  • LAN local area network
  • Wi-Fi registered trademark
  • Bluetooth registered trademark
  • LTE Long Term Evolution
  • 3G third generation mobile communication
  • the feature quantity database 220 accumulates the feature quantities extracted by the feature extraction unit 202. Each feature amount is associated with a unique ID assigned by the modality management unit 102. As described above, the feature amount database 220 exists for each modality.
  • the feature amount database 220 is stored in a storage unit (not shown) included in the feature management server 20.
  • the storage unit of the feature management server 20 is realized by a ROM that stores programs and calculation parameters used for the processing of the control unit 200, and a RAM that temporarily stores parameters that change as appropriate.
  • the configuration of the feature management server 20 according to the present embodiment has been specifically described above.
  • the configuration of the feature management server 20 shown in FIG. 1 is an example, and the present embodiment is not limited to this.
  • at least a part of the configuration of the feature management server 20 may be in an external device.
  • FIG. 3 shows an example of another configuration example of the information processing system according to this embodiment.
  • the feature extraction unit 240 and the feature quantity database 250 may be managed by separate servers (the feature management server 24 and the database server 25), respectively.
  • FIG. 4 is a sequence diagram illustrating an example of a flow of object registration processing in the information processing system according to the present embodiment.
  • the feature space management unit 101 of the information processing apparatus 10 acquires a registration request based on a user operation input or the like (step S103), and the object (obj) included in the registration request Along with the information on the modality (mdl) of the object, a request for generation (conversion data (entity)) is made to the modality management unit 102 (step S106).
  • the modality management unit 102 performs conversion data (entity) generation, unique ID assignment, and obj and unique ID storage processing by the modality definition unit 103 (step S109).
  • the modality definition unit 103 performs processing (common preprocessing) for converting an object into data of a predetermined format in accordance with the definition of the modality.
  • Specific examples of the processing include the following examples. In the case of a still image: JPEG data is converted into char [3] [256] [256] (multidimensional array) and smoothing processing is performed.
  • Handwriting Read the point sequence data and draw it with a white line of thickness 3 on a black image of char [3] [256] [256].
  • the feature space management unit 101 acquires at least an ID and an entity from the modality management unit 102 (step S112). Further, the modality management unit 102 may notify that the ID and obj are saved.
  • the feature space management unit 101 outputs a data addition (registration) request to all corresponding feature spaces (Feature Space) based on the acquired ID and entity (step S115).
  • the addition request includes ID, entity, and modality (mdl).
  • the corresponding feature space is a feature space (feature management server 20) that can handle the modality of the entity.
  • the processes shown in steps S115 to S121 are repeated for each feature space.
  • the feature management unit 201 extracts feature amounts using the feature extraction unit 202 (step S118).
  • the feature management unit 201 adds (registers) the extracted feature quantity to the feature quantity database 220 together with the acquired unique ID (step S121). At this time, the feature management unit 201 registers the feature amount database 220 corresponding to the modality of the extraction source entity.
  • the registration process according to this embodiment has been specifically described above.
  • the modality management unit 102 performs a predetermined conversion process for each modality so that the same data is preprocessed by a different process for each feature extractor. Therefore, the convenience of a system that handles a plurality of feature spaces can be improved. Further, managing the modality management unit 102 and the feature extraction unit 202 individually reduces the responsibility of each function (for example, it becomes easier to identify the cause at the time of an error).
  • the search system is completed by developing only the feature extractor, and the availability of the system increases.
  • the search DB that is, the feature database 220
  • only the ID and the feature are registered, and the original data (data entity) is separately managed by the modality definition unit 103. Therefore, the same data is registered in a plurality of databases. Can be avoided.
  • assigning unique IDs across multiple feature spaces to the same data multiple IDs can be associated with the same data, the same ID can be associated with multiple data, etc. Can be avoided.
  • the feature management unit 201 may register the feature data in the feature amount database 220 only when the original data is stored in the modality management unit 102. This ensures that the feature space management unit 101 can extract the original data from the ID when acquiring a search result to be described later.
  • the feature space (search DB system) development side can develop the feature extraction unit 202 without worrying about the input format. .
  • the registration process according to the present embodiment is not limited to the example shown in FIG.
  • step S115 described above it has been described that the addition instruction is given to the feature space (feature management server 20) corresponding to the modality.
  • the present embodiment is not limited to this, and the feature space management unit 101 can execute all the features.
  • An additional instruction may be given to the space (feature management server 20).
  • the feature space (feature management server 20) can determine whether the entity is a processable entity based on the modality.
  • FIG. 5 is a sequence diagram showing an example of the flow of search processing in the information processing system according to the present embodiment.
  • the feature space management unit 101 of the information processing apparatus 10 acquires a search request based on a user operation input or the like (step S133).
  • the search request includes an object (obj), a modality (mdl1) of the object, and a target modality (mdl2) indicating the modality to be searched.
  • the feature space management unit 101 makes a generation request (conversion data (entity)) to the modality management unit 102 together with information on the object (obj) and the modality (mdl1) of the object included in the search request. (Step S136).
  • the modality management unit 102 uses the modality definition unit 103 to generate conversion data (entity) and assign a unique ID (step S139). Specifically, the modality definition unit 103 performs processing (common preprocessing) for converting an object into data of a predetermined format in accordance with the definition of the modality.
  • the feature space management unit 101 acquires the ID and entity from the modality management unit 102 (step S142).
  • the feature space management unit 101 outputs a search request to all the corresponding feature spaces (Feature Space) based on the acquired entity (step S145).
  • the search request includes entity, mdl1 (original data modality), and mdl2 (target modality).
  • the corresponding feature space is a feature space (feature management server 20) that can handle mdl1 and mdl2.
  • the processing shown in steps S145 to S157 is repeated for each feature space.
  • the feature management unit 201 uses the feature extraction unit 202 to extract feature amounts (step S148).
  • the feature management unit 201 searches for a similar feature amount from the feature amount database 220 based on the extracted feature amount (step S151). At this time, the feature management unit 201 searches the feature amount database 220 corresponding to the requested target modality (mdl2). In the feature quantity database 220, the unique ID is associated with the feature quantity, and the feature management unit 201 searches the feature quantity database 220 for a feature quantity similar to the feature quantity of the requested entity, and is similar. The ID associated with the feature quantity and the similarity of the feature quantity: sim (similarity with the feature quantity of the requested entity, for example, the distance of the N-dimensional vector) is acquired. If there are a plurality of target modalities (mdl2), the processes shown in steps S151 to S154 are repeated for each feature quantity database 220.
  • the feature space management unit 101 acquires a search result (searched feature value ID, searched modality: mdl, and searched feature value similarity: sim) from the feature management unit 201 (step S157).
  • the search result may include a plurality of IDs, mdl, and sim.
  • the feature space management unit 101 specifies, for example, an ID of a feature amount having the highest similarity.
  • the feature space management unit 101 specifies the ID of the upper predetermined number of feature amounts based on, for example, the similarity.
  • the feature space management unit 101 makes a request for original data to the modality management unit 102 together with the identified ID and the corresponding modality information (step S160).
  • the modality management unit 102 acquires the original data (that is, the object) associated with the ID by the modality definition unit 103 (step S163) and outputs it to the feature space management unit 101 (step S166).
  • the processing shown in steps S160 to S166 can be performed for the number of search results to be output.
  • the feature space management unit 101 deletes the record (ID, mdl, sim).
  • the feature space management unit 101 outputs the search result (object, modality, and similarity) to the search request source (step S169).
  • the feature space management unit 101 may display a screen showing the search result on the output unit 130 and present it to the user.
  • the search processing according to this embodiment has been specifically described above.
  • the feature space constructed according to the present embodiment can be used for cross-sectional search in a plurality of feature spaces that handle different modalities.
  • the feature space (search DB) used for the search may be specified.
  • the identified feature space is included in the search request in step S113 as a space ID. For example, “I want to use a search DB created by XX company”, “I want to use a XX search site”, or the like is assumed.
  • FIG. 6 shows an example of a search screen in the present embodiment.
  • the search screen 30 shown in FIG. 6 is presented by the output unit 130, for example.
  • the user inputs a search object 301 and selects a search target (corresponding to a modality. For example, “photograph”, “illustration”, “document”, etc.), and the feature quantity of what is similar is searched for.
  • Select for example, “shape”, “color”, “meaning”, etc., and correspond to each feature space. For example, feature space constructed based on shape features, constructed based on color features
  • an object similar to the search object 301 acquired as a search result is presented.
  • the search object 301 is obtained as the search result.
  • an illustration having a similar shape, color, and / or meaning for example, a “modality: illustration” feature quantity database possessed by the feature management server 20 that handles a feature space constructed based on the feature of the shape. 220) and presented.
  • the feature amount search condition and / or may be arbitrarily selected by the user, or or may be set as a default.
  • the modality definition unit 103 may define a parent-child relationship (inclusion relationship) between modalities. Specific examples include the following.
  • modalities For example, use cases that define modalities including existing modalities are assumed. More specifically, it is assumed that a modality “text” already exists and there is a feature space A (feature extraction unit 202A) that handles “text”.
  • a modality of “mail” including text (body) and a user (sender) and a feature space B (feature extraction unit 202B) that can handle “mail” are added.
  • the first effect is that “can be registered in a plurality of feature spaces at the same time”. That is, since the text can be acquired from the mail, the same ID and object pair can be registered not only in the feature space B but also in the feature space A.
  • the feature space when the focus is on only the text is stored in the feature space A, and the feature space when the focus is on the text and the user is stored in the feature space B.
  • a search can be performed across other texts (in this case, it is necessary to assign a unique ID across all modalities, not just the same modality).
  • FIG. 7 is a diagram illustrating modularization of the feature extractor.
  • each modality using each feature extractor that can handle modalities such as texts and inclusive relations defined with mail and modalities, respectively.
  • a mail feature amount including a sentence feature amount (contents) and a user feature amount (sender).
  • FIG. 8 is a sequence diagram illustrating an example of registration processing of the first application example according to the present embodiment.
  • step S203 when the feature space management unit 101 of the information processing apparatus 10 acquires a registration request based on a user operation input or the like (step S203), the object (obj) included in the registration request and the object Along with information on the modality (mdl) of the object (for example, “Mail”), a request for generation of (entity) is made to the modality management unit 102 (step S206).
  • the modality management unit 102 uses the modality definition unit 103 to generate conversion data (entity), assign a unique ID, and store obj and a unique ID, and have an inclusion relationship with the modality of obj.
  • conversion data entity
  • sub entity is generated (step S209). For example, when the modality of the object is “Mail” and the modality (sub mdl) having an inclusion relation with this is “Text”, the modality definition unit 103 includes the mail data (obj) according to the definition of “Text”. Converts text data into a predetermined data format and generates a sub entity.
  • the feature space management unit 101 acquires at least ID, entity, and sub entity from the modality management unit 102 (step S212). Further, the modality management unit 102 may notify that the ID and obj are saved.
  • the feature space management unit 101 outputs a data addition (registration) request to all corresponding feature spaces (eg, feature space B) based on the acquired ID and entity (eg, Mail Entity) ( Step S215).
  • the subsequent processing related to feature amount extraction shown in steps S218 to S221 is the same as that in steps S118 to S121 shown in FIG. 4 and will not be described in detail.
  • the feature quantity (Mail Vector) is registered.
  • the Get Vector (Step S227) of the feature space A handling the text described below is used. It may be.
  • the feature space management unit 101 outputs a data addition (registration) request for all corresponding feature spaces (eg, feature space A) for the same ID and sub entity (eg, Text Entity) (steps S224 to 230).
  • the feature space A is a feature space corresponding to text only, and the feature amount (Text Vector) of the text extracted from the Text Entity is registered.
  • a feature space having an inclusion relationship can be used, and a feature quantity can be registered in the feature space.
  • FIG. 9 is a sequence diagram illustrating an example of search processing of the first application example according to the present embodiment.
  • the feature space management unit 101 of the information processing apparatus 10 acquires a search request based on a user operation input or the like (step S243).
  • the search request includes an object (obj), a modality (mdl1) of the object, and a target modality (mdl2) indicating the modality to be searched.
  • mdl1 Mail
  • mdl2 Text.
  • the feature space management unit 101 requests the modality management unit 102 to generate (entity) together with information on the object (obj) and the modality (mdl1) of the object included in the search request (step S246). .
  • the modality management unit 102 uses the modality definition unit 103 to generate conversion data (entity) and assign a unique ID, and based on the definition of the modality (sub mdl) having an inclusion relation with the modality of obj. Then, the sub entity is generated (step S249). For example, when the modality of the object is “Mail” and the modality (sub mdl) having an inclusion relation with this is “Text”, the modality definition unit 103 includes the mail data (obj) according to the definition of “Text”. Converts text data into a predetermined data format and generates a sub entity.
  • the feature space management unit 101 acquires ID, entity, and sub entity from the modality management unit 102 (step S252).
  • the feature space management unit 101 outputs a search request to all corresponding feature spaces (Feature Space) based on the acquired entity (for example, Mail Entity) (step S255).
  • the search request includes entity, mdl1 (original data modality, for example, Mail), and mdl2 (target modality, for example, Text).
  • the corresponding feature space is a feature space that can handle mdl1 and mdl2 (for example, a feature space corresponding to both mail and text).
  • steps S258 to S267 are the same as steps S148 to S157 shown in FIG.
  • the feature space management unit 101 outputs a search request to all corresponding feature spaces (Feature Space) based on the ID and sub entity (for example, Text Entity) (step S270).
  • the search request includes sub entity (eg, Text Entity), mdl1 (sub mdl, eg, Text), and mdl2 (target modality, eg, Text).
  • the corresponding feature space is a feature space that can handle mdl1 and mdl2, and here mdl1 and mdl2 are the same “Text”, and therefore feature space A corresponding to text corresponds.
  • a search is performed in the feature space A (steps S273 to S279), and the feature space management unit 101 acquires a search result from the feature management unit 201 (step S282).
  • the feature space management unit 101 sends a request for original data to the modality management unit 102 together with the acquired ID and the corresponding modality (for example, Text), as in steps S160 to S169 shown in FIG.
  • the object acquired based on the ID by the modality definition unit 103 is output from the modality management unit 102 to the feature space management unit 101 (step S291).
  • the feature space management unit 101 outputs the search result (object, modality, and similarity) to the search request source (step S294).
  • the feature space management unit 101 can output the final search result to the search request source after re-evaluating the search result based on the similarity and weighting of the search result from each feature extractor.
  • the weighting is, for example, weighting of the feature space. Such weighting can be arbitrarily set by a search request source (for example, a user).
  • FIG. 10 is a diagram illustrating an example of a search screen in this application example.
  • the search screen 32 includes a search object 321, a search target selection region 322, a region 323 for selecting the feature amount of what is similar to search, and a search button 326. It is displayed.
  • the weight of the selected feature amount can be set by operating the slide bar 324. For example, when it is desired to prioritize “color feature” among “shape feature” and “color feature”, the operation unit 325 of the slide bar 324 is moved toward “color feature”. Thereby, for example, the weight (w) is set as follows on the system side.
  • w (weights) ⁇ space1: 0.8, space2: 0.2 ⁇
  • search results giving priority to “color features” are displayed.
  • the setting of the weighting of the feature space is not limited to the example illustrated in FIG. 10.
  • the weighting may be set based on what is selected by the user from the search results, and the search results may be presented again.
  • An example is shown in FIG.
  • the system outputs the illustration 341 as the search result, assuming that the illustration 341 is close to the user's intention.
  • the weighting may be set so as to prioritize the feature space (feature extractor, that is, the feature extraction unit 202), and the search result may be presented again.
  • FIG. 12 is a sequence diagram illustrating an example of search processing of the second application example.
  • the feature space management unit 101 of the information processing apparatus 10 acquires a search request based on a user operation input or the like (step S303).
  • the search request includes an object (obj), a modality (mdl1) of the object, a target modality (mdl2) indicating a modality to be searched, and a weight (w) of the feature space.
  • step S315 to S321 a search process similar to the process shown in steps S145 to 157 of FIG. 5 described above is performed, and thus detailed description thereof is omitted here.
  • step S318 processing similar to that shown in steps S148 to S154 in FIG. 5 is performed, but detailed illustration is omitted.
  • the object A as a search result is weighted to the first feature space (sim (space1): 0.9) with the similarity (sim (space1): 0.9) when searched from the first feature space (space1).
  • a value obtained by adding the values (sim (new): 0.78) is calculated as a new similarity. This is because an ID associated with the same data may be registered in a plurality of feature spaces.
  • it is assumed that the search result is searched only from one feature space.
  • the second feature space has a similarity (sim (space2): 0.9) when the object C is searched from the second feature space (space2).
  • a value (sim (new): 0.18) obtained by multiplying the weight (space2: 0.2) is calculated as a new similarity.
  • the feature space management unit 101 specifies, for example, a predetermined number of search results (IDs) on the basis of the new similarity.
  • the feature space management unit 101 makes a request for original data to the modality management unit 102 together with the identified ID and the corresponding modality, similarly to steps S160 to S169 shown in FIG. 5 described above (step S327).
  • the object acquired based on the ID by the modality definition unit 103 (step S330) is output from the modality management unit 102 to the feature space management unit 101 (step S333).
  • the feature space management unit 101 outputs the search result (object, modality, and similarity) to the search request source (step S336).
  • the suggestion system will meet the demand according to the usage status of these multiple applications.
  • Content Web page, text, image, etc.
  • FIG. 13 is a functional block diagram showing an example of the configuration of the present system.
  • a suggestion system can be realized by an information processing apparatus 10x.
  • the information processing apparatus 10x functions as one or more applications 105, an information collection unit 106, a suggestion unit 107, a feature space management unit 101x, and a modality management unit 102x. These can be implemented by the control unit 100 of the information processing apparatus 10.
  • the application 105 is various application programs such as a web browser, a map application, and a notebook application.
  • the information collection unit 106 has a function of monitoring the operation of each application 105 and collecting and storing user operation information (that is, application usage status) in each application 105.
  • the information collecting unit 106 may use an OS (Operating System).
  • the suggestion unit 107 generates a search request based on the operation information collected by the information collection unit 106, and makes a search request to the feature space management unit 101x. For example, the suggestion unit 107, based on the request of the modality (mdl1) and content (obj) of the content being browsed / edited acquired from each application 105 from the information collection unit 106, and the modality (mdl2) of the required content, A search request may be generated. For example, the following examples of content modality acquired from each application 105 and required content modality requests are assumed. -Web browser ... Browsing: Web page, Request: Web page / Map application ... Browsing: Address, Request: None * Note app ... Editing: Text / Image, Request: Text / Image
  • the feature space management unit 101x performs a search process using one or more feature spaces in response to a request from the suggestion unit 107.
  • the search process is the same as in the above-described embodiment.
  • the feature space management unit 101x acquires the entity obtained by converting the obj by the modality management unit 102x, and the entity, mdl1 (for example, Web page, address, text), and mdl2
  • mdl1 for example, Web page, address, text
  • mdl2 A search request is made to the feature management server 20 based on (for example, Web page, image).
  • the feature space management unit 101x outputs the search result to the suggestion unit 107.
  • the modality management unit 102x has the same function as that of the modality management unit 102 described with reference to FIG. 1, and the modality definition unit 103 performs processing for converting obj into a predetermined data format according to the definition of the modality of mdl1.
  • the generated entity is output to the feature space management unit 101x.
  • FIG. 14 is a sequence diagram showing an example of the flow of search processing in the suggestion system of this application example.
  • step S403 when one or more applications 105 are operated by the user (step S403), the content being handled (post; obj, mdl1) and the required content request (request) Mdl2) is performed on the information collecting unit 106 (step S406).
  • the post for example, a web page of “Kinkakuji”, an address “Kita-ku, Kyoto ... 1-2-3”, a travel-related text, and the like can be cited.
  • examples of the request include a web page and an image.
  • the information collection unit 106 outputs the collected information (post, request) to the suggestion unit 107 (step S412).
  • the suggestion unit 107 makes a search request to the feature space management unit 101x (step S415).
  • the search request includes the content included in post as obj, the modality thereof as mdl1, and the content modality included in request as mdl2.
  • step S4108 a search process is executed in the feature space management unit 101x (step S418).
  • step S4108 processing similar to that in steps S136 to S166 of FIG. 5 (generation of entity from obj and mdl1, search based on entity, mdl1, and mdl2, acquisition of object from search result ID) is performed. Detailed description is omitted.
  • the suggestion unit 107 acquires a search result from the feature space management unit 101x (step S421).
  • the suggestion unit 107 may rank (re-evaluate) the search results according to the similarity and weighting (W) of the search results (step S424).
  • the suggestion unit 107 may set a weight as shown in Table 2 below for each input and output, and rank by multiplying the similarity. In this application example, such reevaluation may be skipped.
  • the suggestion unit 107 creates a display screen showing the search result (step S427) and presents it to the user (step S430).
  • the suggestion unit 107 may update the weighting (W) used in the above step 424 when the usage status feedback is obtained from the user (step S433).
  • suggestion unit 107 may suggest to the user and obtain feedback from the user via the application 105.
  • FIG. 15 shows an example of operation information and request information acquired from an application according to this application example.
  • the operation information as shown on the left in FIG. 15 and the request information as shown on the right in FIG. 15 are acquired from each application, and the requested information is suggested based on the operation information.
  • the information processing system can further improve the convenience of a system that handles a plurality of feature spaces.
  • a computer program for causing the information processing apparatus 10 or the feature management server 20 to perform the functions of the information processing apparatus 10 or the feature management server 20 on hardware such as a CPU, ROM, and RAM incorporated in the information processing apparatus 10 or the feature management server 20 described above.
  • a computer-readable storage medium storing the computer program is also provided.
  • this technique can also take the following structures.
  • An information processing apparatus comprising a control unit that performs the following.
  • the controller is A control for converting the search object included in the search request according to the definition of the modality of the search object, and generating conversion data for search; Control to output the search conversion data to the feature extractor corresponding to the modality of the search object and the target modality included in the search request;
  • the information processing apparatus according to (3), wherein a corresponding object is acquired from the storage unit based on the second identification information, and is output as a search result.
  • the controller is The degree of similarity indicating the degree of similarity of the feature is acquired from the feature extractor together with the second identification information associated with the feature similar to the feature extracted from the search conversion data, (4) The information processing apparatus described in 1. (6) The information processing apparatus according to (4) or (5), wherein the search request further includes filter information as a search condition.
  • the controller is When the registration request information is input, according to the definition of the sub-modality having a parent-child relationship with the modality of the registration object, control for converting the registration object to generate sub-conversion data for registration; Control for outputting the first identification information and the sub-transformed data to one or more feature extractors corresponding to the sub-modalities;
  • the information processing apparatus according to any one of (1) to (6), further performing: (8)
  • the controller is When the search request is input, sub-converted data for search is generated by converting data corresponding to the sub-modality among the search objects according to the definition of the sub-modality having a parent-child relationship with the modality of the search object.
  • Control Control to output the sub-transform data for search to one or more feature extractors corresponding to the sub-modality and the target modality;
  • the information processing apparatus according to any one of (3) to (6), further performing: (9)
  • the controller is Control for ranking the plurality of second identification information based on the second identification information and similarity obtained from the feature extractor based on the search request, and the weight of the feature extractor; A control for outputting the upper predetermined number of the second identification information as the search results;
  • the information processing apparatus according to any one of (4) to (6), further performing: (10)
  • the controller is Generating the search request for searching for content to be proposed to the user based on information including user operation information output from one or more applications;
  • the information processing apparatus according to any one of (4) to (6), wherein one or more pieces of the second identification information acquired from the feature extractor are output as the search result.
  • the controller is The content included in the information and handled by the application is the search object,
  • the modality of the content is the modality of the search object,
  • the information processing apparatus includes: The information according to any one of (1) to (11), further including a communication unit that transmits the first identification information and the conversion data for registration to a feature management server having the feature extractor. Processing equipment.
  • Processor Control for storing the registration object included in the registration request information in the storage unit in association with the unique first identification information common to the plurality of feature extraction units; Control for converting the registration object according to the definition of the modality of the registration object, and generating conversion data for registration; Control to output the first identification information and the conversion data for registration to a plurality of feature extractors corresponding to the modality;
  • An information processing method including performing.
  • Computer Control for storing the registration object included in the registration request information in the storage unit in association with the unique first identification information common to the plurality of feature extraction units; Control for converting the registration object according to the definition of the modality of the registration object, and generating conversion data for registration; Control to output the first identification information and the conversion data for registration to a plurality of feature extractors corresponding to the modality; A program for functioning as a control unit for performing
  • control unit 101 101x feature space management unit 102, 102x modality management unit 103 modality definition unit 105 application 106 information collection unit 107 suggestion unit 110 input unit 120 communication unit DESCRIPTION OF SYMBOLS 130 Output part 140 Storage part 200 Control part 201 Feature management part 202 Feature extraction part 210 Communication part 220 Feature quantity database 240 Feature extraction part 250 Feature quantity database

Abstract

[Problem] To provide: an information processing device capable of further enhancing the convenience of use of a system in which a plurality of feature spaces are handled; an information processing method; and a program. [Solution] This information processing device is provided with a control unit which performs: a control for storing a registration object, included in registration request information, in association with unique first identification information that is common in a plurality of feature extracting units; a control for converting the registration object according to a definition of the modality of the registration object and generating conversion data for registration; and a control for outputting the first identification information and the conversion data for registration to a plurality of feature extractors corresponding to the modality.

Description

情報処理装置、情報処理方法、および、プログラムInformation processing apparatus, information processing method, and program
 本開示は、情報処理装置、情報処理方法、および、プログラムに関する。 The present disclosure relates to an information processing apparatus, an information processing method, and a program.
 従来、検索技術や機械学習技術の領域では、テキストや画像、表形式データ、時系列データ等の様々なデータを、そのデータの特徴をよく表すN次元ベクトルで代替すること(特徴抽出)が広く行われている。 Conventionally, in the field of search technology and machine learning technology, various data such as text, images, tabular data, and time-series data are widely replaced with N-dimensional vectors that well represent the characteristics of the data (feature extraction). Has been done.
 ベクトル化の例として、自然言語処理の界隈では、語彙数Nを次元数とし、出現した単語のみ値を持つベクトルを用いて文章を代表するBoW(Bag of Words)と呼ばれる手法が一般的に使われている。また、画像処理の界隈では、局所バイナリパターン(local binary pattern:LBP)などの局所特徴をコードワードと見なしたBoVW(Bag of Visual Words)といった技法の他、データを入力とし、特徴ベクトルを出力する深層学習モデルも多数考案されている。表形式データは、カテゴリを1-hotな数次元ベクトルに変換する処理や、各整数値や実数値を正規化する処理によって1つのベクトルに変換される。 As an example of vectorization, in the field of natural language processing, a technique called BoW (Bag of Words) that represents a sentence using a vector in which the number of vocabulary N is the number of dimensions and the value of only the word that appears is generally used. It has been broken. In the field of image processing, in addition to a technique such as BoVW (Bag of Visual Words) that considers local features such as local binary patterns (LBP) as codewords, data is input and feature vectors are output. Many deep learning models have been devised. The tabular data is converted into one vector by the process of converting the category into a 1-hot number-dimensional vector and the process of normalizing each integer value or real value.
 このようなベクトルで張られた空間は特徴空間と呼ばれ、検索技術や機械学習技術の素性として用いられる。今後特徴空間の需要はさらに増大し、特徴空間を複数扱い、横断的に利用したり、切り替えたりといった需要が想定される。 A space spanned by such vectors is called a feature space and is used as a feature of search technology and machine learning technology. The demand for feature spaces will increase further in the future, and there will be demands for handling multiple feature spaces and using them across and switching.
 また、異なるモダリティ(入出力形式)を統一的に扱う横断的な特徴空間も存在する。例えば下記非特許文献1では、テキストと画像を意味的な空間にマッピングする技術(マルチモーダルをシングルスペースにマッピングする技術)が開示されている。 There is also a cross-sectional feature space that handles different modalities (input / output formats) in a unified manner. For example, the following Non-Patent Document 1 discloses a technique for mapping text and images to a semantic space (a technique for mapping multimodal to a single space).
 ここで、特徴抽出を行う際には、特徴抽出器に入れるまでのデータの加工(前処理)が重要であり、決まったプロセスを踏む必要があるが、これらは特徴抽出器毎に異なるため、特徴空間を複数扱う場合は同じデータを特徴抽出器毎に異なるプロセスでそれぞれ前処理を行わなければならず、冗長であった。 Here, when performing feature extraction, it is important to process the data (preprocessing) until it is put into the feature extractor, and it is necessary to follow a predetermined process, but these differ for each feature extractor, When a plurality of feature spaces are handled, the same data must be preprocessed by different processes for each feature extractor, which is redundant.
 そこで、本開示では、複数の特徴空間を扱うシステムの利便性をより向上させることが可能な情報処理装置、情報処理方法、および、プログラムを提案する。 Therefore, the present disclosure proposes an information processing apparatus, an information processing method, and a program that can further improve the convenience of a system that handles a plurality of feature spaces.
 本開示によれば、登録要求情報に含まれる登録オブジェクトを複数の特徴抽出部に共通する一意の第1の識別情報と関連付けて記憶部に記憶する制御と、前記登録オブジェクトを前記登録オブジェクトのモダリティの定義に従って変換し、登録用の変換データを生成する制御と、前記第1の識別情報と前記登録用の変換データを、前記モダリティに対応する複数の特徴抽出器に出力する制御と、を行う制御部を備える、情報処理装置を提案する。 According to the present disclosure, control for storing a registration object included in registration request information in a storage unit in association with unique first identification information common to a plurality of feature extraction units, and the registration object modality of the registration object And converting the first identification information and the registration conversion data to a plurality of feature extractors corresponding to the modality. An information processing apparatus including a control unit is proposed.
 本開示によれば、プロセッサが、登録要求情報に含まれる登録オブジェクトを複数の特徴抽出部に共通する一意の第1の識別情報と関連付けて記憶部に記憶する制御と、前記登録オブジェクトを前記登録オブジェクトのモダリティの定義に従って変換し、登録用の変換データを生成する制御と、前記第1の識別情報と前記登録用の変換データを、前記モダリティに対応する複数の特徴抽出器に出力する制御と、を行うことを含む、情報処理方法を提案する。 According to the present disclosure, the processor stores the registration object included in the registration request information in the storage unit in association with the unique first identification information common to the plurality of feature extraction units, and the registration object Control for converting according to the definition of the modality of the object and generating conversion data for registration; and control for outputting the first identification information and the conversion data for registration to a plurality of feature extractors corresponding to the modality An information processing method including performing the above is proposed.
 本開示によれば、コンピュータを、登録要求情報に含まれる登録オブジェクトを複数の特徴抽出部に共通する一意の第1の識別情報と関連付けて記憶部に記憶する制御と、前記登録オブジェクトを前記登録オブジェクトのモダリティの定義に従って変換し、登録用の変換データを生成する制御と、前記第1の識別情報と前記登録用の変換データを、前記モダリティに対応する複数の特徴抽出器に出力する制御と、を行う制御部として機能させるための、プログラムを提案する。 According to the present disclosure, the computer controls the registration object included in the registration request information to be stored in the storage unit in association with the unique first identification information common to the plurality of feature extraction units, and the registration object is registered in the registration unit. Control for converting according to the definition of the modality of the object and generating conversion data for registration; and control for outputting the first identification information and the conversion data for registration to a plurality of feature extractors corresponding to the modality A program for functioning as a control unit for performing the above is proposed.
 以上説明したように本開示によれば、複数の特徴空間を扱うシステムの利便性をより向上させることが可能となる。 As described above, according to the present disclosure, it is possible to further improve the convenience of a system that handles a plurality of feature spaces.
 なお、上記の効果は必ずしも限定的なものではなく、上記の効果とともに、または上記の効果に代えて、本明細書に示されたいずれかの効果、または本明細書から把握され得る他の効果が奏されてもよい。 Note that the above effects are not necessarily limited, and any of the effects shown in the present specification, or other effects that can be grasped from the present specification, together with or in place of the above effects. May be played.
本開示の一実施形態による情報処理システムの概要について説明する図である。It is a figure explaining an outline of an information processing system by one embodiment of this indication. 本実施形態による情報処理システムの主な機能ブロックにおける処理内容の一例を示す図である。It is a figure which shows an example of the processing content in the main functional block of the information processing system by this embodiment. 本実施形態による情報処理システムの他のシステム構成例の一例を示す図である。It is a figure which shows an example of the other system configuration example of the information processing system by this embodiment. 本実施形態による情報処理システムにおけるオブジェクトの登録処理の流れの一例を示すシーケンス図である。It is a sequence diagram which shows an example of the flow of the registration process of the object in the information processing system by this embodiment. 本実施形態による情報処理システムにおける検索処理の流れの一例を示すシーケンス図である。It is a sequence diagram which shows an example of the flow of the search process in the information processing system by this embodiment. 本実施形態における検索画面の一例を示す図である。It is a figure which shows an example of the search screen in this embodiment. 本実施形態による第1の応用例における特徴抽出器のモジュール化について説明する図である。It is a figure explaining modularization of the feature extractor in the 1st application example by this embodiment. 本実施形態による第1の応用例の登録処理の一例を示すシーケンス図である。It is a sequence diagram which shows an example of the registration process of the 1st application example by this embodiment. 本実施形態による第1の応用例の検索処理の一例を示すシーケンス図である。It is a sequence diagram which shows an example of the search process of the 1st application example by this embodiment. 本実施形態による第2の応用例における検索画面の一例を示す図である。It is a figure which shows an example of the search screen in the 2nd application example by this embodiment. 本実施形態による第2の応用例における検索画面の他の例を示す図である。It is a figure which shows the other example of the search screen in the 2nd application example by this embodiment. 本実施形態による第2の応用例の検索処理の一例を示すシーケンス図である。It is a sequence diagram which shows an example of the search process of the 2nd application example by this embodiment. 本実施形態による第3の応用例のサジェストシステムの構成の一例を示す機能ブロック図である。It is a functional block diagram which shows an example of a structure of the suggestion system of the 3rd application example by this embodiment. 本実施形態による第3の応用例のサジェストシステムにおける検索処理の流れの一例を示すシーケンス図である。It is a sequence diagram which shows an example of the flow of the search process in the suggestion system of the 3rd application example by this embodiment. 本実施形態による第3の応用例におけるアプリケーションから取得する操作情報と要求情報の一例を示す図である。It is a figure which shows an example of the operation information and request information which are acquired from the application in the 3rd application example by this embodiment.
 以下に添付図面を参照しながら、本開示の好適な実施の形態について詳細に説明する。なお、本明細書及び図面において、実質的に同一の機能構成を有する構成要素については、同一の符号を付することにより重複説明を省略する。 Hereinafter, preferred embodiments of the present disclosure will be described in detail with reference to the accompanying drawings. In addition, in this specification and drawing, about the component which has the substantially same function structure, duplication description is abbreviate | omitted by attaching | subjecting the same code | symbol.
 また、説明は以下の順序で行うものとする。
 1.本開示の一実施形態による情報処理システムの概要
 2.構成
  2-1.情報処理装置10の構成
  2-2.特徴管理サーバ20の構成
 3.動作処理
  3-1.登録処理
  3-2.検索処理
 4.応用例
  4-1.第1の応用例:モダリティの包含関係の定義
  4-2.第2の応用例:検索結果のマージ
  4-3.第3の応用例:サジェストシステム
 5.まとめ
The description will be made in the following order.
1. 1. Overview of information processing system according to an embodiment of the present disclosure Configuration 2-1. Configuration of information processing apparatus 10 2-2. 2. Configuration of the feature management server 20 Operation processing 3-1. Registration process 3-2. Search processing Application example 4-1. First application example: definition of modality inclusion 4-2. Second application example: merge of search results 4-3. Third application example: Suggest system 5. Summary
 <<1.本開示の一実施形態による情報処理システムの概要>>
 図1は、本開示の一実施形態による情報処理システムの概要について説明する図である。図1に示すように、本実施形態による情報処理システムは、情報処理装置10と、特徴管理サーバ20とを有する構成となっている。
<< 1. Overview of Information Processing System According to One Embodiment of Present Disclosure >>
FIG. 1 is a diagram illustrating an overview of an information processing system according to an embodiment of the present disclosure. As illustrated in FIG. 1, the information processing system according to the present embodiment includes an information processing apparatus 10 and a feature management server 20.
 特徴管理サーバ20は、テキストや画像、表形式データ、時系列データ等の様々なデータ(以下、「オブジェクト」と称す)を、そのデータの特徴をよく表すN次元ベクトルで代替する処理(特徴抽出処理)を行う特徴抽出部202(特徴抽出器の一例)を有し、かかるベクトルで張られた空間(特徴空間)を管理している。本明細書では、特徴抽出部202と特徴空間は1対1の関係にある。特徴管理サーバ20は、図1に示すように複数存在していてもよく、それぞれ1の特徴抽出部202を有する構成となっている(すなわち、各特徴管理サーバ20が、それぞれ1の特徴空間を管理していると言える)。 The feature management server 20 replaces various data (hereinafter referred to as “objects”) such as text, images, tabular data, and time-series data with N-dimensional vectors that well represent the features of the data (feature extraction). A feature extraction unit 202 (an example of a feature extractor) that performs processing), and manages a space (feature space) spanned by such vectors. In the present specification, the feature extraction unit 202 and the feature space have a one-to-one relationship. A plurality of feature management servers 20 may exist as shown in FIG. 1, each having one feature extraction unit 202 (that is, each feature management server 20 has one feature space. It can be said that it is managing.
 ここで、一般的な特徴空間の扱いについて以下説明する。特徴空間では、オブジェクト間の類似度や関係の表現ができ、例えば概念での検索および推薦が可能となる。また、特徴空間では、他の機械学習技術の精度を向上させることができ、例えば、認識系、データ解析等で素性として利用することが可能となる。また、特徴空間では、異なるモダリティを統一的に扱うことができ、例えば、テキストも手書きもノートも1つの特徴空間で扱うことが可能となる。本明細書において「モダリティ」とは、データの入出力形式であって、例えば下記に挙げるように多岐に渡るモダリティが想定される。
・テキスト-単語、文章、HTML(HyperText Markup Language)など
・メディア-RGB画像、深度画像、ベクタ画像、動画、音声など
・複合文書-オフィス文書、PDF、Webページ、電子メールなど
・メタデータ-ユーザ、日付など
・センサデータ-現在位置、加速度、心拍数など
・アプリケーションデータ-起動ログ、処理中のファイル情報など
このようなモダリティに対して特徴空間を定義でき、自由に拡張可能となる。
Here, the handling of a general feature space will be described below. In the feature space, the degree of similarity and relationship between objects can be expressed, and for example, search and recommendation by concept can be performed. In the feature space, the accuracy of other machine learning techniques can be improved, and for example, it can be used as a feature in a recognition system, data analysis, or the like. In the feature space, different modalities can be handled in a unified manner. For example, text, handwriting, and notes can be handled in one feature space. In this specification, the “modality” is a data input / output format, and for example, various modalities are assumed as described below.
-Text-Word, Sentence, HTML (HyperText Markup Language), etc.-Media-RGB image, depth image, vector image, video, audio, etc.-Compound document-Office document, PDF, Web page, e-mail, etc.-Metadata-User , Date, etc.-sensor data-current position, acceleration, heart rate, etc.-application data-startup log, file information being processed, etc. The feature space can be defined and expanded freely.
 また、複数の特徴空間を横断的に扱うことも可能である。例えば、テキスト、ノート、およびWebページを意味の視点で処理することが可能な第1の特徴空間と、手書きと画像を形の視点で処理することが可能な第2の特徴空間とがある場合に、これらを所定の関数を通してマッピングしてもよい。これにより、例えば、テキストで画像や手書きを検索したり、今見ているWebページに関連したノートを探したり、今書いているノートに近いタグを自動的に付与するシステムをすぐに生成することが可能となる。 It is also possible to handle multiple feature spaces across. For example, there is a first feature space that can process text, notes, and Web pages from a viewpoint of meaning, and a second feature space that can process handwriting and an image from a viewpoint of shape These may be mapped through a predetermined function. This makes it possible to quickly generate a system that, for example, searches for images and handwriting by text, searches for notes related to the web page that you are currently viewing, and automatically assigns tags that are close to the note you are currently writing. It becomes possible.
 また、複数の特徴空間を横断的に扱う方法として、サードパーティによる拡張(プラグインによる拡張)も想定される。より具体的には、他者でも同じ特徴空間に別のモダリティを関連付けたり、同じモダリティから別の特徴空間を構成することが可能である。また、これらを両方行って拡張することも可能である(例えば、「論文」や「判例」のモダリティを関連付けると共に、症例毎のセンサデータを入れることで、論文や判例に近い症例を探すことができる)。また、他者が扱う特徴空間と混ぜて利用することもできる(例えば、テキストや画像から商品を探す)。 Also, third-party extension (extension by plug-in) is assumed as a method of handling multiple feature spaces across. More specifically, it is possible for another person to associate another modality with the same feature space, or to configure another feature space from the same modality. It is also possible to expand by doing both of these (for example, by associating the modalities of “paper” and “case” and entering sensor data for each case, it is possible to search for cases close to the paper or case. it can). It can also be used in combination with the feature space handled by others (for example, searching for products from text or images).
 このようにして、複数のマルチモーダル空間中のあらゆるデータを横断検索することが可能となる。 In this way, it becomes possible to search across all data in multiple multimodal spaces.
 (背景)
 ここで、このような複数の特徴空間を横断的に検索するシステムの構築に関し、以下のような問題が考えられる。
(background)
Here, the following problems can be considered regarding the construction of a system for searching across such a plurality of feature spaces.
 まず、特徴抽出を行う際には、特徴抽出器に入れるまでのデータの加工(前処理)が重要であり、決まったプロセスを踏む必要がある。前処理の決まったプロセスとしては、例えば画像を例に取ると、
・JPEGもしくはPNGデータのみ受け付ける
・RGB3チャンネル、256x256pxに統一する
・アスペクト比が1:1でない場合に短辺を引き伸ばす
・平滑化等のフィルタ処理を掛ける
といった処理が行われる。
First, when performing feature extraction, it is important to process (pre-process) data until it is put into the feature extractor, and it is necessary to follow a predetermined process. For example, taking an image as an example of the pre-process,
-Accept only JPEG or PNG data-Unify to RGB 3 channel, 256x256px-Extend the short side when the aspect ratio is not 1: 1-Apply processing such as smoothing.
 また、テキストの前処理の例としては、
・クリーニング処理(テキスト中のノイズを除去。例えばWebテキストの場合、HTMLタグやJavaScript(登録商標)のソースコードなど)
・文章の単語分割
・単語の正規化
・ストップワード除去
などが挙げられる。
As an example of text preprocessing,
Cleaning process (removes noise in the text. For example, in the case of Web text, an HTML tag or a JavaScript (registered trademark) source code)
-Word division of sentences, word normalization, stop word removal, etc.
 しかしながら、これらは特徴抽出器毎に異なるため、特徴空間を複数扱う場合、同じデータを特徴抽出器毎に異なるプロセスでそれぞれ前処理を行わなければならず、冗長であった。 However, since these differ for each feature extractor, when handling a plurality of feature spaces, the same data must be pre-processed by different processes for each feature extractor, which is redundant.
 そこで、本実施形態では、複数の特徴抽出器における前処理を共通化することで、複数の特徴空間を扱うシステムの利便性をより向上させることを可能とする。 Therefore, in the present embodiment, it is possible to further improve the convenience of a system that handles a plurality of feature spaces by sharing preprocessing in a plurality of feature extractors.
 具体的には、モダリティ毎に所定のデータ形式へ変換するルールを定義付け、情報処理装置10においてオブジェクトのモダリティの定義に従って変換した変換データ(以下、「entity」とも称す)を、当該モダリティに対応する1以上の特徴管理サーバ20(特徴抽出器の一例である特徴抽出部202を有するサーバ装置)に出力する。 Specifically, a rule for conversion to a predetermined data format is defined for each modality, and converted data (hereinafter also referred to as “entity”) converted according to the definition of the object modality in the information processing apparatus 10 corresponds to the modality. To one or more feature management servers 20 (server devices having a feature extraction unit 202 which is an example of a feature extractor).
 また、一般的なプログラミングの型として扱えない一部のデータに関し、モダリティの定義を規定しておくことで、処理の難しいデータも読み込める上、データの前処理を統一することが可能となる。 Also, by defining modality definitions for some data that cannot be handled as general programming types, it is possible to read difficult data and unify data preprocessing.
 例えば以下のようなデータでの需要が想定される。
(パターン1:形式間の変換/抽出)
・ベクタデータ(通常テキストファイル)を他の画像と同様に扱うためにレンダリング(描画)する
・PDF、オフィス文書等をテキストとして扱うためにテキストを抜き出す
・iCalendar(スケジュールの標準フォーマット)から予定の名前と日程、参加者のみを抽出する
(パターン2:形式の統一)
 手書きデータなどは、各社が異なるフォーマットを提唱しており、標準のフォーマットが存在しない。そういったデータの共通項を取り出し、処理可能な形式に変換する。
(パターン3:特殊なデータの読み込み)
 血圧や心拍数等、まだあまりデジタルで扱われていないデータは、テキスト等の一般的なデータで記述される場合が多い。それらを読み込み、処理しやすい(例えば整数列などの)データ形式に変換する。
(パターン4:データの取得)
・URLから画像を取得する
・IDを入力とし、外部の特定のデータベースからデータを引き出す
For example, demand with the following data is assumed.
(Pattern 1: Conversion / extraction between formats)
・ Rendering (drawing) vector data (normal text file) to treat it like other images ・ Extracting text to handle PDF, office documents, etc. ・ Schedule name from iCalendar (standard schedule format) And schedule, extract only participants (Pattern 2: Unification of format)
Each company has proposed a different format for handwritten data, and there is no standard format. Extract common terms of such data and convert them into a processable format.
(Pattern 3: Reading special data)
Data that is not yet handled digitally, such as blood pressure and heart rate, is often described as general data such as text. They are read and converted to a data format that is easy to process (eg, an integer string).
(Pattern 4: Data acquisition)
・ Acquire images from URLs ・ Use ID as input and extract data from specific external database
 また、複数の特徴空間を扱って検索データベース(以下、検索DB)への登録を行う場合、特徴空間毎にデータを保存すると、同じデータが複数のデータベースに登録されてしまい、冗長である。そこで検索DBとは別に、データを格納し、IDからデータを取り出せるデータベースを用意し、検索DBにはIDのみを格納することが考えられる。この際、ユーザがデータと共に任意にIDを入力するようにすると、以下のような問題が生じる可能性があり、ユーザ側でケアしなければならない。
・同じデータに対して複数のIDを関連付けることが可能であること
・複数のデータに対して同じIDを関連付けることが可能であること
・データを一部の検索DBのみに保存する場合、異なる検索DBに保存されているIDが同一のデータを指す保証がない(検索結果の比較やまとめができない)
・結果として、IDから元のデータを取り出せる保証がされない
In addition, when a plurality of feature spaces are handled and registered in a search database (hereinafter referred to as a search DB), if data is stored for each feature space, the same data is registered in a plurality of databases, which is redundant. Therefore, it is conceivable to prepare a database capable of storing data and extracting data from the ID separately from the search DB, and storing only the ID in the search DB. At this time, if the user arbitrarily inputs the ID together with the data, there is a possibility that the following problem may occur and the user must take care.
-It is possible to associate multiple IDs with the same data-It is possible to associate the same ID with multiple data-Different searches when storing data in only some search DBs There is no guarantee that the IDs stored in the DB point to the same data (search results cannot be compared or summarized)
-As a result, there is no guarantee that the original data can be extracted from the ID.
 そこで、本実施形態では、情報処理装置10において、上記前処理の共通化と共に、登録オブジェクトの変換データに複数の特徴抽出器に共通する一意の(ユニークな)ID(識別情報)を付与すると共に、登録オブジェクトを保存することで上記問題を解決し、複数の特徴空間を扱うシステムの利便性をさらに向上させることを可能とする。 Therefore, in the present embodiment, in the information processing apparatus 10, along with the common pre-processing, a unique (unique) ID (identification information) common to a plurality of feature extractors is given to the conversion data of the registered object. By saving the registered object, it is possible to solve the above problem and further improve the convenience of a system that handles a plurality of feature spaces.
 以上説明したように、本実施形態による情報処理システムでは、オブジェクトのモダリティごとに統一的に変換すると共に、複数の特徴空間に共通する一意のIDでオブジェクトを管理することで、複数の特徴空間を扱うシステムの利便性をより向上させることを可能とする。 As described above, in the information processing system according to the present embodiment, a plurality of feature spaces are created by uniformly converting objects for each modality and managing objects with unique IDs common to a plurality of feature spaces. It is possible to further improve the convenience of the handling system.
 このような本実施形態による情報処理システムに含まれる情報処理装置10および特徴管理サーバ20の構成について、以下説明する。 The configurations of the information processing apparatus 10 and the feature management server 20 included in the information processing system according to this embodiment will be described below.
 <<2.構成>>
  <2-1.情報処理装置10の構成>
 図1に示すように、本実施形態による情報処理装置10は、制御部100、通信部120、出力部130、および記憶部140を有する。情報処理装置10は、例えばユーザに利用されるスマートフォン、タブレット端末、またはPC等のローカル端末である。
<< 2. Configuration >>
<2-1. Configuration of Information Processing Device 10>
As illustrated in FIG. 1, the information processing apparatus 10 according to the present embodiment includes a control unit 100, a communication unit 120, an output unit 130, and a storage unit 140. The information processing apparatus 10 is a local terminal such as a smartphone, a tablet terminal, or a PC used by a user, for example.
 (制御部100)
 制御部100は、演算処理装置および制御装置として機能し、各種プログラムに従って情報処理装置10内の動作全般を制御する。制御部100は、例えばCPU(Central Processing Unit)、マイクロプロセッサ等の電子回路によって実現される。また、制御部100は、使用するプログラムや演算パラメータ等を記憶するROM(Read Only Memory)、及び適宜変化するパラメータ等を一時記憶するRAM(Random Access Memory)を含んでいてもよい。
(Control unit 100)
The control unit 100 functions as an arithmetic processing unit and a control unit, and controls overall operations in the information processing apparatus 10 according to various programs. The control unit 100 is realized by an electronic circuit such as a CPU (Central Processing Unit) or a microprocessor, for example. The control unit 100 may include a ROM (Read Only Memory) that stores programs to be used, calculation parameters, and the like, and a RAM (Random Access Memory) that temporarily stores parameters that change as appropriate.
 また、本実施形態による制御部100は、特徴空間管理部101(Feature Space Manager)およびモダリティ管理部102(Modality Manager)としても機能する。 The control unit 100 according to the present embodiment also functions as a feature space management unit 101 (Feature Space Manager) and a modality management unit 102 (Modality Manager).
 特徴空間管理部101は、扱う特徴空間の管理(1以上の特徴管理サーバ20のIDの取得など)や、オブジェクトの登録処理、検索処理等を行う。ここで、図2に、本実施形態による情報処理システムの主な機能ブロック(特徴空間管理部101、モダリティ管理部102、および特徴管理部201)における処理内容の一例を示す。図2に示すように、例えば本実施形態による特徴空間管理部101は、下記のような処理を行い得る。
・Get Manager(space ID: string): Feature Manager
・Register Manager(manager: Feature Manager)
・Get Vector(space ID: string, obj: object, modality: string): vector
・Register Object(obj: object, modality: string, space ID: string=ANY)
・Search(query: object, query Modality: string, target Modality: string=ANY): Search Result
The feature space management unit 101 performs feature space management (acquisition of one or more feature management server 20 IDs), object registration processing, search processing, and the like. Here, FIG. 2 shows an example of processing contents in the main functional blocks (the feature space management unit 101, the modality management unit 102, and the feature management unit 201) of the information processing system according to the present embodiment. As shown in FIG. 2, for example, the feature space management unit 101 according to the present embodiment can perform the following processing.
・ Get Manager (space ID: string): Feature Manager
・ Register Manager (manager: Feature Manager)
・ Get Vector (space ID: string, obj: object, modality: string): vector
・ Register Object (obj: object, modality: string, space ID: string = ANY)
・ Search (query: object, query Modality: string, target Modality: string = ANY): Search Result
 より具体的には、例えば特徴空間管理部101は、登録要求情報が入力された際、登録要求情報に含まれるオブジェクトおよびモダリティをモダリティ管理部102に出力し、モダリティ管理部102においてモダリティの定義に従って変換された変換データ(entity)およびIDを取得し、当該entityおよびIDを特徴管理サーバ20に出力する。この際、特徴空間管理部101は、オブジェクトのモダリティに対応する1以上の特徴管理サーバ20に出力し得る。「モダリティに対応する特徴管理サーバ20」とは、当該モダリティを扱うことが可能な特徴管理サーバ20である。特徴空間管理部101は、各特徴空間がどのような情報を扱っているかを特徴管理サーバ20のID(space ID: stringなど)から把握することが可能である。従って、例えば特徴空間管理部101は、modality: "Text"の場合に、各space IDを参照し、"Text"を扱う特徴空間を管理している特徴管理サーバ20を特定することが可能となる。若しくは、特徴空間管理部101は、全ての特徴管理サーバ20に出力するようにしてもよい(この場合、特徴管理サーバ20側で適宜処理可否が判断され得る)。 More specifically, for example, when the registration request information is input, the feature space management unit 101 outputs the object and modality included in the registration request information to the modality management unit 102, and the modality management unit 102 follows the definition of the modality. The converted conversion data (entity) and ID are acquired, and the entity and ID are output to the feature management server 20. At this time, the feature space management unit 101 can output to one or more feature management servers 20 corresponding to the modality of the object. The “feature management server 20 corresponding to the modality” is the feature management server 20 capable of handling the modality. The feature space management unit 101 can grasp what information each feature space handles from the ID (space ID: string, etc.) of the feature management server 20. Therefore, for example, in the case of modality: “Text”, the feature space management unit 101 can refer to each space ID and identify the feature management server 20 that manages the feature space that handles “Text”. . Alternatively, the feature space management unit 101 may output the information to all the feature management servers 20 (in this case, the feature management server 20 may determine whether or not processing can be appropriately performed).
 また、特徴空間管理部101は、検索要求情報が入力された際、検索要求に含まれるオブジェクトおよびモダリティをモダリティ管理部102に出力し、モダリティ管理部102においてモダリティの定義に従って変換された変換データ(entity)を取得し、当該entityを特徴管理サーバ20に出力する。この際、特徴空間管理部101は、オブジェクトのモダリティに対応する1以上の特徴管理サーバ20に出力し得る。また、特徴空間管理部101は、検索要求に検索条件としてspace IDが含まれていた場合、指定されたspace IDに対応する特徴管理サーバ20に出力するようにしてもよい。若しくは、特徴空間管理部101は、全ての特徴管理サーバ20に出力するようにしてもよい(この場合、特徴管理サーバ20側で適宜処理可否が判断され得る)。 Further, when the search request information is input, the feature space management unit 101 outputs the object and modality included in the search request to the modality management unit 102, and the converted data (converted by the modality management unit 102 according to the definition of the modality ( entity), and outputs the entity to the feature management server 20. At this time, the feature space management unit 101 can output to one or more feature management servers 20 corresponding to the modality of the object. The feature space management unit 101 may output the feature request to the feature management server 20 corresponding to the designated space ID when the search request includes a space ID as a search condition. Alternatively, the feature space management unit 101 may output the information to all the feature management servers 20 (in this case, the feature management server 20 may determine whether or not processing can be appropriately performed).
 そして、特徴空間管理部101は、特徴管理サーバ20において検索された1以上のIDに基づいて、記憶部140から対応するオブジェクト(すなわち元のデータ)を取り出し、検索結果として検索要求元に出力する。なお、検索条件には、追加条件としてフィルター情報がさらに含まれていてもよい。フィルター情報としては、例えば、検索DB(検索対象の特徴空間に相当)の指定や、検索数の指定等が挙げられる。特徴空間管理部101は、例えば検索数が指定されている場合、各検索結果の類似度(各検索結果の、検索オブジェクトの変換データ(entity)の特徴量との類似度合いを示す類似度)に基づいて、上位所定数の検索結果を検索要求元に出力するようにしてもよい。 Then, the feature space management unit 101 extracts a corresponding object (that is, original data) from the storage unit 140 based on one or more IDs searched in the feature management server 20, and outputs them as search results to the search request source. . Note that the search condition may further include filter information as an additional condition. Examples of the filter information include designation of a search DB (corresponding to a feature space to be searched), designation of the number of searches, and the like. For example, when the number of searches is specified, the feature space management unit 101 sets the similarity of each search result (similarity indicating the similarity between each search result and the feature amount of the conversion data (entity) of the search object). Based on this, a predetermined upper number of search results may be output to the search request source.
 モダリティ管理部102は、モダリティの管理を行う。例えば図2にも示すように、モダリティ管理部102は下記のような処理を行い得る。
・Get Modalities(): string[]
・Register Modality(modality: string, definer: Modality Definer)
・Create(obj: object, modality: string): entity
The modality management unit 102 manages modalities. For example, as shown in FIG. 2, the modality management unit 102 can perform the following processing.
・ Get Modalities (): string []
・ Register Modality (modality: string, definer: Modality Definer)
Create (obj: object, modality: string): entity
 より具体的には、例えばモダリティ管理部102は、モダリティの定義(データ)を登録したり、モダリティ定義部103により、特徴空間管理部101から入力されたオブジェクト(登録オブジェクト)を当該オブジェクトのモダリティの定義に従って所定のデータ形式に変換し、変換データ(entity)を生成したりする。entityは、特徴抽出部202に直接渡すことのできる、整形されたデータである。また、モダリティ定義部103は、生成したentityに、複数の特徴空間に共通する一意のIDを付与し、entityおよびIDを、特徴空間管理部101に出力する。さらに、モダリティ定義部103は、オブジェクト(登録オブジェクト)と、当該登録オブジェクトのentityに付与したIDを関連付けて記憶部140に記憶する。かかるIDは、同じモダリティを扱う複数の特徴空間にまたがって一意な文字列である。また、同じentityに対しては同じ値を返すよう、ハッシュ値等を用いてもよい。 More specifically, for example, the modality management unit 102 registers the modality definition (data), or the modality definition unit 103 inputs the object (registered object) input from the feature space management unit 101 to the modality of the object. The data is converted into a predetermined data format according to the definition, and converted data (entity) is generated. The entity is shaped data that can be directly passed to the feature extraction unit 202. In addition, the modality definition unit 103 assigns a unique ID common to the plurality of feature spaces to the generated entity, and outputs the entity and ID to the feature space management unit 101. Further, the modality definition unit 103 stores the object (registered object) and the ID assigned to the entity of the registered object in the storage unit 140 in association with each other. Such an ID is a unique character string across a plurality of feature spaces that handle the same modality. A hash value or the like may be used so that the same value is returned for the same entity.
 また、モダリティの定義のデータは、入力データ(obj)として受け取れる形式の定義データ(例えば、ファイル名(string)、OpenCVのMat形式)(複数可)と、出力データ(entity)の形式の定義データ(1つのみ)(例えば、char[3][256][256])とを有する。モダリティの定義データは、例えば、データ形式毎や、上述したパターン毎(形式間の変換、形式の統一、特殊なデータの読み込み等)に存在し、記憶部140に記憶されている。モダリティ管理部102は、かかるモダリティの定義データを用いて、登録オブジェクト(入力データ)を変換し、変換データ(entity)を出力する。また、モダリティ定義部103は、IDに対応するデータを必要に応じて保存するか、保存されていることを確認する機能と、IDに対応するデータを取り出す機能と、IDに対応するデータを削除する機能を有する。モダリティ定義部103は、モダリティ毎に存在する。 Modality definition data includes definition data in a format that can be received as input data (obj) (for example, file name (string), OpenCV Mat format) (multiple) and output data (entity) format definition data (Only one) (for example, char [3] [256] [256]). Modality definition data exists, for example, for each data format or for each pattern described above (for example, conversion between formats, unification of formats, reading of special data, etc.), and is stored in the storage unit 140. The modality management unit 102 converts the registration object (input data) using the definition data of the modality, and outputs the conversion data (entity). Also, the modality definition unit 103 saves the data corresponding to the ID as necessary or confirms that the data is saved, the function of retrieving the data corresponding to the ID, and deletes the data corresponding to the ID. Has the function of The modality definition unit 103 exists for each modality.
 (入力部110)
 入力部110は、ユーザによる操作指示を受け付ける操作入力部や、ユーザのよる音声指示を受け付ける音声入力部など、ユーザによる指示内容を受け付ける機能を有し、その指示内容を制御部100に出力する。操作入力部は、タッチセンサ、圧力センサ、若しくは近接センサであってもよい。あるいは、入力部110は、ボタン、スイッチ、およびレバーなど、物理的構成であってもよい。
(Input unit 110)
The input unit 110 has a function of receiving user instruction content such as an operation input unit that receives an operation instruction from the user and a voice input unit that receives a voice instruction from the user, and outputs the instruction content to the control unit 100. The operation input unit may be a touch sensor, a pressure sensor, or a proximity sensor. Alternatively, the input unit 110 may be a physical configuration such as a button, a switch, and a lever.
 (通信部120)
 通信部120は、有線または無線により外部装置と接続し、外部装置とデータの送受信を行う。例えば通信部120は、有線/無線LAN(Local Area Network)、またはWi-Fi(登録商標)、Bluetooth(登録商標)、携帯通信網(LTE:Long Term Evolution、3G(第3世代の移動体通信方式))等により、ネットワーク(不図示)に接続し、ネットワークを介して特徴管理サーバ20とデータの送受信を行い得る。
(Communication unit 120)
The communication unit 120 is connected to an external device by wire or wireless, and transmits / receives data to / from the external device. For example, the communication unit 120 may be a wired / wireless local area network (LAN), Wi-Fi (registered trademark), Bluetooth (registered trademark), a mobile communication network (LTE: Long Term Evolution, 3G (third generation mobile communication)). For example, it is possible to connect to a network (not shown) and transmit / receive data to / from the feature management server 20 via the network.
 (出力部130)
 出力部130は、表示部および音声出力部等、ユーザへの情報提示(出力)を行う機能を有する。例えば出力部130は、制御部100の制御に従って、検索画面を出力したり、検索結果を出力したりする。
(Output unit 130)
The output unit 130 has a function of presenting (outputting) information to the user, such as a display unit and an audio output unit. For example, the output unit 130 outputs a search screen or outputs a search result under the control of the control unit 100.
 (記憶部140)
 記憶部140は、制御部100の処理に用いられるプログラムや演算パラメータ等を記憶するROM(Read Only Memory)、および適宜変化するパラメータ等を一時記憶するRAM(Random Access Memory)により実現される。
(Storage unit 140)
The storage unit 140 is realized by a ROM (Read Only Memory) that stores programs used in the processing of the control unit 100, calculation parameters, and the like, and a RAM (Random Access Memory) that temporarily stores parameters that change as appropriate.
 例えば、記憶部140には、特徴空間の管理情報、モダリティの定義情報、および、一意のIDが付与されたオブジェクト(実体データ)が格納される。 For example, the storage unit 140 stores feature space management information, modality definition information, and an object (substance data) assigned with a unique ID.
 以上、本実施形態による情報処理装置10の構成について具体的に説明した。なお情報処理装置10の構成は、図1に示す例に限定されない。例えば、情報処理装置10の制御部100による各処理を複数の装置で実行するようにしてもよいし、ネットワーク上のサーバで実行するようにしてもよい。 The configuration of the information processing apparatus 10 according to the present embodiment has been specifically described above. The configuration of the information processing apparatus 10 is not limited to the example illustrated in FIG. For example, each process by the control unit 100 of the information processing apparatus 10 may be executed by a plurality of apparatuses or may be executed by a server on the network.
  <2-2.特徴管理サーバ20の構成>
 図2に示すように、特徴管理サーバ20は、制御部200、通信部210、および特徴量データベース220を有する。なお、本実施形態において、特徴管理サーバ20は1の特徴抽出部202を有するため、各特徴管理サーバ20はそれぞれ1の特徴空間を管理していると言えるが、本開示は、これに限定されず、例えば特徴管理サーバ20が複数の特徴抽出部202を有していれば、複数の特徴空間を管理することも可能である。
<2-2. Configuration of Feature Management Server 20>
As illustrated in FIG. 2, the feature management server 20 includes a control unit 200, a communication unit 210, and a feature amount database 220. In this embodiment, since the feature management server 20 has one feature extraction unit 202, it can be said that each feature management server 20 manages one feature space, but the present disclosure is limited to this. For example, if the feature management server 20 includes a plurality of feature extraction units 202, a plurality of feature spaces can be managed.
 (制御部200)
 制御部200は、演算処理装置および制御装置として機能し、各種プログラムに従って特徴管理サーバ20内の動作全般を制御する。制御部200は、例えばCPU(Central Processing Unit)、マイクロプロセッサ等の電子回路によって実現される。また、制御部200は、使用するプログラムや演算パラメータ等を記憶するROM(Read Only Memory)、及び適宜変化するパラメータ等を一時記憶するRAM(Random Access Memory)を含んでいてもよい。
(Control unit 200)
The control unit 200 functions as an arithmetic processing unit and a control unit, and controls the overall operation in the feature management server 20 according to various programs. The control unit 200 is realized by an electronic circuit such as a CPU (Central Processing Unit) or a microprocessor, for example. The control unit 200 may include a ROM (Read Only Memory) that stores programs to be used, calculation parameters, and the like, and a RAM (Random Access Memory) that temporarily stores parameters that change as appropriate.
 また、本実施形態による制御部200は、特徴管理部201としても機能する。特徴管理部201は、情報処理装置10から送信されたentityに対し、特徴抽出部202により特徴抽出(例えば、N次元ベクトルへの代替処理)を行い、抽出した特徴量を、当該entityのIDに関連付けて、特徴量データベース220へ登録する処理を行う。特徴量抽出のアルゴリズムは既存の技術を用いることが可能であり、ここでは特に限定しない。また、特徴抽出部202は、異なるモダリティを統一的に扱うことが可能である。特徴抽出部202で抽出した特徴量は、特徴量データベース220に登録されるが、ここで、特徴量データベース220は、モダリティ毎に存在する。特徴管理部201は、特徴抽出部202で抽出した特徴量を、元データ(情報処理装置10から送信されたentity)のモダリティに対応する特徴量データベース220に、上記IDと関連付けて登録する。例えば、色特徴という視点で異なるモダリティ(例えば手書き(Strokes)と画像(Image)等)から特徴量を抽出することができる特徴抽出部202a(色特徴の特徴空間a)が存在したとする。この場合、特徴抽出部202aにより抽出された特徴量は、手書きデータから抽出した場合は手書きデータに対応する特徴量データベース220-1に格納され、画像データから抽出した場合は画像データに対応する特徴量データベース220-2に格納される。 In addition, the control unit 200 according to the present embodiment also functions as the feature management unit 201. The feature management unit 201 performs feature extraction (for example, substitution processing for an N-dimensional vector) on the entity transmitted from the information processing apparatus 10 by using the feature extraction unit 202, and sets the extracted feature amount as the ID of the entity. A process of registering in the feature amount database 220 in association with each other is performed. An existing technique can be used for the feature quantity extraction algorithm, and is not particularly limited here. In addition, the feature extraction unit 202 can handle different modalities in a unified manner. The feature quantity extracted by the feature extraction unit 202 is registered in the feature quantity database 220. Here, the feature quantity database 220 exists for each modality. The feature management unit 201 registers the feature amount extracted by the feature extraction unit 202 in the feature amount database 220 corresponding to the modality of the original data (entity transmitted from the information processing apparatus 10) in association with the ID. For example, it is assumed that there is a feature extraction unit 202a (color feature feature space a) that can extract feature quantities from different modalities (for example, handwriting (Strokes) and image (Image)) from the viewpoint of color features. In this case, the feature amount extracted by the feature extraction unit 202a is stored in the feature amount database 220-1 corresponding to the handwritten data when extracted from the handwritten data, and the feature corresponding to the image data when extracted from the image data. It is stored in the quantity database 220-2.
 また、特徴管理部201は、特徴空間管理部101からの要求に応じて、特徴空間を用いた検索処理(類似検索)を行うことも可能である。特徴量データベース220はモダリティ毎に存在するため、特徴管理部201は、ターゲットモダリティ(検索対象のモダリティ)に対応する特徴量データベース220を用いて類似検索を行えばよい。例えば特徴管理部201は、図2に示すように、下記のような処理を行い得る。
・Get Space ID(): string
・Register Database(modality: string, database: Feature Database)
・Get Vector(entity: object, modality: string): vector
・Add(id: string, entity: object, modality: string)
・Most Similar(query: object, modality: string, target Modality: string): Search Result[]
The feature management unit 201 can also perform a search process (similarity search) using a feature space in response to a request from the feature space management unit 101. Since the feature quantity database 220 exists for each modality, the feature management unit 201 may perform a similarity search using the feature quantity database 220 corresponding to the target modality (search target modality). For example, the feature management unit 201 can perform the following processing as shown in FIG.
・ Get Space ID (): string
・ Register Database (modality: string, database: Feature Database)
・ Get Vector (entity: object, modality: string): vector
Add (id: string, entity: object, modality: string)
・ Most Similar (query: object, modality: string, target Modality: string): Search Result []
 (通信部210)
 通信部210は、有線または無線により外部装置と接続し、外部装置とデータの送受信を行う。例えば通信部210は、有線/無線LAN(Local Area Network)、またはWi-Fi(登録商標)、Bluetooth(登録商標)、携帯通信網(LTE:Long Term Evolution、3G(第3世代の移動体通信方式))等により、ネットワーク(不図示)に接続し、ネットワークを介して情報処理装置10とデータの送受信を行い得る。
(Communication unit 210)
The communication unit 210 is connected to an external device by wire or wireless, and transmits / receives data to / from the external device. For example, the communication unit 210 may be a wired / wireless local area network (LAN), Wi-Fi (registered trademark), Bluetooth (registered trademark), a mobile communication network (LTE: Long Term Evolution, 3G (third generation mobile communication). By connecting to a network (not shown) and transmitting / receiving data to / from the information processing apparatus 10 via the network.
 (特徴量データベース220)
 特徴量データベース220は、特徴抽出部202により抽出された特徴量を蓄積する。各特徴量には、モダリティ管理部102により付与された一意のIDが関連付けられる。特徴量データベース220は、上述したように、モダリティ毎に存在する。
(Feature database 220)
The feature quantity database 220 accumulates the feature quantities extracted by the feature extraction unit 202. Each feature amount is associated with a unique ID assigned by the modality management unit 102. As described above, the feature amount database 220 exists for each modality.
 また、特徴量データベース220は、特徴管理サーバ20が有する記憶部(不図示)に記憶される。特徴管理サーバ20の記憶部は、制御部200の処理に用いられるプログラムや演算パラメータ等を記憶するROM、および適宜変化するパラメータ等を一時記憶するRAMにより実現される。 The feature amount database 220 is stored in a storage unit (not shown) included in the feature management server 20. The storage unit of the feature management server 20 is realized by a ROM that stores programs and calculation parameters used for the processing of the control unit 200, and a RAM that temporarily stores parameters that change as appropriate.
 以上、本実施形態による特徴管理サーバ20の構成について具体的に説明した。なお図1に示す特徴管理サーバ20の構成は一例であって、本実施形態はこれに限定されない。例えば特徴管理サーバ20の少なくとも一部の構成が外部装置にあってもよい。 The configuration of the feature management server 20 according to the present embodiment has been specifically described above. The configuration of the feature management server 20 shown in FIG. 1 is an example, and the present embodiment is not limited to this. For example, at least a part of the configuration of the feature management server 20 may be in an external device.
 ここで、図3に、本実施例による情報処理システムの他の構成例の一例を示す。図3に示すように、例えば特徴抽出部240と特徴量データベース250を別のサーバ(特徴管理サーバ24およびデータベースサーバ25)でそれぞれ管理するようにしてもよい。 Here, FIG. 3 shows an example of another configuration example of the information processing system according to this embodiment. As shown in FIG. 3, for example, the feature extraction unit 240 and the feature quantity database 250 may be managed by separate servers (the feature management server 24 and the database server 25), respectively.
 <<3.動作処理>>
 続いて、本実施形態による情報処理システムの動作処理について図面を用いて具体的に説明する。
<< 3. Action processing >>
Next, the operation processing of the information processing system according to the present embodiment will be specifically described with reference to the drawings.
  <3-1.登録処理>
 図4は、本実施形態による情報処理システムにおけるオブジェクトの登録処理の流れの一例を示すシーケンス図である。
<3-1. Registration process>
FIG. 4 is a sequence diagram illustrating an example of a flow of object registration processing in the information processing system according to the present embodiment.
 図4に示すように、まず、情報処理装置10の特徴空間管理部101は、ユーザの操作入力等に基づいて登録要求を取得すると(ステップS103)、登録要求に含まれるオブジェクト(obj)と当該オブジェクトのモダリティ(mdl)の情報と共に、モダリティ管理部102に対して、(変換データ(entity)の)生成依頼を行う(ステップS106)。 As illustrated in FIG. 4, first, the feature space management unit 101 of the information processing apparatus 10 acquires a registration request based on a user operation input or the like (step S103), and the object (obj) included in the registration request Along with the information on the modality (mdl) of the object, a request for generation (conversion data (entity)) is made to the modality management unit 102 (step S106).
 次に、モダリティ管理部102は、モダリティ定義部103により、変換データ(entity)の生成、一意のIDの付与、およびobjと一意のIDの保存処理を行う(ステップS109)。具体的には、モダリティ定義部103は、モダリティの定義に従って、オブジェクトを所定の形式のデータに変換する処理(共通化した前処理)を行う。処理の具体例として、例えば以下のような例が挙げられる。
 ・静止画の場合:JPEGデータをchar[3][256][256](多次元配列)に変換し、平滑化処理を実施する。
 ・音声の場合:mp3データをshort型の任意長配列として読み取る。
 ・テキストの場合:HTMLタグを除去し、全ての大文字を小文字に変換する(形式は変換しない)。
 ・手書きの場合:点列データを読み取り、char[3][256][256]の黒画像に太さ3の白線で描画する。
Next, the modality management unit 102 performs conversion data (entity) generation, unique ID assignment, and obj and unique ID storage processing by the modality definition unit 103 (step S109). Specifically, the modality definition unit 103 performs processing (common preprocessing) for converting an object into data of a predetermined format in accordance with the definition of the modality. Specific examples of the processing include the following examples.
In the case of a still image: JPEG data is converted into char [3] [256] [256] (multidimensional array) and smoothing processing is performed.
・ For voice: Read mp3 data as short type arbitrary length array.
For text: Remove HTML tags and convert all uppercase letters to lowercase (no format).
・ Handwriting: Read the point sequence data and draw it with a white line of thickness 3 on a black image of char [3] [256] [256].
 次いで、特徴空間管理部101は、モダリティ管理部102から、少なくともIDおよびentityを取得する(ステップS112)。また、モダリティ管理部102からは、IDとobjを保存した旨が通知されてもよい。 Next, the feature space management unit 101 acquires at least an ID and an entity from the modality management unit 102 (step S112). Further, the modality management unit 102 may notify that the ID and obj are saved.
 次に、特徴空間管理部101は、取得したIDおよびentityに基づいて、対応する全ての特徴空間(Feature Space)に対してデータの追加(登録)要求を出力する(ステップS115)。追加要求には、ID、entity、モダリティ(mdl)が含まれる。対応する特徴空間とは、当該entityのモダリティを扱い得る特徴空間(特徴管理サーバ20)である。なお、当該entityのモダリティを扱い得る特徴空間が複数ある場合は、特徴空間毎に、ステップS115~S121に示す処理を繰り返す。 Next, the feature space management unit 101 outputs a data addition (registration) request to all corresponding feature spaces (Feature Space) based on the acquired ID and entity (step S115). The addition request includes ID, entity, and modality (mdl). The corresponding feature space is a feature space (feature management server 20) that can handle the modality of the entity. When there are a plurality of feature spaces that can handle the modality of the entity, the processes shown in steps S115 to S121 are repeated for each feature space.
 次いで、特徴管理部201は、特徴抽出部202により、特徴量の抽出を行う(ステップS118)。 Next, the feature management unit 201 extracts feature amounts using the feature extraction unit 202 (step S118).
 そして、特徴管理部201は、抽出された特徴量を、上記取得した一意のIDと共に、特徴量データベース220に追加(登録)する(ステップS121)。この際、特徴管理部201は、抽出元のentityのモダリティに対応する特徴量データベース220に登録する。 Then, the feature management unit 201 adds (registers) the extracted feature quantity to the feature quantity database 220 together with the acquired unique ID (step S121). At this time, the feature management unit 201 registers the feature amount database 220 corresponding to the modality of the extraction source entity.
 以上、本実施形態による登録処理について具体的に説明した。このように各特徴空間にデータを入力する前に、モダリティ管理部102において、モダリティ毎に所定の変換処理を行うことで、同じデータを特徴抽出器毎に異なるプロセスでそれぞれ前処理を行うといった手間が省け、複数の特徴空間を扱うシステムの利便性を向上させることができる。また、モダリティ管理部102と特徴抽出部202を個別に管理することで、各機能の責任が軽くなる(例えばエラー時の原因を特定し易くなる)。 The registration process according to this embodiment has been specifically described above. In this way, before inputting data to each feature space, the modality management unit 102 performs a predetermined conversion process for each modality so that the same data is preprocessed by a different process for each feature extractor. Therefore, the convenience of a system that handles a plurality of feature spaces can be improved. Further, managing the modality management unit 102 and the feature extraction unit 202 individually reduces the responsibility of each function (for example, it becomes easier to identify the cause at the time of an error).
 また、例えば、IDと特徴ベクトルを保管できる一般的なデータベースを提供すれば、特徴抽出器のみの開発で検索システムが完成し、システムの可用性も上がる。また、検索DB(すなわち、特徴量データベース220)では、IDと特徴量のみ登録し、元データ(データの実体)はモダリティ定義部103により別で管理するため、同じデータを複数のデータベースに登録して冗長となることを回避することができる。また、複数の特徴空間にまたがって一意なIDを同じデータに付与することで、同じデータに対して複数のIDを関連付けてしまうことや、複数のデータに対して同じIDを関連付けてしまうこと等を回避することができる。 Also, for example, if a general database that can store IDs and feature vectors is provided, the search system is completed by developing only the feature extractor, and the availability of the system increases. In the search DB (that is, the feature database 220), only the ID and the feature are registered, and the original data (data entity) is separately managed by the modality definition unit 103. Therefore, the same data is registered in a plurality of databases. Can be avoided. In addition, by assigning unique IDs across multiple feature spaces to the same data, multiple IDs can be associated with the same data, the same ID can be associated with multiple data, etc. Can be avoided.
 また、特徴管理部201は、モダリティ管理部102で元データが保存されている場合のみ特徴量データベース220に登録するようにしてもよい。これにより、後述する検索結果取得の際に特徴空間管理部101がIDから元のデータを取り出せる保証がなされる。 In addition, the feature management unit 201 may register the feature data in the feature amount database 220 only when the original data is stored in the modality management unit 102. This ensures that the feature space management unit 101 can extract the original data from the ID when acquiring a search result to be described later.
 また、所定のデータ形式への前処理が別で管理されるため、特徴空間(検索DBシステム)の開発側としては、入力形式を気にすることなく特徴抽出部202の開発を行うことができる。 In addition, since the preprocessing to a predetermined data format is managed separately, the feature space (search DB system) development side can develop the feature extraction unit 202 without worrying about the input format. .
 また、本実施形態による登録処理は、図4に示す例に限定されない。例えば、上記ステップS115では、モダリティに対応する特徴空間(特徴管理サーバ20)に追加指示を行う旨を説明したが、本実施形態はこれに限定されず、特徴空間管理部101は、全ての特徴空間(特徴管理サーバ20)に追加指示を行ってもよい。この場合、特徴空間(特徴管理サーバ20)側で、モダリティに基づき、処理可能なentityであるか否かを判断し得る。 Further, the registration process according to the present embodiment is not limited to the example shown in FIG. For example, in step S115 described above, it has been described that the addition instruction is given to the feature space (feature management server 20) corresponding to the modality. However, the present embodiment is not limited to this, and the feature space management unit 101 can execute all the features. An additional instruction may be given to the space (feature management server 20). In this case, the feature space (feature management server 20) can determine whether the entity is a processable entity based on the modality.
  <3-2.検索処理>
 続いて、上述したように特徴空間を構築する本実施形態による情報処理システムにおける検索処理について、図5を参照して説明する。図5は、本実施形態による情報処理システムにおける検索処理の流れの一例を示すシーケンス図である。
<3-2. Search process>
Next, search processing in the information processing system according to the present embodiment that constructs a feature space as described above will be described with reference to FIG. FIG. 5 is a sequence diagram showing an example of the flow of search processing in the information processing system according to the present embodiment.
 図5に示すように、まず、情報処理装置10の特徴空間管理部101は、ユーザの操作入力等に基づいて検索要求を取得する(ステップS133)。検索要求には、オブジェクト(obj)と、当該オブジェクトのモダリティ(mdl1)と、検索対象のモダリティを示すターゲットモダリティ(mdl2)とが含まれる。 As shown in FIG. 5, first, the feature space management unit 101 of the information processing apparatus 10 acquires a search request based on a user operation input or the like (step S133). The search request includes an object (obj), a modality (mdl1) of the object, and a target modality (mdl2) indicating the modality to be searched.
 次いで、特徴空間管理部101は、検索要求に含まれるオブジェクト(obj)と当該オブジェクトのモダリティ(mdl1)の情報と共に、モダリティ管理部102に対して、(変換データ(entity)の)生成依頼を行う(ステップS136)。 Next, the feature space management unit 101 makes a generation request (conversion data (entity)) to the modality management unit 102 together with information on the object (obj) and the modality (mdl1) of the object included in the search request. (Step S136).
 次に、モダリティ管理部102は、モダリティ定義部103により、変換データ(entity)の生成、および一意のIDの付与を行う(ステップS139)。具体的には、モダリティ定義部103は、モダリティの定義に従って、オブジェクトを所定の形式のデータに変換する処理(共通化した前処理)を行う。 Next, the modality management unit 102 uses the modality definition unit 103 to generate conversion data (entity) and assign a unique ID (step S139). Specifically, the modality definition unit 103 performs processing (common preprocessing) for converting an object into data of a predetermined format in accordance with the definition of the modality.
 次いで、特徴空間管理部101は、モダリティ管理部102から、IDおよびentityを取得する(ステップS142)。 Next, the feature space management unit 101 acquires the ID and entity from the modality management unit 102 (step S142).
 次に、特徴空間管理部101は、取得したentityに基づいて、対応する全ての特徴空間(Feature Space)に対して検索要求を出力する(ステップS145)。検索要求には、entity、mdl1(元データのモダリティ)、mdl2(ターゲットモダリティ)が含まれる。対応する特徴空間とは、mdl1およびmdl2を扱い得る特徴空間(特徴管理サーバ20)である。なお、mdl1およびmdl2を扱い得る特徴空間が複数ある場合は、特徴空間毎に、ステップS145~S157に示す処理を繰り返す。 Next, the feature space management unit 101 outputs a search request to all the corresponding feature spaces (Feature Space) based on the acquired entity (step S145). The search request includes entity, mdl1 (original data modality), and mdl2 (target modality). The corresponding feature space is a feature space (feature management server 20) that can handle mdl1 and mdl2. When there are a plurality of feature spaces that can handle mdl1 and mdl2, the processing shown in steps S145 to S157 is repeated for each feature space.
 次いで、特徴管理部201は、特徴抽出部202により、特徴量の抽出を行う(ステップS148)。 Next, the feature management unit 201 uses the feature extraction unit 202 to extract feature amounts (step S148).
 続いて、特徴管理部201は、抽出された特徴量に基づいて、特徴量データベース220から、類似している特徴量の検索を行う(ステップS151)。この際、特徴管理部201は、要求されたターゲットモダリティ(mdl2)に対応する特徴量データベース220から検索する。特徴量データベース220では、特徴量に、上記一意のIDが関連付けられており、特徴管理部201は、検索要求されたentityの特徴量と類似する特徴量を特徴量データベース220から検索し、類似する特徴量に関連付けられたIDと、当該特徴量の類似度:sim(検索要求されたentityの特徴量との類似度であって、例えばN次元ベクトルの距離)を取得する。なお、ターゲットモダリティ(mdl2)が複数ある場合は、特徴量データベース220毎に、ステップS151~S154に示す処理を繰り返す。 Subsequently, the feature management unit 201 searches for a similar feature amount from the feature amount database 220 based on the extracted feature amount (step S151). At this time, the feature management unit 201 searches the feature amount database 220 corresponding to the requested target modality (mdl2). In the feature quantity database 220, the unique ID is associated with the feature quantity, and the feature management unit 201 searches the feature quantity database 220 for a feature quantity similar to the feature quantity of the requested entity, and is similar. The ID associated with the feature quantity and the similarity of the feature quantity: sim (similarity with the feature quantity of the requested entity, for example, the distance of the N-dimensional vector) is acquired. If there are a plurality of target modalities (mdl2), the processes shown in steps S151 to S154 are repeated for each feature quantity database 220.
 次に、特徴空間管理部101は、特徴管理部201から、検索結果(検索した特徴量のID、検索したモダリティ:mdl、および検索した特徴量の類似度:sim)を取得する(ステップS157)。検索結果には、複数のID、mdl、およびsimが含まれていてもよい。検索結果として単数が求められている場合、特徴空間管理部101は、例えば類似度が最も高い特徴量のIDを特定する。検索結果として所定数が求められている場合、特徴空間管理部101は、例えば類似度に基づいて上位所定数の特徴量のIDを特定する。 Next, the feature space management unit 101 acquires a search result (searched feature value ID, searched modality: mdl, and searched feature value similarity: sim) from the feature management unit 201 (step S157). . The search result may include a plurality of IDs, mdl, and sim. When a singular number is obtained as a search result, the feature space management unit 101 specifies, for example, an ID of a feature amount having the highest similarity. When the predetermined number is obtained as the search result, the feature space management unit 101 specifies the ID of the upper predetermined number of feature amounts based on, for example, the similarity.
 次いで、特徴空間管理部101は、特定したIDおよび対応するモダリティの情報と共に、モダリティ管理部102に対して元データの要求を行う(ステップS160)。 Next, the feature space management unit 101 makes a request for original data to the modality management unit 102 together with the identified ID and the corresponding modality information (step S160).
 次に、モダリティ管理部102は、モダリティ定義部103により、IDに関連付けられた元データ(すなわち、オブジェクト)を取得し(ステップS163)、特徴空間管理部101に出力する(ステップS166)。かかるステップS160~S166に示す処理は、出力する検索結果数分行い得る。元データの取得ができなかった場合、特徴空間管理部101は、レコード(ID,mdl,sim)の削除を行う。 Next, the modality management unit 102 acquires the original data (that is, the object) associated with the ID by the modality definition unit 103 (step S163) and outputs it to the feature space management unit 101 (step S166). The processing shown in steps S160 to S166 can be performed for the number of search results to be output. When the original data cannot be acquired, the feature space management unit 101 deletes the record (ID, mdl, sim).
 そして、特徴空間管理部101は、検索結果(オブジェクト、モダリティ、および類似度)を、検索要求元に出力する(ステップS169)。例えば特徴空間管理部101は、検索結果を示す画面を、出力部130で表示し、ユーザに提示してもよい。 Then, the feature space management unit 101 outputs the search result (object, modality, and similarity) to the search request source (step S169). For example, the feature space management unit 101 may display a screen showing the search result on the output unit 130 and present it to the user.
 以上、本実施形態による検索処理について具体的に説明した。このように、本実施形態により構築した特徴空間を、異なるモダリティを扱う複数の特徴空間における横断的な検索に利用することができる。 The search processing according to this embodiment has been specifically described above. Thus, the feature space constructed according to the present embodiment can be used for cross-sectional search in a plurality of feature spaces that handle different modalities.
 上記最後の例に示したように、検索に用いる特徴空間(検索DB)を特定してもよい。特定した特徴空間は、space IDとして、上記ステップS113の検索要求に含まれる。例えば、「○○社が作成した検索DBを利用したい」、「○○検索サイトを利用したい」等が想定される。 As shown in the last example above, the feature space (search DB) used for the search may be specified. The identified feature space is included in the search request in step S113 as a space ID. For example, “I want to use a search DB created by XX company”, “I want to use a XX search site”, or the like is assumed.
 ここで、図6に、本実施形態における検索画面の一例を示す。図6に示す検索画面30は、例えば出力部130で提示される。ユーザは、検索オブジェクト301を入力し、検索対象を選択し(モダリティに相当。例えば、「写真」、「イラスト」、「書類」等)、何が似ているものを検索したいのかその特徴量を選択し(例えば、「形」、「色」、「意味」等であって、各特徴空間に相当する。例えば、形の特徴に基づいて構築された特徴空間、色の特徴に基づいて構築された特徴空間等である)、検索ボタン302を選択すると、検索結果として取得された検索オブジェクト301に似ているオブジェクトが提示される。例えば検索対象として「イラスト」を選択し、特徴量として「形」(space ID1)、「色」(space ID2)、および「意味」(space ID3)を選択した場合、検索結果として、検索オブジェクト301と、形、色、および/または意味が似ているイラストが取得され(例えば、形の特徴に基づいて構築された特徴空間を扱う特徴管理サーバ20が保有する「モダリティ:イラスト」の特徴量データベース220から検索され)、提示される。特徴量の検索条件のand/orはユーザが任意に選択できるようにしてもよいし、orをデフォルトにしてもよい。 Here, FIG. 6 shows an example of a search screen in the present embodiment. The search screen 30 shown in FIG. 6 is presented by the output unit 130, for example. The user inputs a search object 301 and selects a search target (corresponding to a modality. For example, “photograph”, “illustration”, “document”, etc.), and the feature quantity of what is similar is searched for. Select (for example, “shape”, “color”, “meaning”, etc., and correspond to each feature space. For example, feature space constructed based on shape features, constructed based on color features When the search button 302 is selected, an object similar to the search object 301 acquired as a search result is presented. For example, when “illustration” is selected as the search target and “shape” (space ID1), “color” (space ID2), and “meaning” (space ID3) are selected as the feature quantities, the search object 301 is obtained as the search result. And an illustration having a similar shape, color, and / or meaning (for example, a “modality: illustration” feature quantity database possessed by the feature management server 20 that handles a feature space constructed based on the feature of the shape. 220) and presented. The feature amount search condition and / or may be arbitrarily selected by the user, or or may be set as a default.
 <<4.応用例>>
 続いて、本実施形態による情報処理システムの応用例について説明する。
<< 4. Application example >>
Subsequently, an application example of the information processing system according to the present embodiment will be described.
  <4-1.第1の応用例:モダリティの包含関係の定義>
 まず、第1の応用例として、モダリティの包含関係の定義について説明する。本実施形態によるモダリティ定義部103は、モダリティ同士に親子関係(包含関係)を定義してもよい。具体例として下記が挙げられる。
<4-1. First Application Example: Definition of Modality Inclusion Relationship>
First, as a first application example, the definition of the modality inclusion relationship will be described. The modality definition unit 103 according to the present embodiment may define a parent-child relationship (inclusion relationship) between modalities. Specific examples include the following.
・モダリティ「RGB画像」は、モダリティ「グレースケール画像」を子として持つ
・モダリティ「メール」は、モダリティ「テキスト」、「ユーザ」、「日付」を子として持つ
-Modality "RGB image" has modality "grayscale image" as child-Modality "mail" has modality "text", "user", "date" as children
 本応用例によれば、新規にモダリティを定義する際、子モダリティを定義すれば、容易に既存の特徴空間に組込むことができる。また、新しいモダリティに対しても、複数の特徴抽出器(特徴空間、特徴抽出部202)を組み合わせることによって特徴抽出を行うことが可能となる。 According to this application example, when a new modality is defined, if a child modality is defined, it can be easily incorporated into an existing feature space. In addition, it is possible to perform feature extraction for a new modality by combining a plurality of feature extractors (feature space, feature extraction unit 202).
 例えば、既存モダリティを含むモダリティを定義するユースケースが想定される。より具体的には、“テキスト”というモダリティが既に存在し、“テキスト”を扱う特徴空間A(特徴抽出部202A)があるとする。ここに、テキスト(本文)とユーザ(送信者)を含む“メール”というモダリティと、“メール”を扱える特徴空間B(特徴抽出部202B)を追加した場合を想定する。この場合、第1の効果として、「同時に複数の特徴空間に登録できる」ということが挙げられる。すなわち、メールからはテキストが取得できるため、同じIDとオブジェクトのペアで、特徴空間Bだけでなく、特徴空間Aにも登録できる。すなわち、特徴空間Aには、テキストだけに着目した場合の特徴量が、特徴空間Bには、テキストとユーザに着目した場合の特徴量が格納される。また、他のテキストと横断的に検索が可能となる(この場合、IDは、同じモダリティだけではなく、全てのモダリティにまたがって一意なIDを付与する必要がある)。 For example, use cases that define modalities including existing modalities are assumed. More specifically, it is assumed that a modality “text” already exists and there is a feature space A (feature extraction unit 202A) that handles “text”. Here, it is assumed that a modality of “mail” including text (body) and a user (sender) and a feature space B (feature extraction unit 202B) that can handle “mail” are added. In this case, the first effect is that “can be registered in a plurality of feature spaces at the same time”. That is, since the text can be acquired from the mail, the same ID and object pair can be registered not only in the feature space B but also in the feature space A. That is, the feature space when the focus is on only the text is stored in the feature space A, and the feature space when the focus is on the text and the user is stored in the feature space B. In addition, a search can be performed across other texts (in this case, it is necessary to assign a unique ID across all modalities, not just the same modality).
 また、第2の効果として、「既存の特徴抽出器(特徴抽出部202)を再利用することで、容易に新規の特徴抽出器(特徴抽出部202)を構築できる」ということが挙げられる。すなわち、特徴空間Bは、特徴空間Aを利用してテキストの特徴抽出を行うことができるため、特徴空間Bではユーザの特徴抽出のみを行えばよく、実装が容易となる。また、既存の特徴抽出器をモジュール化して利用することも可能である。図7は、特徴抽出器のモジュール化について説明する図である。 Also, as a second effect, “a new feature extractor (feature extractor 202) can be easily constructed by reusing an existing feature extractor (feature extractor 202)” is mentioned. That is, since the feature space B can extract the features of the text using the feature space A, only the feature extraction of the user needs to be performed in the feature space B, and the implementation becomes easy. It is also possible to modularize existing feature extractors. FIG. 7 is a diagram illustrating modularization of the feature extractor.
 図7左に示すように、例えばメールの特徴抽出を行う際には、メールと包含関係が定義されている文章、およびユーザといったモダリティをそれぞれ扱うことが可能な各特徴抽出器を用いて各モダリティの特徴量を抽出することで、図7右に示すように、文章特徴量(内容)、およびユーザ特徴量(送信者)を含むメール特徴量を取得することが可能となる。 As shown in the left of FIG. 7, for example, when extracting mail features, each modality using each feature extractor that can handle modalities such as texts and inclusive relations defined with mail and modalities, respectively. As shown in the right part of FIG. 7, it is possible to acquire a mail feature amount including a sentence feature amount (contents) and a user feature amount (sender).
 以下、本応用例における登録処理と検索処理について、図8および図9を参照してそれぞれ順次説明する。 Hereinafter, registration processing and search processing in this application example will be sequentially described with reference to FIG. 8 and FIG.
 (モダリティの包含関係の定義を考慮した登録処理)
 図8は、本実施形態による第1の応用例の登録処理の一例を示すシーケンス図である。
(Registration processing taking into account the definition of modality inclusion)
FIG. 8 is a sequence diagram illustrating an example of registration processing of the first application example according to the present embodiment.
 図8に示すように、まず、情報処理装置10の特徴空間管理部101は、ユーザの操作入力等に基づいて登録要求を取得すると(ステップS203)、登録要求に含まれるオブジェクト(obj)と当該オブジェクトのモダリティ(mdl)の情報(例えば、「Mail」)と共に、モダリティ管理部102に対して、(entityの)生成依頼を行う(ステップS206)。 As shown in FIG. 8, first, when the feature space management unit 101 of the information processing apparatus 10 acquires a registration request based on a user operation input or the like (step S203), the object (obj) included in the registration request and the object Along with information on the modality (mdl) of the object (for example, “Mail”), a request for generation of (entity) is made to the modality management unit 102 (step S206).
 次に、モダリティ管理部102は、モダリティ定義部103により、変換データ(entity)の生成、一意のIDの付与、およびobjと一意のIDの保存処理を行うと共に、objのモダリティと包含関係を有するモダリティ(sub mdl)の定義に基づいて、sub entityの生成を行う(ステップS209)。例えばオブジェクトのモダリティが「Mail」であって、これと包含関係を有するモダリティ(sub mdl)が「Text」の場合、モダリティ定義部103は、「Text」の定義に従って、メールデータ(obj)のうちテキストのデータを所定のデータ形式に変換し、sub entityとして生成する処理を行う。 Next, the modality management unit 102 uses the modality definition unit 103 to generate conversion data (entity), assign a unique ID, and store obj and a unique ID, and have an inclusion relationship with the modality of obj. Based on the definition of modality (sub mdl), sub entity is generated (step S209). For example, when the modality of the object is “Mail” and the modality (sub mdl) having an inclusion relation with this is “Text”, the modality definition unit 103 includes the mail data (obj) according to the definition of “Text”. Converts text data into a predetermined data format and generates a sub entity.
 次いで、特徴空間管理部101は、モダリティ管理部102から、少なくともID、entity、およびsub entityを取得する(ステップS212)。また、モダリティ管理部102からは、IDとobjを保存した旨が通知されてもよい。 Next, the feature space management unit 101 acquires at least ID, entity, and sub entity from the modality management unit 102 (step S212). Further, the modality management unit 102 may notify that the ID and obj are saved.
 次に、特徴空間管理部101は、取得したIDおよびentity(例えばMail Entity)に基づいて、対応する全ての特徴空間(例えば特徴空間B)に対してデータの追加(登録)要求を出力する(ステップS215)。続くステップS218~S221に示す特徴量の抽出に関する処理については、図4に示すステップS118~S121と同様であるため、詳細な説明は省略するが、例えばメールを扱う特徴空間Bには、メールの特徴量(Mail Vector)を登録するが、この際、メールの特徴量のうち、テキスト(Text Vector)については、次に説明するテキストを扱う特徴空間AのGet Vector(ステップS227)を利用するようにしてもよい。 Next, the feature space management unit 101 outputs a data addition (registration) request to all corresponding feature spaces (eg, feature space B) based on the acquired ID and entity (eg, Mail Entity) ( Step S215). The subsequent processing related to feature amount extraction shown in steps S218 to S221 is the same as that in steps S118 to S121 shown in FIG. 4 and will not be described in detail. For example, in the feature space B handling mail, The feature quantity (Mail Vector) is registered. At this time, for the text (Text Vector) among the feature quantities of the mail, the Get Vector (Step S227) of the feature space A handling the text described below is used. It may be.
 特徴空間管理部101は、同IDおよびsub entity(例えばText Entity)について、対応する全ての特徴空間(例えば特徴空間A)に対してデータの追加(登録)要求を出力する(ステップS224~230)。特徴空間Aは、テキストのみに対応した特徴空間であり、Text Entityから抽出したテキストの特徴量(Text Vector)が登録される。 The feature space management unit 101 outputs a data addition (registration) request for all corresponding feature spaces (eg, feature space A) for the same ID and sub entity (eg, Text Entity) (steps S224 to 230). . The feature space A is a feature space corresponding to text only, and the feature amount (Text Vector) of the text extracted from the Text Entity is registered.
 このように、本変形例では、特徴量の抽出において、包含関係を有する特徴空間を利用することができると共に、当該特徴空間にも特徴量を登録することが可能となる。 As described above, in this modification, in extracting feature quantities, a feature space having an inclusion relationship can be used, and a feature quantity can be registered in the feature space.
 (モダリティの包含関係の定義を考慮した検索処理)
 続いて、本変形例による検索処理について図9を参照して説明する。図9は、本実施形態による第1の応用例の検索処理の一例を示すシーケンス図である。
(Search processing taking into account the definition of modality inclusion)
Next, a search process according to this modification will be described with reference to FIG. FIG. 9 is a sequence diagram illustrating an example of search processing of the first application example according to the present embodiment.
 図9に示すように、まず、情報処理装置10の特徴空間管理部101は、ユーザの操作入力等に基づいて検索要求を取得する(ステップS243)。検索要求には、オブジェクト(obj)と、当該オブジェクトのモダリティ(mdl1)と、検索対象のモダリティを示すターゲットモダリティ(mdl2)とが含まれる。ここでは、例えばmdl1=Mail、mdl2=Textとする。 As shown in FIG. 9, first, the feature space management unit 101 of the information processing apparatus 10 acquires a search request based on a user operation input or the like (step S243). The search request includes an object (obj), a modality (mdl1) of the object, and a target modality (mdl2) indicating the modality to be searched. Here, for example, mdl1 = Mail and mdl2 = Text.
 次いで、特徴空間管理部101は、検索要求に含まれるオブジェクト(obj)と当該オブジェクトのモダリティ(mdl1)の情報と共に、モダリティ管理部102に対して、(entityの)生成依頼を行う(ステップS246)。 Next, the feature space management unit 101 requests the modality management unit 102 to generate (entity) together with information on the object (obj) and the modality (mdl1) of the object included in the search request (step S246). .
 次に、モダリティ管理部102は、モダリティ定義部103により、変換データ(entity)の生成、および一意のIDの付与を行うと共に、objのモダリティと包含関係を有するモダリティ(sub mdl)の定義に基づいて、sub entityの生成を行う(ステップS249)。例えばオブジェクトのモダリティが「Mail」であって、これと包含関係を有するモダリティ(sub mdl)が「Text」の場合、モダリティ定義部103は、「Text」の定義に従って、メールデータ(obj)のうちテキストのデータを所定のデータ形式に変換し、sub entityとして生成する処理を行う。 Next, the modality management unit 102 uses the modality definition unit 103 to generate conversion data (entity) and assign a unique ID, and based on the definition of the modality (sub mdl) having an inclusion relation with the modality of obj. Then, the sub entity is generated (step S249). For example, when the modality of the object is “Mail” and the modality (sub mdl) having an inclusion relation with this is “Text”, the modality definition unit 103 includes the mail data (obj) according to the definition of “Text”. Converts text data into a predetermined data format and generates a sub entity.
 次いで、特徴空間管理部101は、モダリティ管理部102から、ID、entity、およびsub entityを取得する(ステップS252)。 Next, the feature space management unit 101 acquires ID, entity, and sub entity from the modality management unit 102 (step S252).
 次に、特徴空間管理部101は、取得したentity(例えばMail Entity)に基づいて、対応する全ての特徴空間(Feature Space)に対して検索要求を出力する(ステップS255)。検索要求には、entity、mdl1(元データのモダリティ、例えばMail)、mdl2(ターゲットモダリティ、例えばText)が含まれる。対応する特徴空間とは、mdl1およびmdl2を扱い得る特徴空間(例えば、メールとテキスト双方に対応した特徴空間)である。 Next, the feature space management unit 101 outputs a search request to all corresponding feature spaces (Feature Space) based on the acquired entity (for example, Mail Entity) (step S255). The search request includes entity, mdl1 (original data modality, for example, Mail), and mdl2 (target modality, for example, Text). The corresponding feature space is a feature space that can handle mdl1 and mdl2 (for example, a feature space corresponding to both mail and text).
 続くステップS258~S267に示す特徴量の抽出に関する処理については、図5に示すステップS148~S157と同様であるため、詳細な説明は省略する。ここで、対応する特徴空間が存在しない場合も想定される。例えば、メールを扱う特徴空間Bと、テキストを扱う特徴空間Aが存在する場合、いずれも上記メールとテキストの双方を扱う特徴空間ではないため、検索結果は返されないが、次に説明するsub mdlを用いた場合には、特徴空間Aから検索結果が返され得る。 The subsequent processes related to feature amount extraction shown in steps S258 to S267 are the same as steps S148 to S157 shown in FIG. Here, it is assumed that there is no corresponding feature space. For example, if there is a feature space B that handles e-mails and a feature space A that handles texts, both are not feature spaces that handle both e-mail and text, so no search results will be returned. When is used, a search result can be returned from the feature space A.
 特徴空間管理部101は、同IDおよびsub entity(例えばText Entity)に基づいて、対応する全ての特徴空間(Feature Space)に対して検索要求を出力する(ステップS270)。検索要求には、sub entity(例えばText Entity)、mdl1(sub mdl、例えばText)、mdl2(ターゲットモダリティ、例えばText)が含まれる。対応する特徴空間とは、mdl1およびmdl2を扱い得る特徴空間、ここではmdl1およびmdl2が同じ「Text」であるため、テキストに対応した特徴空間Aが相当する。特徴空間Aにおいて検索が行われ(ステップS273~S279)、特徴空間管理部101は、特徴管理部201から、検索結果を取得する(ステップS282)。 The feature space management unit 101 outputs a search request to all corresponding feature spaces (Feature Space) based on the ID and sub entity (for example, Text Entity) (step S270). The search request includes sub entity (eg, Text Entity), mdl1 (sub mdl, eg, Text), and mdl2 (target modality, eg, Text). The corresponding feature space is a feature space that can handle mdl1 and mdl2, and here mdl1 and mdl2 are the same “Text”, and therefore feature space A corresponding to text corresponds. A search is performed in the feature space A (steps S273 to S279), and the feature space management unit 101 acquires a search result from the feature management unit 201 (step S282).
 続いて、特徴空間管理部101は、上述した図5に示すステップS160~S169と同様に、取得したIDおよび対応するモダリティ(例えば、Text)と共に、モダリティ管理部102に対して元データの要求を行い(ステップS285)、モダリティ定義部103によりIDに基づいて取得されたオブジェクトが(ステップS288)、モダリティ管理部102から特徴空間管理部101に出力される(ステップS291)。 Subsequently, the feature space management unit 101 sends a request for original data to the modality management unit 102 together with the acquired ID and the corresponding modality (for example, Text), as in steps S160 to S169 shown in FIG. In step S285, the object acquired based on the ID by the modality definition unit 103 (step S288) is output from the modality management unit 102 to the feature space management unit 101 (step S291).
 そして、特徴空間管理部101は、検索結果(オブジェクト、モダリティ、および類似度)を、検索要求元に出力する(ステップS294)。 Then, the feature space management unit 101 outputs the search result (object, modality, and similarity) to the search request source (step S294).
 以上、本応用例によるモダリティの包含関係を考慮した検索処理について具体的に説明した。 The search processing in consideration of the modality inclusion relation according to this application example has been specifically described above.
  <4-2.第2の応用例:検索結果のマージ>
 次に、第2の応用例として、検索結果のマージについて説明する。特徴空間管理部101は、各特徴抽出器からの検索結果の類似度と重み付けに基づいて、検索結果を再評価した上で、検索要求元に最終的な検索結果を出力することが可能である。重み付けとは、例えば特徴空間の重み付けである。かかる重み付けは、検索要求元(例えばユーザ)が任意に設定することも可能である。
<4-2. Second Application Example: Merging Search Results>
Next, search result merging will be described as a second application example. The feature space management unit 101 can output the final search result to the search request source after re-evaluating the search result based on the similarity and weighting of the search result from each feature extractor. . The weighting is, for example, weighting of the feature space. Such weighting can be arbitrarily set by a search request source (for example, a user).
 図10は、本応用例における検索画面の一例を示す図である。図10に示すように、検索画面32には、検索オブジェクト321と、検索対象の選択領域322と、何が似ているものを検索したいのかその特徴量を選択する領域323と、検索ボタン326が表示されている。特徴量を選択する領域323では、スライドバー324を操作して、選択した特徴量の重み付けを設定することが可能である。例えば、「形特徴」と「色特徴」のうち「色特徴」を優先したい場合は、スライドバー324の操作部325を「色特徴」の方に動かす。これにより、例えばシステム側で、以下のように重み付け(w)を設定する。ここで、色の特徴空間:space1、形の特徴空間:space2とする。
w( weights)={ space1: 0.8, space2: 0.2}
FIG. 10 is a diagram illustrating an example of a search screen in this application example. As shown in FIG. 10, the search screen 32 includes a search object 321, a search target selection region 322, a region 323 for selecting the feature amount of what is similar to search, and a search button 326. It is displayed. In the area 323 for selecting the feature amount, the weight of the selected feature amount can be set by operating the slide bar 324. For example, when it is desired to prioritize “color feature” among “shape feature” and “color feature”, the operation unit 325 of the slide bar 324 is moved toward “color feature”. Thereby, for example, the weight (w) is set as follows on the system side. Here, the color feature space: space1 and the shape feature space: space2.
w (weights) = {space1: 0.8, space2: 0.2}
 この場合、図10に示すように、「色特徴」を優先した検索結果(色が似ているイラストが優先された検索結果)が表示される。 In this case, as shown in FIG. 10, search results giving priority to “color features” (search results giving priority to illustrations with similar colors) are displayed.
 なお、特徴空間の重み付けの設定は、図10に示す例に限定されず、例えば検索結果からユーザが選択したものに基づいて重み付けを設定し、再度検索結果を提示するようにしてもよい。図11に一例を示す。例えば、図11の検索画面34に提示された検索結果のうち、イラスト341が選択されると、システムは、イラスト341が、ユーザの意図に近い結果であったとして、イラスト341を検索結果として出力した特徴空間(特徴抽出器、すなわち特徴抽出部202)を優先するよう重み付けを設定し、再度検索結果を提示するようにしてもよい。 Note that the setting of the weighting of the feature space is not limited to the example illustrated in FIG. 10. For example, the weighting may be set based on what is selected by the user from the search results, and the search results may be presented again. An example is shown in FIG. For example, when the illustration 341 is selected from the search results presented on the search screen 34 in FIG. 11, the system outputs the illustration 341 as the search result, assuming that the illustration 341 is close to the user's intention. The weighting may be set so as to prioritize the feature space (feature extractor, that is, the feature extraction unit 202), and the search result may be presented again.
 (動作処理)
 次に、本応用例の動作処理について図12を参照して説明する。図12は、第2の応用例の検索処理の一例を示すシーケンス図である。
(Operation processing)
Next, operation processing of this application example will be described with reference to FIG. FIG. 12 is a sequence diagram illustrating an example of search processing of the second application example.
 図12に示すように、まず、情報処理装置10の特徴空間管理部101は、ユーザの操作入力等に基づいて検索要求を取得する(ステップS303)。検索要求には、オブジェクト(obj)と、当該オブジェクトのモダリティ(mdl1)と、検索対象のモダリティを示すターゲットモダリティ(mdl2)と、特徴空間の重み付け(w)が含まれる。 As shown in FIG. 12, first, the feature space management unit 101 of the information processing apparatus 10 acquires a search request based on a user operation input or the like (step S303). The search request includes an object (obj), a modality (mdl1) of the object, a target modality (mdl2) indicating a modality to be searched, and a weight (w) of the feature space.
 続くステップS315~S321では、上述した図5のステップS145~157に示す処理と同様の検索処理が行われるため、ここでの詳細な説明は省略する。なお、ステップS318では、図5のステップS148~S154に示す処理と同様の処理が行われるが、詳細な図示は省略している。 In subsequent steps S315 to S321, a search process similar to the process shown in steps S145 to 157 of FIG. 5 described above is performed, and thus detailed description thereof is omitted here. In step S318, processing similar to that shown in steps S148 to S154 in FIG. 5 is performed, but detailed illustration is omitted.
 次に、特徴空間管理部101は、検索結果の類似度と重み付けに応じて、検索結果の順位付け(再評価)を行う(ステップS324)。具体的には、例えば特徴空間管理部101は、検索結果の類似度と、当該検索結果を出力した特徴空間(特徴抽出器、すなわち特徴抽出部202)の重みとを乗算し、新たな類似度を算出した上で、再評価を行い得る。下記表1に、再評価の一例を示す。ここで、w (weights)= {space1: 0.8, space2: 0.2}とする。 Next, the feature space management unit 101 ranks (re-evaluates) the search results according to the similarity and weighting of the search results (step S324). Specifically, for example, the feature space management unit 101 multiplies the similarity of the search result by the weight of the feature space (feature extractor, that is, the feature extraction unit 202) that outputs the search result, and creates a new similarity. Can be re-evaluated. Table 1 below shows an example of re-evaluation. Here, w (weights) = {space1: 0.8, space2: 0.2}.
Figure JPOXMLDOC01-appb-T000001
Figure JPOXMLDOC01-appb-T000001
 上記表1に示すように、例えば検索結果であるオブジェクトAが、第1の特徴空間(space1)から検索された際の類似度(sim(space1):0.9)に第1の特徴空間の重み(space1:0.8)を乗算した値と、第2の特徴空間(space2)から検索された際の類似度(sim(space2):0.3)に第2の特徴空間の重み(space2:0.2)を乗算した値とを加算した値(sim(new):0.78)が、新たな類似度として算出される。同じデータに関連付くIDが複数の特徴空間に登録されている場合も考えられるためである。また、検索結果が1つの特徴空間からのみ検索された場合も想定される。この場合、上記表1のオブジェクトCの例のように、例えばオブジェクトCが、第2の特徴空間(space2)から検索された際の類似度(sim(space2):0.9)に第2の特徴空間の重み(space2:0.2)を乗算した値(sim(new):0.18)が、新たな類似度として算出される。 As shown in Table 1 above, for example, the object A as a search result is weighted to the first feature space (sim (space1): 0.9) with the similarity (sim (space1): 0.9) when searched from the first feature space (space1). space1: 0.8) multiplied by the second feature space weight (space2: 0.2) multiplied by the similarity (sim (space2): 0.3) retrieved from the second feature space (space2) A value obtained by adding the values (sim (new): 0.78) is calculated as a new similarity. This is because an ID associated with the same data may be registered in a plurality of feature spaces. In addition, it is assumed that the search result is searched only from one feature space. In this case, as in the example of the object C in Table 1 above, for example, the second feature space has a similarity (sim (space2): 0.9) when the object C is searched from the second feature space (space2). A value (sim (new): 0.18) obtained by multiplying the weight (space2: 0.2) is calculated as a new similarity.
 特徴空間管理部101は、新たな類似度に基づいて、例えば上位所定数の検索結果(ID)を特定する。 The feature space management unit 101 specifies, for example, a predetermined number of search results (IDs) on the basis of the new similarity.
 次いで、特徴空間管理部101は、上述した図5に示すステップS160~S169と同様に、特定したIDおよび対応するモダリティと共に、モダリティ管理部102に対して元データの要求を行い(ステップS327)、モダリティ定義部103によりIDに基づいて取得されたオブジェクトが(ステップS330)、モダリティ管理部102から特徴空間管理部101に出力される(ステップS333)。 Next, the feature space management unit 101 makes a request for original data to the modality management unit 102 together with the identified ID and the corresponding modality, similarly to steps S160 to S169 shown in FIG. 5 described above (step S327). The object acquired based on the ID by the modality definition unit 103 (step S330) is output from the modality management unit 102 to the feature space management unit 101 (step S333).
 そして、特徴空間管理部101は、検索結果(オブジェクト、モダリティ、および類似度)を、検索要求元に出力する(ステップS336)。 Then, the feature space management unit 101 outputs the search result (object, modality, and similarity) to the search request source (step S336).
 以上、本応用例による検索結果のマージについて具体的に説明した。 So far, the search result merging according to this application example has been specifically described.
  <4-3.第3の応用例:サジェストシステム>
 次に、第3の応用例として、サジェストシステムについて図13~図15を参照して説明する。サジェストシステムは、複数のアプリケーションが動作するシステム上で用いることで、各アプリケーションにおけるユーザの操作情報(閲覧しているコンテンツや、操作しているコンテンツ等)に基づいて、状況に合ったコンテンツを検索し、ユーザに提案することを可能とする。
<4-3. Third application example: Suggest system>
Next, a suggestion system will be described with reference to FIGS. 13 to 15 as a third application example. The suggest system can be used on a system that runs multiple applications to search for content that suits the situation based on user operation information (browsing content, operating content, etc.) in each application. It is possible to make a proposal to the user.
 例えば、ユーザが色々なアプリケーションを用いて旅行計画を立てている場合を想定する。ユーザが、Webブラウザで観光地を探し、地図アプリで現地の地図を検索し、さらにノートアプリに計画をまとめている場合、サジェストシステムは、これらの複数アプリケーションの利用状況に応じて、需要に合ったコンテンツ(Webページやテキスト、画像など)を提案することが可能となる。 Suppose, for example, that the user is planning a travel plan using various applications. When a user searches for a sightseeing spot with a web browser, searches for a local map with a map application, and further summarizes a plan in a notebook application, the suggestion system will meet the demand according to the usage status of these multiple applications. Content (Web page, text, image, etc.) can be proposed.
 (構成例)
 図13は、本システムの構成の一例を示す機能ブロック図である。図13に示すように、例えばサジェストシステムは、情報処理装置10xにより実現され得る。情報処理装置10xは、1以上のアプリ105と、情報収集部106と、サジェスト部107と、特徴空間管理部101xと、モダリティ管理部102xと、として機能する。これらは、情報処理装置10の制御部100により実施され得る。
(Configuration example)
FIG. 13 is a functional block diagram showing an example of the configuration of the present system. As shown in FIG. 13, for example, a suggestion system can be realized by an information processing apparatus 10x. The information processing apparatus 10x functions as one or more applications 105, an information collection unit 106, a suggestion unit 107, a feature space management unit 101x, and a modality management unit 102x. These can be implemented by the control unit 100 of the information processing apparatus 10.
 アプリ105は、Webブラウザ、地図アプリケーション、ノートアプリケーション等の、各種アプリケーションプログラムである。 The application 105 is various application programs such as a web browser, a map application, and a notebook application.
 情報収集部106は、各アプリ105の動作を監視し、各アプリ105におけるユーザ操作情報(すなわちアプリケーションの利用状況)を収集、蓄積する機能を有する。また、情報収集部106には、OS(Operating System)を利用してもよい。 The information collection unit 106 has a function of monitoring the operation of each application 105 and collecting and storing user operation information (that is, application usage status) in each application 105. The information collecting unit 106 may use an OS (Operating System).
 サジェスト部107は、情報収集部106により収集された操作情報に基づいて検索要求を生成し、特徴空間管理部101xに対して検索要求を行う。例えばサジェスト部107は、情報収集部106から各アプリ105から取得した閲覧中/編集中のコンテンツのモダリティ(mdl1)と内容(obj)、および必要なコンテンツのモダリティ(mdl2)の要求に基づいて、検索要求を生成し得る。各アプリ105から取得されるコンテンツのモダリティと必要なコンテンツのモダリティの要求は、例えば以下のような例が想定される。
・Webブラウザ…閲覧:Webページ、要求:Webページ
・地図アプリ…閲覧:住所、要求:なし
・ノートアプリ…編集:テキスト/画像、要求:テキスト/画像
The suggestion unit 107 generates a search request based on the operation information collected by the information collection unit 106, and makes a search request to the feature space management unit 101x. For example, the suggestion unit 107, based on the request of the modality (mdl1) and content (obj) of the content being browsed / edited acquired from each application 105 from the information collection unit 106, and the modality (mdl2) of the required content, A search request may be generated. For example, the following examples of content modality acquired from each application 105 and required content modality requests are assumed.
-Web browser ... Browsing: Web page, Request: Web page / Map application ... Browsing: Address, Request: None * Note app ... Editing: Text / Image, Request: Text / Image
 特徴空間管理部101xは、サジェスト部107からの要求に応じて、1以上の特徴空間を用いた検索処理を行う。検索処理は、上述した実施形態と同様であり、まず特徴空間管理部101xがモダリティ管理部102xによりobjを変換処理したentityを取得し、entity、mdl1(例えばWebページ、住所、テキスト)、およびmdl2(例えばWebページ、画像)に基づいて、特徴管理サーバ20に対して検索要求を行う。そして、特徴空間管理部101xは、検索結果をサジェスト部107に出力する。 The feature space management unit 101x performs a search process using one or more feature spaces in response to a request from the suggestion unit 107. The search process is the same as in the above-described embodiment. First, the feature space management unit 101x acquires the entity obtained by converting the obj by the modality management unit 102x, and the entity, mdl1 (for example, Web page, address, text), and mdl2 A search request is made to the feature management server 20 based on (for example, Web page, image). Then, the feature space management unit 101x outputs the search result to the suggestion unit 107.
 モダリティ管理部102xは、図1を参照して説明したモダリティ管理部102と同様の機能を有し、モダリティ定義部103により、objを、mdl1のモダリティの定義に従って所定のデータ形式に変換する処理を行い、生成したentityを特徴空間管理部101xに出力する。 The modality management unit 102x has the same function as that of the modality management unit 102 described with reference to FIG. 1, and the modality definition unit 103 performs processing for converting obj into a predetermined data format according to the definition of the modality of mdl1. The generated entity is output to the feature space management unit 101x.
 以上、本応用例によるサジェストシステムを実行する情報処理装置10xの構成の一例について具体的に説明した。 The example of the configuration of the information processing apparatus 10x that executes the suggestion system according to this application example has been specifically described above.
 (動作処理)
 続いて、本応用例によるサジェストシステムの動作処理について図14を参照して説明する。図14は、本応用例のサジェストシステムにおける検索処理の流れの一例を示すシーケンス図である。
(Operation processing)
Next, an operation process of the suggest system according to this application example will be described with reference to FIG. FIG. 14 is a sequence diagram showing an example of the flow of search processing in the suggestion system of this application example.
 図14に示すように、まず、1以上のアプリ105は、ユーザにより操作が行われると(ステップS403)、扱っているコンテンツの送信(post;obj,mdl1)と、必要なコンテンツの要求(request;mdl2)を、情報収集部106に対して行う(ステップS406)。postの一例としては、例えば、「金閣寺」のWebページ、「京都市北区・・・1-2-3」という住所、および旅行関連のテキスト等が挙げられる。また、requestとしては、例えば、Webページ、画像が挙げられる。 As shown in FIG. 14, first, when one or more applications 105 are operated by the user (step S403), the content being handled (post; obj, mdl1) and the required content request (request) Mdl2) is performed on the information collecting unit 106 (step S406). As an example of the post, for example, a web page of “Kinkakuji”, an address “Kita-ku, Kyoto ... 1-2-3”, a travel-related text, and the like can be cited. Further, examples of the request include a web page and an image.
 次いで、情報収集部106は、収集した情報(post、request)を、サジェスト部107に出力する(ステップS412)。 Next, the information collection unit 106 outputs the collected information (post, request) to the suggestion unit 107 (step S412).
 次に、サジェスト部107は、特徴空間管理部101xに対し、検索要求を行う(ステップS415)。検索要求には、postに含まれるコンテンツがobj、そのモダリティがmdl1、また、requestに含まれるコンテンツのモダリティがmdl2として含まれる。 Next, the suggestion unit 107 makes a search request to the feature space management unit 101x (step S415). The search request includes the content included in post as obj, the modality thereof as mdl1, and the content modality included in request as mdl2.
 次いで、特徴空間管理部101xにおいて検索処理が実行される(ステップS418)。ステップS418では、図5のステップS136~S166と同様の処理(objとmdl1からentityの生成、entityとmdl1とmdl2に基づく検索、検索結果のIDからオブジェクトの取得)が行われるが、ここでの詳細な説明は省略する。 Next, a search process is executed in the feature space management unit 101x (step S418). In step S418, processing similar to that in steps S136 to S166 of FIG. 5 (generation of entity from obj and mdl1, search based on entity, mdl1, and mdl2, acquisition of object from search result ID) is performed. Detailed description is omitted.
 次に、サジェスト部107は、特徴空間管理部101xから検索結果を取得する(ステップS421)。 Next, the suggestion unit 107 acquires a search result from the feature space management unit 101x (step S421).
 次いで、サジェスト部107は、検索結果の類似度と重み付け(W)に応じて、検索結果の順位付け(再評価)を行ってもよい(ステップS424)。例えば、サジェスト部107は、入出力毎に下記表2のような重みを設定しておき、類似度に掛け合わせてランキングしてもよい。なお、本応用例において、かかる再評価はスキップされてもよい。 Next, the suggestion unit 107 may rank (re-evaluate) the search results according to the similarity and weighting (W) of the search results (step S424). For example, the suggestion unit 107 may set a weight as shown in Table 2 below for each input and output, and rank by multiplying the similarity. In this application example, such reevaluation may be skipped.
Figure JPOXMLDOC01-appb-T000002
Figure JPOXMLDOC01-appb-T000002
 そして、サジェスト部107は、検索結果を示す表示画面の作成を行い(ステップS427)、ユーザに提示する(ステップS430)。 The suggestion unit 107 creates a display screen showing the search result (step S427) and presents it to the user (step S430).
 また、サジェスト部107は、ユーザから利用状況のフィードバックを得た場合は、上記ステップ424で用いた重み付け(W)を更新等してもよい(ステップS433)。 Further, the suggestion unit 107 may update the weighting (W) used in the above step 424 when the usage status feedback is obtained from the user (step S433).
 なお、サジェスト部107によるユーザへのサジェストやユーザからのフィードバックの取得は、アプリ105を介して行うようにしてもよい。 Note that the suggestion unit 107 may suggest to the user and obtain feedback from the user via the application 105.
 ここで、図15に、本応用例によるアプリケーションから取得する操作情報と要求情報の一例を示す。本システムでは、各アプリケーションから、図15の左に示すような操作情報と、図15の右に示すような要求情報を取得し、操作情報に基づいて、要求された情報をサジェストする。 Here, FIG. 15 shows an example of operation information and request information acquired from an application according to this application example. In this system, the operation information as shown on the left in FIG. 15 and the request information as shown on the right in FIG. 15 are acquired from each application, and the requested information is suggested based on the operation information.
 <<5.まとめ>>
 上述したように、本開示の実施形態による情報処理システムでは、複数の特徴空間を扱うシステムの利便性をより向上させることが可能となる。
<< 5. Summary >>
As described above, the information processing system according to the embodiment of the present disclosure can further improve the convenience of a system that handles a plurality of feature spaces.
 以上、添付図面を参照しながら本開示の好適な実施形態について詳細に説明したが、本技術はかかる例に限定されない。本開示の技術分野における通常の知識を有する者であれば、特許請求の範囲に記載された技術的思想の範疇内において、各種の変更例または修正例に想到し得ることは明らかであり、これらについても、当然に本開示の技術的範囲に属するものと了解される。 The preferred embodiments of the present disclosure have been described in detail above with reference to the accompanying drawings, but the present technology is not limited to such examples. It is obvious that a person having ordinary knowledge in the technical field of the present disclosure can come up with various changes or modifications within the scope of the technical idea described in the claims. Of course, it is understood that it belongs to the technical scope of the present disclosure.
 例えば、上述した情報処理装置10、または特徴管理サーバ20に内蔵されるCPU、ROM、およびRAM等のハードウェアに、情報処理装置10、または特徴管理サーバ20の機能を発揮させるためのコンピュータプログラムも作成可能である。また、当該コンピュータプログラムを記憶させたコンピュータ読み取り可能な記憶媒体も提供される。 For example, a computer program for causing the information processing apparatus 10 or the feature management server 20 to perform the functions of the information processing apparatus 10 or the feature management server 20 on hardware such as a CPU, ROM, and RAM incorporated in the information processing apparatus 10 or the feature management server 20 described above. Can be created. A computer-readable storage medium storing the computer program is also provided.
 また、本明細書に記載された効果は、あくまで説明的または例示的なものであって限定的ではない。つまり、本開示に係る技術は、上記の効果とともに、または上記の効果に代えて、本明細書の記載から当業者には明らかな他の効果を奏しうる。 In addition, the effects described in this specification are merely illustrative or illustrative, and are not limited. That is, the technology according to the present disclosure can exhibit other effects that are apparent to those skilled in the art from the description of the present specification in addition to or instead of the above effects.
 なお、本技術は以下のような構成も取ることができる。
(1)
 登録要求情報に含まれる登録オブジェクトを複数の特徴抽出部に共通する一意の第1の識別情報と関連付けて記憶部に記憶する制御と、
 前記登録オブジェクトを前記登録オブジェクトのモダリティの定義に従って変換し、登録用の変換データを生成する制御と、
 前記第1の識別情報と前記登録用の変換データを、前記モダリティに対応する複数の特徴抽出器に出力する制御と、
を行う制御部を備える、情報処理装置。
(2)
 前記モダリティの定義は、モダリティに対応する所定のデータ形式への変換ルールである、前記(1)に記載の情報処理装置。
(3)
 前記制御部は、
  検索要求に含まれる検索オブジェクトを前記検索オブジェクトのモダリティの定義に従って変換し、検索用の変換データを生成する制御と、
  前記検索用の変換データを、前記検索オブジェクトのモダリティと前記検索要求に含まれるターゲットモダリティとに対応する前記特徴抽出器に出力する制御と、
を行う、前記(1)または(2)に記載の情報処理装置。
(4)
 前記制御部は、
  1以上の前記特徴抽出器において前記検索用の変換データに基づいて検索された第2の識別情報を取得し、
  前記第2の識別情報に基づいて、前記記憶部から対応するオブジェクトを取得し、検索結果として出力する、前記(3)に記載の情報処理装置。
(5)
 前記制御部は、
  前記特徴抽出器から、前記検索用の変換データから抽出された特徴と類似する特徴に関連付けられた前記第2の識別情報と共に、前記特徴の類似度合いを示す類似度を取得する、前記(4)に記載の情報処理装置。
(6)
 前記検索要求には、検索条件としてフィルター情報がさらに含まれる、前記(4)または(5)に記載の情報処理装置。
(7)
 前記制御部は、
  前記登録要求情報が入力された際、前記登録オブジェクトのモダリティと親子関係を有するサブモダリティの定義に従って、前記登録オブジェクトを変換して登録用のサブ変換データを生成する制御と、
 前記第1の識別情報と前記サブ変換データを、前記サブモダリティに対応する1以上の特徴抽出器に出力する制御と、
をさらに行う、前記(1)~(6)のいずれか1項に記載の情報処理装置。
(8)
 前記制御部は、
  前記検索要求が入力された際、前記検索オブジェクトのモダリティと親子関係を有するサブモダリティの定義に従って、前記検索オブジェクトのうち前記サブモダリティに対応するデータを変換して検索用のサブ変換データを生成する制御と、
  前記検索用のサブ変換データを、前記サブモダリティおよび前記ターゲットモダリティに対応する1以上の前記特徴抽出器に出力する制御と、
をさらに行う、前記(3)~(6)のいずれか1項に記載の情報処理装置。
(9)
 前記制御部は、
  前記検索要求に基づいて前記特徴抽出器から取得した前記第2の識別情報および類似度と、前記特徴抽出器の重み付けに基づいて、複数の前記第2の識別情報を順位付けする制御と、
  上位所定数の前記第2の識別情報を前記検索結果として出力する制御と、
をさらに行う、前記(4)~(6)のいずれか1項に記載の情報処理装置。
(10)
 前記制御部は、
  1以上のアプリケーションから出力されたユーザの操作情報を含む情報に基づいて、前記ユーザに提案するコンテンツを検索する前記検索要求を生成し、
 前記特徴抽出器から取得した1以上の前記第2の識別情報を、前記検索結果として出力する、前記(4)~(6)のいずれか1項に記載の情報処理装置。
(11)
 前記制御部は、
  前記情報に含まれる、前記アプリケーションで扱われているコンテンツを前記検索オブジェクトとし、
  前記コンテンツのモダリティを、前記検索オブジェクトのモダリティとし、
  前記アプリケーションで要求されているコンテンツのモダリティを、前記ターゲットモダリティとして、前記検索要求を生成する、前記(10)に記載の情報処理装置。
(12)
 前記情報処理装置は、
  前記第1の識別情報と前記登録用の変換データを、前記特徴抽出器を有する特徴管理サーバに送信する通信部をさらに備える、前記(1)~(11)のいずれか1項に記載の情報処理装置。
(13)
 プロセッサが、
 登録要求情報に含まれる登録オブジェクトを複数の特徴抽出部に共通する一意の第1の識別情報と関連付けて記憶部に記憶する制御と、
 前記登録オブジェクトを前記登録オブジェクトのモダリティの定義に従って変換し、登録用の変換データを生成する制御と、
 前記第1の識別情報と前記登録用の変換データを、前記モダリティに対応する複数の特徴抽出器に出力する制御と、
を行うことを含む、情報処理方法。
(14)
 コンピュータを、
 登録要求情報に含まれる登録オブジェクトを複数の特徴抽出部に共通する一意の第1の識別情報と関連付けて記憶部に記憶する制御と、
 前記登録オブジェクトを前記登録オブジェクトのモダリティの定義に従って変換し、登録用の変換データを生成する制御と、
 前記第1の識別情報と前記登録用の変換データを、前記モダリティに対応する複数の特徴抽出器に出力する制御と、
を行う制御部として機能させるための、プログラム。
In addition, this technique can also take the following structures.
(1)
Control for storing the registration object included in the registration request information in the storage unit in association with the unique first identification information common to the plurality of feature extraction units;
Control for converting the registration object according to the definition of the modality of the registration object, and generating conversion data for registration;
Control to output the first identification information and the conversion data for registration to a plurality of feature extractors corresponding to the modality;
An information processing apparatus comprising a control unit that performs the following.
(2)
The information processing apparatus according to (1), wherein the definition of the modality is a rule for conversion to a predetermined data format corresponding to the modality.
(3)
The controller is
A control for converting the search object included in the search request according to the definition of the modality of the search object, and generating conversion data for search;
Control to output the search conversion data to the feature extractor corresponding to the modality of the search object and the target modality included in the search request;
The information processing apparatus according to (1) or (2), wherein:
(4)
The controller is
Obtaining second identification information searched based on the search conversion data in the one or more feature extractors;
The information processing apparatus according to (3), wherein a corresponding object is acquired from the storage unit based on the second identification information, and is output as a search result.
(5)
The controller is
The degree of similarity indicating the degree of similarity of the feature is acquired from the feature extractor together with the second identification information associated with the feature similar to the feature extracted from the search conversion data, (4) The information processing apparatus described in 1.
(6)
The information processing apparatus according to (4) or (5), wherein the search request further includes filter information as a search condition.
(7)
The controller is
When the registration request information is input, according to the definition of the sub-modality having a parent-child relationship with the modality of the registration object, control for converting the registration object to generate sub-conversion data for registration;
Control for outputting the first identification information and the sub-transformed data to one or more feature extractors corresponding to the sub-modalities;
The information processing apparatus according to any one of (1) to (6), further performing:
(8)
The controller is
When the search request is input, sub-converted data for search is generated by converting data corresponding to the sub-modality among the search objects according to the definition of the sub-modality having a parent-child relationship with the modality of the search object. Control,
Control to output the sub-transform data for search to one or more feature extractors corresponding to the sub-modality and the target modality;
The information processing apparatus according to any one of (3) to (6), further performing:
(9)
The controller is
Control for ranking the plurality of second identification information based on the second identification information and similarity obtained from the feature extractor based on the search request, and the weight of the feature extractor;
A control for outputting the upper predetermined number of the second identification information as the search results;
The information processing apparatus according to any one of (4) to (6), further performing:
(10)
The controller is
Generating the search request for searching for content to be proposed to the user based on information including user operation information output from one or more applications;
The information processing apparatus according to any one of (4) to (6), wherein one or more pieces of the second identification information acquired from the feature extractor are output as the search result.
(11)
The controller is
The content included in the information and handled by the application is the search object,
The modality of the content is the modality of the search object,
The information processing apparatus according to (10), wherein the search request is generated using the modality of the content requested by the application as the target modality.
(12)
The information processing apparatus includes:
The information according to any one of (1) to (11), further including a communication unit that transmits the first identification information and the conversion data for registration to a feature management server having the feature extractor. Processing equipment.
(13)
Processor
Control for storing the registration object included in the registration request information in the storage unit in association with the unique first identification information common to the plurality of feature extraction units;
Control for converting the registration object according to the definition of the modality of the registration object, and generating conversion data for registration;
Control to output the first identification information and the conversion data for registration to a plurality of feature extractors corresponding to the modality;
An information processing method including performing.
(14)
Computer
Control for storing the registration object included in the registration request information in the storage unit in association with the unique first identification information common to the plurality of feature extraction units;
Control for converting the registration object according to the definition of the modality of the registration object, and generating conversion data for registration;
Control to output the first identification information and the conversion data for registration to a plurality of feature extractors corresponding to the modality;
A program for functioning as a control unit for performing
 10、10x 情報処理装置
 20、24 特徴管理サーバ
 25 データベースサーバ
 100 制御部
 101、101x 特徴空間管理部
 102、102x  モダリティ管理部
 103 モダリティ定義部
 105 アプリ
 106 情報収集部
 107 サジェスト部
 110 入力部
 120 通信部
 130 出力部
 140 記憶部
 200 制御部
 201 特徴管理部
 202 特徴抽出部
 210 通信部
 220 特徴量データベース
 240 特徴抽出部
 250 特徴量データベース
10, 10x information processing device 20, 24 feature management server 25 database server 100 control unit 101, 101x feature space management unit 102, 102x modality management unit 103 modality definition unit 105 application 106 information collection unit 107 suggestion unit 110 input unit 120 communication unit DESCRIPTION OF SYMBOLS 130 Output part 140 Storage part 200 Control part 201 Feature management part 202 Feature extraction part 210 Communication part 220 Feature quantity database 240 Feature extraction part 250 Feature quantity database

Claims (14)

  1.  登録要求情報に含まれる登録オブジェクトを複数の特徴抽出部に共通する一意の第1の識別情報と関連付けて記憶部に記憶する制御と、
     前記登録オブジェクトを前記登録オブジェクトのモダリティの定義に従って変換し、登録用の変換データを生成する制御と、
     前記第1の識別情報と前記登録用の変換データを、前記モダリティに対応する複数の特徴抽出器に出力する制御と、
    を行う制御部を備える、情報処理装置。
    Control for storing the registration object included in the registration request information in the storage unit in association with the unique first identification information common to the plurality of feature extraction units;
    Control for converting the registration object according to the definition of the modality of the registration object, and generating conversion data for registration;
    Control to output the first identification information and the conversion data for registration to a plurality of feature extractors corresponding to the modality;
    An information processing apparatus comprising a control unit that performs the following.
  2.  前記モダリティの定義は、モダリティに対応する所定のデータ形式への変換ルールである、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the definition of the modality is a rule for conversion to a predetermined data format corresponding to the modality.
  3.  前記制御部は、
      検索要求に含まれる検索オブジェクトを前記検索オブジェクトのモダリティの定義に従って変換し、検索用の変換データを生成する制御と、
      前記検索用の変換データを、前記検索オブジェクトのモダリティと前記検索要求に含まれるターゲットモダリティとに対応する前記特徴抽出器に出力する制御と、
    を行う、請求項1に記載の情報処理装置。
    The controller is
    A control for converting the search object included in the search request according to the definition of the modality of the search object, and generating conversion data for search;
    Control to output the search conversion data to the feature extractor corresponding to the modality of the search object and the target modality included in the search request;
    The information processing apparatus according to claim 1, wherein:
  4.  前記制御部は、
      1以上の前記特徴抽出器において前記検索用の変換データに基づいて検索された第2の識別情報を取得し、
      前記第2の識別情報に基づいて、前記記憶部から対応するオブジェクトを取得し、検索結果として出力する、請求項3に記載の情報処理装置。
    The controller is
    Obtaining second identification information searched based on the search conversion data in the one or more feature extractors;
    The information processing apparatus according to claim 3, wherein a corresponding object is acquired from the storage unit based on the second identification information, and is output as a search result.
  5.  前記制御部は、
      前記特徴抽出器から、前記検索用の変換データから抽出された特徴と類似する特徴に関連付けられた前記第2の識別情報と共に、前記特徴の類似度合いを示す類似度を取得する、請求項4に記載の情報処理装置。
    The controller is
    The degree of similarity indicating the degree of similarity of the feature is acquired from the feature extractor together with the second identification information associated with the feature similar to the feature extracted from the search conversion data. The information processing apparatus described.
  6.  前記検索要求には、検索条件としてフィルター情報がさらに含まれる、請求項4に記載の情報処理装置。 The information processing apparatus according to claim 4, wherein the search request further includes filter information as a search condition.
  7.  前記制御部は、
      前記登録要求情報が入力された際、前記登録オブジェクトのモダリティと親子関係を有するサブモダリティの定義に従って、前記登録オブジェクトを変換して登録用のサブ変換データを生成する制御と、
     前記第1の識別情報と前記サブ変換データを、前記サブモダリティに対応する1以上の特徴抽出器に出力する制御と、
    をさらに行う、請求項1に記載の情報処理装置。
    The controller is
    When the registration request information is input, according to the definition of the sub-modality having a parent-child relationship with the modality of the registration object, control for converting the registration object to generate sub-conversion data for registration;
    Control for outputting the first identification information and the sub-transformed data to one or more feature extractors corresponding to the sub-modalities;
    The information processing apparatus according to claim 1, further performing:
  8.  前記制御部は、
      前記検索要求が入力された際、前記検索オブジェクトのモダリティと親子関係を有するサブモダリティの定義に従って、前記検索オブジェクトのうち前記サブモダリティに対応するデータを変換して検索用のサブ変換データを生成する制御と、
      前記検索用のサブ変換データを、前記サブモダリティおよび前記ターゲットモダリティに対応する1以上の前記特徴抽出器に出力する制御と、
    をさらに行う、請求項3に記載の情報処理装置。
    The controller is
    When the search request is input, sub-converted data for search is generated by converting data corresponding to the sub-modality among the search objects according to the definition of the sub-modality having a parent-child relationship with the modality of the search object. Control,
    Control to output the sub-transform data for search to one or more feature extractors corresponding to the sub-modality and the target modality;
    The information processing apparatus according to claim 3, further performing:
  9.  前記制御部は、
      前記検索要求に基づいて前記特徴抽出器から取得した前記第2の識別情報および類似度と、前記特徴抽出器の重み付けに基づいて、複数の前記第2の識別情報を順位付けする制御と、
      上位所定数の前記第2の識別情報を前記検索結果として出力する制御と、
    をさらに行う、請求項4に記載の情報処理装置。
    The controller is
    Control for ranking the plurality of second identification information based on the second identification information and similarity obtained from the feature extractor based on the search request, and the weight of the feature extractor;
    A control for outputting the upper predetermined number of the second identification information as the search results;
    The information processing apparatus according to claim 4, further performing:
  10.  前記制御部は、
      1以上のアプリケーションから出力されたユーザの操作情報を含む情報に基づいて、前記ユーザに提案するコンテンツを検索する前記検索要求を生成し、
     前記特徴抽出器から取得した1以上の前記第2の識別情報を、前記検索結果として出力する、請求項4に記載の情報処理装置。
    The controller is
    Generating the search request for searching for content to be proposed to the user based on information including user operation information output from one or more applications;
    The information processing apparatus according to claim 4, wherein one or more pieces of the second identification information acquired from the feature extractor are output as the search result.
  11.  前記制御部は、
      前記情報に含まれる、前記アプリケーションで扱われているコンテンツを前記検索オブジェクトとし、
      前記コンテンツのモダリティを、前記検索オブジェクトのモダリティとし、
      前記アプリケーションで要求されているコンテンツのモダリティを、前記ターゲットモダリティとして、前記検索要求を生成する、請求項10に記載の情報処理装置。
    The controller is
    The content included in the information and handled by the application is the search object,
    The modality of the content is the modality of the search object,
    The information processing apparatus according to claim 10, wherein the search request is generated with the content modality requested by the application as the target modality.
  12.  前記情報処理装置は、
      前記第1の識別情報と前記登録用の変換データを、前記特徴抽出器を有する特徴管理サーバに送信する通信部をさらに備える、請求項1に記載の情報処理装置。
    The information processing apparatus includes:
    The information processing apparatus according to claim 1, further comprising a communication unit that transmits the first identification information and the conversion data for registration to a feature management server having the feature extractor.
  13.  プロセッサが、
     登録要求情報に含まれる登録オブジェクトを複数の特徴抽出部に共通する一意の第1の識別情報と関連付けて記憶部に記憶する制御と、
     前記登録オブジェクトを前記登録オブジェクトのモダリティの定義に従って変換し、登録用の変換データを生成する制御と、
     前記第1の識別情報と前記登録用の変換データを、前記モダリティに対応する複数の特徴抽出器に出力する制御と、
    を行うことを含む、情報処理方法。
    Processor
    Control for storing the registration object included in the registration request information in the storage unit in association with the unique first identification information common to the plurality of feature extraction units;
    Control for converting the registration object according to the definition of the modality of the registration object, and generating conversion data for registration;
    Control to output the first identification information and the conversion data for registration to a plurality of feature extractors corresponding to the modality;
    An information processing method including performing.
  14.  コンピュータを、
     登録要求情報に含まれる登録オブジェクトを複数の特徴抽出部に共通する一意の第1の識別情報と関連付けて記憶部に記憶する制御と、
     前記登録オブジェクトを前記登録オブジェクトのモダリティの定義に従って変換し、登録用の変換データを生成する制御と、
     前記第1の識別情報と前記登録用の変換データを、前記モダリティに対応する複数の特徴抽出器に出力する制御と、
    を行う制御部として機能させるための、プログラム。
    Computer
    Control for storing the registration object included in the registration request information in the storage unit in association with the unique first identification information common to the plurality of feature extraction units;
    Control for converting the registration object according to the definition of the modality of the registration object, and generating conversion data for registration;
    Control to output the first identification information and the conversion data for registration to a plurality of feature extractors corresponding to the modality;
    A program for functioning as a control unit for performing
PCT/JP2019/004534 2018-03-16 2019-02-08 Information processing device, information processing method, and program WO2019176398A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2020505682A JP7255585B2 (en) 2018-03-16 2019-02-08 Information processing device, information processing method, and program

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2018049304 2018-03-16
JP2018-049304 2018-03-16

Publications (1)

Publication Number Publication Date
WO2019176398A1 true WO2019176398A1 (en) 2019-09-19

Family

ID=67907731

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2019/004534 WO2019176398A1 (en) 2018-03-16 2019-02-08 Information processing device, information processing method, and program

Country Status (2)

Country Link
JP (1) JP7255585B2 (en)
WO (1) WO2019176398A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112905829A (en) * 2021-03-25 2021-06-04 王芳 Cross-modal artificial intelligence information processing system and retrieval method

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004348706A (en) * 2003-04-30 2004-12-09 Canon Inc Information processing device, information processing method, storage medium, and program
JP2006285612A (en) * 2005-03-31 2006-10-19 Canon Inc Information processor, and its method
JP2006343850A (en) * 2005-06-07 2006-12-21 Fuji Xerox Co Ltd Recommendation information providing system

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004348706A (en) * 2003-04-30 2004-12-09 Canon Inc Information processing device, information processing method, storage medium, and program
JP2006285612A (en) * 2005-03-31 2006-10-19 Canon Inc Information processor, and its method
JP2006343850A (en) * 2005-06-07 2006-12-21 Fuji Xerox Co Ltd Recommendation information providing system

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112905829A (en) * 2021-03-25 2021-06-04 王芳 Cross-modal artificial intelligence information processing system and retrieval method

Also Published As

Publication number Publication date
JP7255585B2 (en) 2023-04-11
JPWO2019176398A1 (en) 2021-04-22

Similar Documents

Publication Publication Date Title
CN111753060B (en) Information retrieval method, apparatus, device and computer readable storage medium
CN111125422B (en) Image classification method, device, electronic equipment and storage medium
US11030445B2 (en) Sorting and displaying digital notes on a digital whiteboard
CN110301117B (en) Method and apparatus for providing response in session
CN110209897B (en) Intelligent dialogue method, device, storage medium and equipment
US11899681B2 (en) Knowledge graph building method, electronic apparatus and non-transitory computer readable storage medium
US10437868B2 (en) Providing images for search queries
JP6381775B2 (en) Information processing system and information processing method
CN107491655B (en) Liver disease information intelligent consultation system based on machine learning
JP2022002075A (en) Information recommendation method and device, electronic apparatus, program and computer readable storage medium
CN110675944A (en) Triage method and device, computer equipment and medium
JP6033697B2 (en) Image evaluation device
JP3220886B2 (en) Document search method and apparatus
CN106874397B (en) Automatic semantic annotation method for Internet of things equipment
US10191921B1 (en) System for expanding image search using attributes and associations
JP2018014058A (en) Medical information processing system, medical information processing device and medical information processing method
JP5876396B2 (en) Information collection program, information collection method, and information processing apparatus
WO2019176398A1 (en) Information processing device, information processing method, and program
JP2019101944A (en) Intellectual Property System, Intellectual Property Support Method and Intellectual Property Support Program
JP6802332B1 (en) Information processing method and information processing equipment
JP2019114308A (en) Intellectual Property System, Intellectual Property Support Method and Intellectual Property Support Program
Xu et al. Estimating similarity of rich internet pages using visual information
US10095802B2 (en) Methods and systems for using field characteristics to index, search for, and retrieve forms
JP6531302B1 (en) Intellectual Property System, Intellectual Property Support Method and Intellectual Property Support Program
JP5389764B2 (en) Microblog text classification apparatus, method and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19768066

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2020505682

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19768066

Country of ref document: EP

Kind code of ref document: A1