US20170147712A1 - Memory equipped with information retrieval function, method for using same, device, and information processing method - Google Patents

Memory equipped with information retrieval function, method for using same, device, and information processing method Download PDF

Info

Publication number
US20170147712A1
US20170147712A1 US15/425,555 US201715425555A US2017147712A1 US 20170147712 A1 US20170147712 A1 US 20170147712A1 US 201715425555 A US201715425555 A US 201715425555A US 2017147712 A1 US2017147712 A1 US 2017147712A1
Authority
US
United States
Prior art keywords
memory
logical
information
data
search function
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/425,555
Inventor
Katsumi Inoue
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US15/425,555 priority Critical patent/US20170147712A1/en
Publication of US20170147712A1 publication Critical patent/US20170147712A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/903Querying
    • G06F16/90335Query processing
    • G06F16/90339Query processing by using parallel associative memories or content-addressable memories
    • G06F17/30982
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11CSTATIC STORES
    • G11C15/00Digital stores in which information comprising one or more characteristic parts is written into the store and in which information is read-out by searching for one or more of these characteristic parts, i.e. associative or content-addressed stores
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11CSTATIC STORES
    • G11C15/00Digital stores in which information comprising one or more characteristic parts is written into the store and in which information is read-out by searching for one or more of these characteristic parts, i.e. associative or content-addressed stores
    • G11C15/04Digital stores in which information comprising one or more characteristic parts is written into the store and in which information is read-out by searching for one or more of these characteristic parts, i.e. associative or content-addressed stores using semiconductor elements
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11CSTATIC STORES
    • G11C7/00Arrangements for writing information into, or reading information out from, a digital store
    • G11C7/10Input/output [I/O] data interface arrangements, e.g. I/O data control circuits, I/O data buffers
    • G11C7/1006Data managing, e.g. manipulating data before writing or reading out, data bus switches or control circuits therefor
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/28Constructional details of speech recognition systems
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11CSTATIC STORES
    • G11C15/00Digital stores in which information comprising one or more characteristic parts is written into the store and in which information is read-out by searching for one or more of these characteristic parts, i.e. associative or content-addressed stores
    • G11C15/04Digital stores in which information comprising one or more characteristic parts is written into the store and in which information is read-out by searching for one or more of these characteristic parts, i.e. associative or content-addressed stores using semiconductor elements
    • G11C15/043Digital stores in which information comprising one or more characteristic parts is written into the store and in which information is read-out by searching for one or more of these characteristic parts, i.e. associative or content-addressed stores using semiconductor elements using capacitive charge storage elements

Definitions

  • in-memory database technologies have become a focus of attention as fast-processing techniques for big data.
  • One of the in-memory database technologies is used in data mining, where an enormous amount of data resides on a semiconductor memory (in-memory database) to speed up data accesses and, therefore, searches for information of interest.
  • moving the information data from a hard disk device to a semiconductor memory alone only offers expectation of 1 to 2 orders of speed improvement.
  • CPU's do all information processing regardless of whether of not the CPU's are suited for the processing they have to perform.
  • data items are analogous to playing cards with their faces down, and the CPU's have to turn over each card (access each address) when searching for information.
  • information processing such as one to sequentially search for specific information on a memory
  • an amount of information processing becomes extremely large, leading to a long overall waiting time. This is the bus bottleneck, an inevitable problem of the von Neumann-architecture computers.
  • any of the above algorithms is a method for, for example, organizing types and in-memory locations of information pieces ahead of time, creating headers and their routes so that a CPU can easily find information, arranging data items according to their sizes.
  • the CPU load is reduced during the search, but pre- and post-processing mandates complex information processing.
  • data rearrangement or reordering is required every time a data item is added or deleted as the pre- and post-processing for these algorithms.
  • CAM Content-addressable memories
  • CAMs Content-addressable memories
  • CAMs content-addressable memories
  • CAMs have challenges that they each needs a completely parallel comparison circuit which increases their overall circuit size, and that parallel circuits constituting the comparison circuit consume high current. For this reason, utilization of the content-addressable memories (CAMs) has been limited to special applications where super fast search is essential as in communication routers.
  • the purpose of the present invention is to provide a memory with a new concept to enable a big data search at a speed comparable to that of content-addressable memories (CAM) by simply incorporating an extremely small number of circuits into a common memory.
  • CAM content-addressable memories
  • a memory 101 of the present invention may be applied to the above two prior inventions.
  • Japanese Patent Application No. 10-232531 “Memory with Operation Function” has an objective, as it illustrates, to improve a chip efficiency by providing an operation circuit for each block unit. This reduces the size of operation circuitry compared to providing an operation circuit for each memory, but deteriorates the operation efficiency as well as having a poor chip efficiency and undesirable cost issues.
  • the present memory leverages the concept of content-addressable memory (CAM) capable of parallel information searches, and addresses the CAM's challenges such as its circuit size and high power consumption to thereby provide a memory element applicable to big data and based on a novel concept free from conventional information processing scheme.
  • CAM content-addressable memory
  • a memory capable of reading and writing information, provided with an information search function
  • the memory has one word bit width of n and N word addresses, that is a storage cell structure of N ⁇ n bits,
  • the memory comprising: a logical operation unit having a set of n bits for the entirety of the memory;
  • the memory provided with the information search function of claim 1 , wherein the logical operation unit is configured to be capable of logical storage, logical product, logical sum, logical negation, exclusive logic or any combinatorial operation thereof between single bits of the entered (substituted) memory storage cell information having the one word bit width of n.
  • the memory provided with the information search function of claim 1 , wherein the logical operation unit is provided with a shift register function.
  • the memory provided with the information search function of claim 1 , wherein the memory performs an exact match search for a data value and a range search for data values using the logical storage, logical product, logical sum, logical negation, exclusive logic or any combinatorial operation thereof.
  • the memory provided with the information search function of claim 1 , wherein the memory is integrated with a semiconductor device of another type such as a CPU.
  • the memory provided with the information search function of claim 1 , wherein the memory is implemented in an FPGA.
  • a method for using the memory provided with the information search function of claim 1 comprising the step of assigning a database record to any one of columns within the one word bit width of n to generate a database, wherein a total bit number of fields per record of the database equals the number of word addresses N.
  • a method for using the memory provided with the information search function of claim 1 comprising the step of:
  • a device including the memory of claim 1 .
  • An information processing method comprising the step of repeating logical product, logical sum, logical negation, exclusive logic or any combinatorial logical operation thereof between single bits of the memory storage cell information within the memory to thereby obtain a prescribed operation result.
  • the information processing method of claim 11 further comprising the step of performing the logical operation in parallel between the single bits of the information.
  • FIG. 1 is a structural view of a common memory
  • FIG. 2 is a structural view of a memory provided with an information search function
  • FIG. 3 is an example of publication search using the memory provided with the information search function (Example 1);
  • FIG. 4 is an example of exact match data search using the memory provided with the information search function (Example 2);
  • FIG. 5 is an example of range data search using the memory provided with the information search function (Example 3);
  • FIG. 6 is an example of serial and parallel connections of the memories each provided with the information search function (Example 4).
  • FIG. 7 is an example of hierarchical connections of the memories each provided with the information search function (Example 5).
  • FIG. 1 is a structural view showing a general memory according to this one embodiment.
  • this memory 100 is configured so that information data may be freely written into and read out of this memory, wherein this memory 100 consists of a storage cells 102 consisting of N ⁇ n bits of cells, wherein the storage cells 102 comprise N word addresses 104 with each word having a width 103 of n bits, and wherein generally the 1-N word addresses may be externally selected/specified with means such as an address decoder.
  • the data width 103 of the memory 100 is a constant value such as 8 bits, 16 bits, 32 bits, etc., and for information data searches, the CPU sequentially accesses addresses in the memory's address space, reads data stored therein and serially processes the data, wherein a given number of addresses in the memory's address space is 1 M, 1 G, or the like.
  • Information processing by the memory of the present invention is based on a concept which defies the common sense of data width and address of the above general memory structure and database table structure, and is based on single-bit-based parallel logic operations.
  • FIG. 2 shows an example memory structure provided with an information search function according to the present invention.
  • this memory 101 is configured so that information data may be freely written into and read out of this memory, wherein this memory 101 consists of storage cells 102 consisting of N ⁇ n bits of cells, wherein the storage cells 102 comprise N word addresses 104 with each word having a width 103 of n bits, and wherein the 1-N word addresses may be externally selected/specified 110 .
  • the one word bit width of n 103 corresponds with a number of database records (n), and for the ease of understanding, one may analogously consider that each record is arranged as a column and N of the word address 104 corresponds to fields in each record.
  • this memory is a data table having n records with each record having N bits.
  • a logical operation unit 105 is provided in parallel to n-bit storage cells 102 in this memory's row (horizontal) direction, which n-bit storage cells 102 may be selected/specified 110 using the word address 104 , and this logical operation unit 105 is configured so that circuits capable of logical storage 116 for respective n-bit storage cells 102 of a selected/specified 110 word address 104 , logical product 112 , logical sum 113 , logical negation (NOT) 114 , exclusive logic 115 and combinations thereof may be freely specified.
  • NOT logical negation
  • an operation result output 106 function of a priority address encoder output circuit or the like is provided in order to output operation results of this logical operation unit 105 .
  • This memory is mostly made of memory cells themselves and only small part of this memory is dedicated for the logical operation unit 105 and the operation result output 106 function, and therefore, by incorporating these functions in a miniscule area of respective general memories, those general memories may become high-capacity memories usable for databases.
  • the memory may be adapted so that multiple word addresses may be selected/specified 110 simultaneously and multiple logical operation units 105 may be provided.
  • a current memory capacity per semiconductor die (chip) is about 8 Gbits.
  • the main memory 101 may be implemented for the main memory 101 , for example, 1 M word addresses and 8 Kbit word width, or 8 K word addresses and 1 Mbit word width.
  • This memory 101 may be effective for any information search, especially, big data search, data mining, genome analysis, etc., and some practical examples of this memory will be discussed below.
  • One example of our everyday search is an Internet search and its concept is embodied by information refinement using a keyword. For example, by providing a keyword such as “information processing,” “information search” or “CPU,” the search is refined and suitable Internet sites may be found.
  • a keyword such as “information processing,” “information search” or “CPU”
  • FIG. 3 shows an example of publication search using a memory provided with an information search function.
  • each of word addresses from 1 to N may be assigned to one vocabulary term such as “information processing,” “information search,” “patent” or “CPU,” and each of vertical lines (columns) of word width n may be assigned to one publication.
  • the word addresses 1-N constitute one record as a whole, and the word addresses 1-N correspond to fields.
  • N vocabulary terms and n publications (n records) are registered in a database.
  • a word address 18 corresponds to “information processing”
  • a word address 5 corresponds to “information search”
  • a word address 24 corresponds to “patent”
  • a word address 10 corresponds to “CPU,” wherein here we assume that the search operation formula is (publication(s) containing either “information processing” or “information search”) ⁇ (publication(s) not containing “patent”) ⁇ (publication(s) containing “CPU”).
  • Records of publications containing either “information processing” of the word address 18 or “information search” of the word address 5 are records 3 , 4 , 5 , 13 , 14 , 16 , 19 , 21 and 25 .
  • records of publications not containing “patent” of the word address 24 are records 4 , 8 , 11 , 16 , 22 and 25 , and taking a logical product (AND) between this logical negation operation 114 output and the previous operation output, i.e., publication records 3 , 4 , 5 , 13 , 14 , 16 , 19 , 21 and 25 results in surviving publications 4 , 16 and 25 .
  • word addresses which have not been previously selected/specified 110 are ignored (Don't care) during in the course of operations, resulting in a similar effect to that of a content-addressable memory (CAM) with three values. That is, the publication 16 is a result of (publication(s) containing either “information processing” or “information search”) ⁇ (publication(s) not containing “patent”) ⁇ (publication(s) containing “CPU”).
  • the above results may be sequentially read out from the operation result output 106 from a priority address encoder output circuit or the like.
  • the CPU may simply perform word address selection/specification 110 and operation specification 111 on this memory 101 to search for information of interest without having to search for the information through the entire memory space.
  • FIG. 4 shows an example of exact match data search using the memory provided with the information search function.
  • an 8-bit data is assigned to a field, wherein a word address 10 is the most significant bit (MSB) of the field and a word address 17 is the least significant bit (MSB) of the field.
  • MSB most significant bit
  • MSB 17 the least significant bit
  • the data may store data in 256 different ways, allowing 256 ways of arbitrary searches from an exact match to a size comparison by properly selecting the eight word addresses from the word address 10 to the word address 17 . For example, when performing the exact match search for a data value “10,” which is “00001010” in binary, search may be performed eight times starting with the word address 10 as the most significant bit (MSB) and the word address 17 as the least significant bit (MSB).
  • the operation is sequentially performed from the MSB word address 10 in this example.
  • “0” and “1” of “00001010” are searched with logical negation and positive logic, respectively, and logical product operations (tournament operations) are repeated eight times to obtain surviving two records 13 and 25 , both of which has the data value “10.”
  • FIG. 5 shows an example of range data search using the memory provided with the information search function.
  • records with the data values equal to or greater than “10” and less than “16” may be obtained by taking a logical product between a logical sum of word addresses 15 and 16 , and a word address 14 ; and records with the data value equal to or greater than “10” may be obtained by taking a logical sum of the records with the data value equal to or greater than “10” and less than “16” and the previously obtained records with the data value equal to or greater than “16.” Further, by negating records with the data value equal to or greater than “10,” records with the data value less than “10,” i.e., equal to or less than “9” may be detected. Other data values or ranges may be searched by repeating single-bit-based operations analogous to the above.
  • the above operations may be performed about 10 times to generate results of all records processed in parallel. For 16-bit and 32-bit data values, it takes twice and four times as many operations of the above, respectively, to achieve the exact match or range searches. Also, operations may be kept extremely simple when the data width is increased from 8 bits to 9 bits or 10 bits, and the data width may be changed to even 17 bits, 33 bits, etc. with the word addresses not necessarily being continuous.
  • this memory has no limitation of in-field assignment as to searches for single-bit data with “1” or “0” to searches for a range with any data width.
  • the memory may be configured to divide the operation result output 106 such as a priority address encoder output circuit into several blocks, and read out at per block basis.
  • Data may be written and read using a data bus with a constant width, and may be serially written and read using a first in, first out (FIFO) interface in vertical and horizontal directions. In this case, several operations may be processed in parallel.
  • FIFO first in, first out
  • FIG. 6 shows an example of the present memories connected in series and parallel.
  • the present memories may be mutually connected to expand the system in the vertical (word address) direction and the horizontal (data width) direction to thereby allow extremely simple system expansion while providing durability to the system.
  • required memory capacity is determined in the vertical (word address) direction and the horizontal (data width) direction.
  • the required number of word addresses may be several hundred-thousands for the full-text search example discussed above, whereas it may be several thousands to several tens of thousands for the personal information example.
  • the overall connected memory may be processed in parallel no matter how its data is arranged in serial and parallel even for 10 TB data, for example, by repeating the word address specification 110 and the operation specification 111 several times, several dozen times or several hundred times.
  • An access speed depends on a storage element, but if one logical operation takes 10 ns, for example, 100,000 operations may be possible within an information search time from several hundred ns, several ⁇ s to 1 ms.
  • This fully parallel processing allows information search on any big data within a fixed amount of time from several hundred ns, several ⁇ s to about 1 ms, and that is the most important characteristic of the present technology.
  • the concept of the present invention reverses the conventional notion of memory structure and vertical and horizontal relationship of data, and clearly demonstrates that it can significantly reduce the number of information processing procedures and the processing time. This is extremely effective in data mining on big data, where searches have to be repeated based on various assumptions.
  • FIG. 7 shows an example of hierarchical connections of the memories each provided with the information search function.
  • a memory 101 at the top provided with the information search function is a master memory, and each of its records corresponds with more detailed data stored in respective sub memories, wherein the memory 101 is also provided with a sub information search function for the sub memories.
  • big data of any scale may be searched by utilizing the above hierarchical database.
  • This memory 101 may be used to build a database by simply assigning its records and fields, and determining the operation specifications 111 for the logical operation units 105 to use the database. Therefore, this memory 101 eliminates the need for conventional search algorithms and database concepts themselves such as SQL and the like.
  • This algorithm is one of the most established information processing techniques as it can significantly reduce the number of information data searches, but it requires preprocessing such as arranging data in ascending order when writing data values in a data table on a memory, and also requires data rearrangement (data maintenance) on the memory every time data increases or decreases. That is, this algorithm reduces the CPU's load when it searches for a particular data value, but the load for the preprocessing and the data maintenance is never small.
  • the above algorithms become unnecessary, eliminating the need for information processing such as the preprocessing and the data maintenance.
  • Data maintenance such as data rearrangement becomes entirely unnecessary because data may be registered or deleted by simply specifying the record(s) and/or field(s) to be registered or deleted. Therefore, a CPU for controlling the present memory 101 and the entire information processing does not need to be fast, allowing a significant reduction in information processing, especially electric power for information processing. As a result, the CPU load will be eliminated and technologies used for peripherals and information searches will be simplified.
  • Japanese Patent No. 4588114 “Memory Provided with Information Refinement Detection Function” discloses a memory with a strong capability of logical product operations such as pattern matching.
  • PCT/JP2013/059260 “Memory Provided with Set Operation Function” discloses a memory capable of expanding the concept of the above memory provided with the information refinement detection function to thereby freely enable logical product operations, logical sum operations and logical negation operations, etc.
  • the foregoing inventions are both applied technologies based on the content-addressable memory (CAM) and capable of utilizing the present invention.
  • a memory provided with an information refinement detection function, or a memory provided with a set operation function may be realized.
  • the present scheme Compared to typical content-addressable memories (CAMs), the present scheme has a somewhat reduced speed of parallel operations themselves, but it can implement a large-scale equivalent content-addressable memory (CAM) to thereby dramatically reduce the processing time of overall information search.
  • CAM content-addressable memory
  • Such a memory 101 is exceptionally effective for genome analysis.
  • human DNA has big data of several G base information. Analyses on such big data require extremely long time. If several G base information is stored in this memory 101 , its DNA analysis will be faster and more accurate than using a supercomputer.
  • the data width of the memory 100 is a constant value such as 32 bits, 64 bits, 128 bits, etc., and a CPU sequentially accesses addresses, reads data stored therein and serially processes the data.
  • Information processing efficiency increases as the data width (bus width) widens, but at the same time, the number of device I/O pins and wiring load of the printed circuit board implementing the device increase as well, and therefore there is a limit to how wide the data bus width can be.
  • the memory 101 of the present invention vertically and horizontally reverses the conventional concept of memory structure and information processing for a database, and performs single-bit-based parallel operations. Even with single-bit-based operations, if parallel information processing can be done on massive storage cell 102 information, such information processing will be extremely efficient with its significantly reduced number of operations compared to conventional techniques, and further, as one of the characteristics of memory devices, information in a memory chip does not have to be externally retrieved, but instead may be internally processed by driving a large number of memory cells in parallel, entering (substituting) the information in the parallel logical operation units and performing parallel operations within the memory chip.
  • the present memory may be implemented in conventional DRAMs, SRAMs and flash memories as well as magnetic storage memory cells, which have been actively researched recently for their nonvolatile, power-saving features, by adding logical operation functions to only a small part of the chips, thus enabling massive, super fast and simple information processing.
  • Algorithms of this memory 101 may be easily implemented in FPGAs.
  • This memory provided with an in-memory database and a self-search function independent of a CPU, may be also combined with a CPU into an integrated device configuration.
  • the content-addressable memory has been know as a device having a fast operation function, and various applications have been studied including various knowledge processing, but it is used for only limited purposes such as a communication router for reasons including its inability to have a large storage capacity and its high power consumption.
  • the present invention is a novel type of information processing memory having a large capacity and an information search speed comparable to that of the content-addressable memory (CAM), it has a wide variety of applications such as various databases, data mining of big data, data analysis and knowledge processing.
  • CAM content-addressable memory
  • this technology may be used to dramatically reduce the power consumption of information processing, it has a great significance in eliminating environmental problems caused by ICT equipment and it will be one of novel information processing techniques.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Memory System (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

CPUs are not effective for search processing for information on a memory. Content-addressable memories (CAMs) are effective for information searches, but it is difficult to build a large-capacity memory usable for big data using the CAMs. A large-capacity memory may be turned into an active memory having an information search capability comparable to that of a content-addressable memory (CAM) by incorporating an extremely small, single-bit-based parallel logical operation unit into a common memory. With this memory, a super fast in-memory database capable of fully parallel searches may be realized.

Description

    FIELD OF THE INVENTION
  • The present invention relates to a memory provided with an information search function and the memory's usage, device and information processing method.
  • BACKGROUND OF THE INVENTION
  • With massive semiconductor memories becoming available at low prices, in-memory database technologies have become a focus of attention as fast-processing techniques for big data. One of the in-memory database technologies is used in data mining, where an enormous amount of data resides on a semiconductor memory (in-memory database) to speed up data accesses and, therefore, searches for information of interest. However, moving the information data from a hard disk device to a semiconductor memory alone only offers expectation of 1 to 2 orders of speed improvement.
  • It is necessary to clarify challenges of current von Neumann-architecture computers before considering an essence of big data usage.
  • In current computers, CPU's do all information processing regardless of whether of not the CPU's are suited for the processing they have to perform. For CPU's, for example, data items are analogous to playing cards with their faces down, and the CPU's have to turn over each card (access each address) when searching for information. When a CPU performs information processing such as one to sequentially search for specific information on a memory, an amount of information processing becomes extremely large, leading to a long overall waiting time. This is the bus bottleneck, an inevitable problem of the von Neumann-architecture computers.
  • Parallel processing (distributed processing) by each CPU is an attempt to solve these challenges, but complicates peripheral circuits to overly enlarge the system.
  • Based on such a background, various techniques (software algorithms) have been devised and utilized in order to reduce the CPU load and the number of information processing procedures for the current computers with the bus bottleneck.
  • For example, representative algorithms utilized for information search include hash table, index, tree structure, binary search and clustering algorithms and, considering their combinations, the number of possible algorithms is practically infinite. These techniques (software algorithms) are merely means for reducing the CPU load and the number of information processing procedures to thereby take full advantage of the CPUs with the above inherent problem. In other words, any of the above algorithms is a method for, for example, organizing types and in-memory locations of information pieces ahead of time, creating headers and their routes so that a CPU can easily find information, arranging data items according to their sizes.
  • According to such algorithms, the CPU load is reduced during the search, but pre- and post-processing mandates complex information processing. For example, in data insertion or deletion, data rearrangement or reordering is required every time a data item is added or deleted as the pre- and post-processing for these algorithms.
  • In order to build an optimal system for a particular database, it is necessary to select some of the software algorithms described above according to a type and/or a scale of the database, and this can be done only by experts with appropriate knowledge and experience.
  • These and other inherent problems of the current computers arise from the fact that the CPU performs all the information processing, but from a different point of view, if the memory can find specific information by itself, the information processing discussed above will totally change.
  • Content-addressable memories (CAM) exist as a technology to eliminate the above and other problems. However, content-addressable memories (CAMs) have challenges that they each needs a completely parallel comparison circuit which increases their overall circuit size, and that parallel circuits constituting the comparison circuit consume high current. For this reason, utilization of the content-addressable memories (CAMs) has been limited to special applications where super fast search is essential as in communication routers.
  • Considering the above situation, the purpose of the present invention is to provide a memory with a new concept to enable a big data search at a speed comparable to that of content-addressable memories (CAM) by simply incorporating an extremely small number of circuits into a common memory.
  • In order to attain the same objective, the present inventor has been devising various inventions. For example, Japanese Patent No. 4588114, “Memory Provided with Information Refinement Detection Function” by the present inventor discloses a memory with a strong capability of logical product operations such as pattern matching. Also, PCT/JP2013/059260 “Memory Provided with Set Operation Function” discloses a memory capable of expanding the concept of the above memory provided with the information refinement detection function to thereby freely enable logical product operations, logical sum operations and logical negation operations, etc. The disclosures of these two applications are incorporated herein by reference in their entirety.
  • A memory 101 of the present invention may be applied to the above two prior inventions.
  • Also, Japanese Patent Application No. 10-232531 “Memory with Operation Function” has an objective, as it illustrates, to improve a chip efficiency by providing an operation circuit for each block unit. This reduces the size of operation circuitry compared to providing an operation circuit for each memory, but deteriorates the operation efficiency as well as having a poor chip efficiency and undesirable cost issues.
  • Memories with an operation function disclosed in other patent publications are similar to this, and there is no prior invention found for performing parallel information processing on data in a memory with a minimum configuration of one set of operation function as the present invention does.
  • SUMMARY OF THE INVENTION
  • An object of the invention is to provide a memory provided with an information search function based on a novel information processing concept, applicable to big data and capable of fundamentally eliminating the biggest problem in CPU-based information processing, i.e., information searches, for which information processing is too complex for non-experts and burdensome to CPUs, and complicates peripheral circuits to increase required power consumption.
  • Specifically, the present memory leverages the concept of content-addressable memory (CAM) capable of parallel information searches, and addresses the CAM's challenges such as its circuit size and high power consumption to thereby provide a memory element applicable to big data and based on a novel concept free from conventional information processing scheme.
  • In order to overcome the above challenges, the following are provided according to a principal aspect of the present invention.
  • In claim 1:
  • A memory capable of reading and writing information, provided with an information search function,
  • (1) wherein the memory has one word bit width of n and N word addresses, that is a storage cell structure of N×n bits,
  • (2) the memory comprising: a logical operation unit having a set of n bits for the entirety of the memory;
  • (3) a function to enter (substitute) storage cell information having the one word bit width of n into the logical operation unit having the set of n bits, and logically operate the storage cell information, wherein the storage cell information is repeatedly selected/specified from among the N word addresses; and
  • (4) a function to output contents of the logical operation unit.
  • In claim 2:
  • The memory provided with the information search function of claim 1, wherein the logical operation unit is configured to be capable of logical storage, logical product, logical sum, logical negation, exclusive logic or any combinatorial operation thereof between single bits of the entered (substituted) memory storage cell information having the one word bit width of n.
  • In claim 3:
  • The memory provided with the information search function of claim 1, wherein the logical operation unit is provided with a shift register function.
  • In claim 4:
  • The memory provided with the information search function of claim 1, wherein the memory performs an exact match search for a data value and a range search for data values using the logical storage, logical product, logical sum, logical negation, exclusive logic or any combinatorial operation thereof.
  • In claim 5:
  • The memory provided with the information search function of claim 1, wherein the memory performs an addition or a subtraction of data values using the logical storage, logical product, logical sum, logical negation, exclusive logic or any combinatorial operation thereof.
  • In claim 6:
  • The memory provided with the information search function of claim 1, wherein the memory is integrated with a semiconductor device of another type such as a CPU.
  • In claim 7:
  • The memory provided with the information search function of claim 1, wherein the memory is implemented in an FPGA.
  • In claim 8:
  • A method for using the memory provided with the information search function of claim 1, comprising the step of assigning a database record to any one of columns within the one word bit width of n to generate a database, wherein a total bit number of fields per record of the database equals the number of word addresses N.
  • In claim 9:
  • A method for using the memory provided with the information search function of claim 1, comprising the step of:
  • connecting for use, memories each provided with the information search function
  • (1) in series, in parallel or in series and parallel;
  • (2) hierarchically; or
  • (1) or (2).
  • In claim 10:
  • A device including the memory of claim 1.
  • In claim 11:
  • An information processing method comprising the step of repeating logical product, logical sum, logical negation, exclusive logic or any combinatorial logical operation thereof between single bits of the memory storage cell information within the memory to thereby obtain a prescribed operation result.
  • In claim 12:
  • The information processing method of claim 11, further comprising the step of performing the logical operation in parallel between the single bits of the information.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a structural view of a common memory;
  • FIG. 2 is a structural view of a memory provided with an information search function;
  • FIG. 3 is an example of publication search using the memory provided with the information search function (Example 1);
  • FIG. 4 is an example of exact match data search using the memory provided with the information search function (Example 2);
  • FIG. 5 is an example of range data search using the memory provided with the information search function (Example 3);
  • FIG. 6 is an example of serial and parallel connections of the memories each provided with the information search function (Example 4); and
  • FIG. 7 is an example of hierarchical connections of the memories each provided with the information search function (Example 5).
  • DETAILED DESCRIPTION OF THE INVENTION
  • One embodiment of the present invention will be described below in accordance with accompanying drawings.
  • FIG. 1 is a structural view showing a general memory according to this one embodiment.
  • In the memory 100 of FIG. 1, functional circuits such as an address decoder and a data bus are omitted, wherein this memory 100 is configured so that information data may be freely written into and read out of this memory, wherein this memory 100 consists of a storage cells 102 consisting of N×n bits of cells, wherein the storage cells 102 comprise N word addresses 104 with each word having a width 103 of n bits, and wherein generally the 1-N word addresses may be externally selected/specified with means such as an address decoder.
  • In the current information processing done by a CPU, the data width 103 of the memory 100 is a constant value such as 8 bits, 16 bits, 32 bits, etc., and for information data searches, the CPU sequentially accesses addresses in the memory's address space, reads data stored therein and serially processes the data, wherein a given number of addresses in the memory's address space is 1 M, 1 G, or the like.
  • Information processing by the memory of the present invention is based on a concept which defies the common sense of data width and address of the above general memory structure and database table structure, and is based on single-bit-based parallel logic operations.
  • FIG. 2 shows an example memory structure provided with an information search function according to the present invention.
  • Similarly, to FIG. 1, in FIG. 2, functional circuits such as an address decoder and a data bus are omitted, wherein this memory 101 is configured so that information data may be freely written into and read out of this memory, wherein this memory 101 consists of storage cells 102 consisting of N×n bits of cells, wherein the storage cells 102 comprise N word addresses 104 with each word having a width 103 of n bits, and wherein the 1-N word addresses may be externally selected/specified 110. The one word bit width of n 103 corresponds with a number of database records (n), and for the ease of understanding, one may analogously consider that each record is arranged as a column and N of the word address 104 corresponds to fields in each record.
  • In other words, this memory is a data table having n records with each record having N bits. A logical operation unit 105 is provided in parallel to n-bit storage cells 102 in this memory's row (horizontal) direction, which n-bit storage cells 102 may be selected/specified 110 using the word address 104, and this logical operation unit 105 is configured so that circuits capable of logical storage 116 for respective n-bit storage cells 102 of a selected/specified 110 word address 104, logical product 112, logical sum 113, logical negation (NOT) 114, exclusive logic 115 and combinations thereof may be freely specified.
  • Also, an operation result output 106 function of a priority address encoder output circuit or the like is provided in order to output operation results of this logical operation unit 105. This memory is mostly made of memory cells themselves and only small part of this memory is dedicated for the logical operation unit 105 and the operation result output 106 function, and therefore, by incorporating these functions in a miniscule area of respective general memories, those general memories may become high-capacity memories usable for databases.
  • Of course for faster operations, the memory may be adapted so that multiple word addresses may be selected/specified 110 simultaneously and multiple logical operation units 105 may be provided.
  • Next, an example is considered where the present invention is implemented with a DRAM. A current memory capacity per semiconductor die (chip) is about 8 Gbits.
  • With the assumption that the scale of circuits for the logical operation unit 105 and operation result output 106 functions is negligible, different combinations of the number of word addresses and the word width may be implemented for the main memory 101, for example, 1 M word addresses and 8 Kbit word width, or 8 K word addresses and 1 Mbit word width. This memory 101 may be effective for any information search, especially, big data search, data mining, genome analysis, etc., and some practical examples of this memory will be discussed below.
  • One example of our everyday search is an Internet search and its concept is embodied by information refinement using a keyword. For example, by providing a keyword such as “information processing,” “information search” or “CPU,” the search is refined and suitable Internet sites may be found.
  • Exemplary embodiments of utilizing this memory for publication search will be discussed below.
  • Example 1
  • FIG. 3 shows an example of publication search using a memory provided with an information search function.
  • In this example, each of word addresses from 1 to N may be assigned to one vocabulary term such as “information processing,” “information search,” “patent” or “CPU,” and each of vertical lines (columns) of word width n may be assigned to one publication. Also, the word addresses 1-N constitute one record as a whole, and the word addresses 1-N correspond to fields. In other words, if any of the terms including “information processing,” “information search,” “patent” and “CPU,” exist in one publication, “1” will be written into memory cells (fields) corresponding to the existing terms (“0” is omitted and likewise hereafter). Therefore, in this example, N vocabulary terms and n publications (n records) are registered in a database.
  • An example of searching for a particular publication from within this stored database will be discussed below.
  • For registered database keywords, a word address 18 corresponds to “information processing,” a word address 5 corresponds to “information search,” a word address 24 corresponds to “patent” and a word address 10 corresponds to “CPU,” wherein here we assume that the search operation formula is (publication(s) containing either “information processing” or “information search”)×(publication(s) not containing “patent”)×(publication(s) containing “CPU”).
  • An operation process for the above keyword search is shown at the lower section of FIG. 3.
  • Records of publications containing either “information processing” of the word address 18 or “information search” of the word address 5 (logical sum (OR)) are records 3, 4, 5, 13, 14, 16, 19, 21 and 25. Next, records of publications not containing “patent” of the word address 24 are records 4, 8, 11, 16, 22 and 25, and taking a logical product (AND) between this logical negation operation 114 output and the previous operation output, i.e., publication records 3, 4, 5, 13, 14, 16, 19, 21 and 25 results in surviving publications 4, 16 and 25. Lastly, taking a logical product (AND) between publication records 3, 7, 9, 12, 15, 16 and 22 containing “CPU” of the word address 10 and the previous surviving publications records yields a final surviving publication 107 of record 16
  • As can be seen from the foregoing discussion, word addresses which have not been previously selected/specified 110 are ignored (Don't care) during in the course of operations, resulting in a similar effect to that of a content-addressable memory (CAM) with three values. That is, the publication 16 is a result of (publication(s) containing either “information processing” or “information search”)×(publication(s) not containing “patent”)×(publication(s) containing “CPU”).
  • The above results may be sequentially read out from the operation result output 106 from a priority address encoder output circuit or the like.
  • The CPU may simply perform word address selection/specification 110 and operation specification 111 on this memory 101 to search for information of interest without having to search for the information through the entire memory space.
  • Although the above example was discussed on full-text searches, it may be utilized for a database for Internet search by populating the records with URLs.
  • Also, the above full-text searches were all done on single-bit-based data, but data subject to search will be values in the next exemplary embodiment below.
  • Example 2
  • FIG. 4 shows an example of exact match data search using the memory provided with the information search function.
  • In this example, an 8-bit data is assigned to a field, wherein a word address 10 is the most significant bit (MSB) of the field and a word address 17 is the least significant bit (MSB) of the field. Since the data has 8 bits, it may store data in 256 different ways, allowing 256 ways of arbitrary searches from an exact match to a size comparison by properly selecting the eight word addresses from the word address 10 to the word address 17. For example, when performing the exact match search for a data value “10,” which is “00001010” in binary, search may be performed eight times starting with the word address 10 as the most significant bit (MSB) and the word address 17 as the least significant bit (MSB).
  • As shown in the lower section of FIG. 4, the operation is sequentially performed from the MSB word address 10 in this example. In this case, “0” and “1” of “00001010” are searched with logical negation and positive logic, respectively, and logical product operations (tournament operations) are repeated eight times to obtain surviving two records 13 and 25, both of which has the data value “10.” When incorporating adding and subtracting functions to the logical operation unit 105 for parallel operations, four arithmetic operations on record data will be available.
  • Example 3
  • FIG. 5 shows an example of range data search using the memory provided with the information search function.
  • The above described the exact match search for the data value “10,” but when searching for data values equal to or greater than “10,” records with the data values equal to or greater than “16” may be detected by taking a logical sum four times from the MSB of the word address 10 to a word address 13, as shown in the figure.
  • Additionally for the lower 4 bits, records with the data values equal to or greater than “10” and less than “16” may be obtained by taking a logical product between a logical sum of word addresses 15 and 16, and a word address 14; and records with the data value equal to or greater than “10” may be obtained by taking a logical sum of the records with the data value equal to or greater than “10” and less than “16” and the previously obtained records with the data value equal to or greater than “16.” Further, by negating records with the data value equal to or greater than “10,” records with the data value less than “10,” i.e., equal to or less than “9” may be detected. Other data values or ranges may be searched by repeating single-bit-based operations analogous to the above.
  • The above operations may be performed about 10 times to generate results of all records processed in parallel. For 16-bit and 32-bit data values, it takes twice and four times as many operations of the above, respectively, to achieve the exact match or range searches. Also, operations may be kept extremely simple when the data width is increased from 8 bits to 9 bits or 10 bits, and the data width may be changed to even 17 bits, 33 bits, etc. with the word addresses not necessarily being continuous.
  • In other words, this memory has no limitation of in-field assignment as to searches for single-bit data with “1” or “0” to searches for a range with any data width.
  • In case of personal information as an example to show an important characteristic of this memory, information may be refined by searching for “males who live in Chiba, who work in Tokyo, who are 170 cm to 175 cm tall and who are 50 years old to 60 years old” for example, to thereby detect only required records. Considering that a large number of records are subject to searches, the memory may be configured to divide the operation result output 106 such as a priority address encoder output circuit into several blocks, and read out at per block basis.
  • Data may be written and read using a data bus with a constant width, and may be serially written and read using a first in, first out (FIFO) interface in vertical and horizontal directions. In this case, several operations may be processed in parallel.
  • Example 4
  • FIG. 6 shows an example of the present memories connected in series and parallel.
  • Each being fully independent, the present memories may be mutually connected to expand the system in the vertical (word address) direction and the horizontal (data width) direction to thereby allow extremely simple system expansion while providing durability to the system. According to a data type, required memory capacity is determined in the vertical (word address) direction and the horizontal (data width) direction. The required number of word addresses may be several hundred-thousands for the full-text search example discussed above, whereas it may be several thousands to several tens of thousands for the personal information example.
  • Usually, in order for one CPU to find particular information from a memory without any arrangement definitions or indices, it takes about 10 ms for 1 M word addresses, about 10 s for 1 G word addresses and about 10,000 s (about 3 hours) for 1 T word addresses just to access and match the memory contents at 10 ns/procedure. If CPUs are used in parallel with distributed processing, the processing time may be essentially reduced in inverse proportion to the number of CPUs. However, it is difficult to search through a massive database in realtime (e.g., within 1 second).
  • In the present example, the overall connected memory may be processed in parallel no matter how its data is arranged in serial and parallel even for 10 TB data, for example, by repeating the word address specification 110 and the operation specification 111 several times, several dozen times or several hundred times.
  • An access speed depends on a storage element, but if one logical operation takes 10 ns, for example, 100,000 operations may be possible within an information search time from several hundred ns, several μs to 1 ms. This fully parallel processing allows information search on any big data within a fixed amount of time from several hundred ns, several μs to about 1 ms, and that is the most important characteristic of the present technology.
  • The concept of the present invention reverses the conventional notion of memory structure and vertical and horizontal relationship of data, and clearly demonstrates that it can significantly reduce the number of information processing procedures and the processing time. This is extremely effective in data mining on big data, where searches have to be repeated based on various assumptions.
  • Example 5
  • FIG. 7 shows an example of hierarchical connections of the memories each provided with the information search function.
  • In the example of FIG. 7, a memory 101 at the top provided with the information search function is a master memory, and each of its records corresponds with more detailed data stored in respective sub memories, wherein the memory 101 is also provided with a sub information search function for the sub memories. In particular, big data of any scale may be searched by utilizing the above hierarchical database.
  • This memory 101 may be used to build a database by simply assigning its records and fields, and determining the operation specifications 111 for the logical operation units 105 to use the database. Therefore, this memory 101 eliminates the need for conventional search algorithms and database concepts themselves such as SQL and the like. For information search using a CPU, various techniques exist for reducing the CPU's load during its use. The binary search is one typical example of such techniques.
  • This algorithm is one of the most established information processing techniques as it can significantly reduce the number of information data searches, but it requires preprocessing such as arranging data in ascending order when writing data values in a data table on a memory, and also requires data rearrangement (data maintenance) on the memory every time data increases or decreases. That is, this algorithm reduces the CPU's load when it searches for a particular data value, but the load for the preprocessing and the data maintenance is never small.
  • The binary search has been discussed above, but other algorithms such as hash tables have similar problems.
  • With the present invention, the above algorithms become unnecessary, eliminating the need for information processing such as the preprocessing and the data maintenance. Data maintenance such as data rearrangement becomes entirely unnecessary because data may be registered or deleted by simply specifying the record(s) and/or field(s) to be registered or deleted. Therefore, a CPU for controlling the present memory 101 and the entire information processing does not need to be fast, allowing a significant reduction in information processing, especially electric power for information processing. As a result, the CPU load will be eliminated and technologies used for peripherals and information searches will be simplified.
  • The present inventor has been conducting research on various memory devices. Japanese Patent No. 4588114, “Memory Provided with Information Refinement Detection Function” discloses a memory with a strong capability of logical product operations such as pattern matching.
  • Also, PCT/JP2013/059260 “Memory Provided with Set Operation Function” discloses a memory capable of expanding the concept of the above memory provided with the information refinement detection function to thereby freely enable logical product operations, logical sum operations and logical negation operations, etc.
  • The foregoing inventions are both applied technologies based on the content-addressable memory (CAM) and capable of utilizing the present invention.
  • As an example, by combining the logical operation unit 105 with a shift register function, a memory provided with an information refinement detection function, or a memory provided with a set operation function may be realized.
  • Compared to typical content-addressable memories (CAMs), the present scheme has a somewhat reduced speed of parallel operations themselves, but it can implement a large-scale equivalent content-addressable memory (CAM) to thereby dramatically reduce the processing time of overall information search.
  • Such a memory 101 is exceptionally effective for genome analysis. Among others, human DNA has big data of several G base information. Analyses on such big data require extremely long time. If several G base information is stored in this memory 101, its DNA analysis will be faster and more accurate than using a supercomputer. In the current information processing, the data width of the memory 100 is a constant value such as 32 bits, 64 bits, 128 bits, etc., and a CPU sequentially accesses addresses, reads data stored therein and serially processes the data. Information processing efficiency increases as the data width (bus width) widens, but at the same time, the number of device I/O pins and wiring load of the printed circuit board implementing the device increase as well, and therefore there is a limit to how wide the data bus width can be.
  • As one of the characteristics of the memory 101 of the present invention, it vertically and horizontally reverses the conventional concept of memory structure and information processing for a database, and performs single-bit-based parallel operations. Even with single-bit-based operations, if parallel information processing can be done on massive storage cell 102 information, such information processing will be extremely efficient with its significantly reduced number of operations compared to conventional techniques, and further, as one of the characteristics of memory devices, information in a memory chip does not have to be externally retrieved, but instead may be internally processed by driving a large number of memory cells in parallel, entering (substituting) the information in the parallel logical operation units and performing parallel operations within the memory chip.
  • The present memory may be implemented in conventional DRAMs, SRAMs and flash memories as well as magnetic storage memory cells, which have been actively researched recently for their nonvolatile, power-saving features, by adding logical operation functions to only a small part of the chips, thus enabling massive, super fast and simple information processing.
  • Algorithms of this memory 101 may be easily implemented in FPGAs.
  • This memory, provided with an in-memory database and a self-search function independent of a CPU, may be also combined with a CPU into an integrated device configuration.
  • Since before, the content-addressable memory (CAM) has been know as a device having a fast operation function, and various applications have been studied including various knowledge processing, but it is used for only limited purposes such as a communication router for reasons including its inability to have a large storage capacity and its high power consumption.
  • Since the present invention is a novel type of information processing memory having a large capacity and an information search speed comparable to that of the content-addressable memory (CAM), it has a wide variety of applications such as various databases, data mining of big data, data analysis and knowledge processing.
  • Also, since this technology may be used to dramatically reduce the power consumption of information processing, it has a great significance in eliminating environmental problems caused by ICT equipment and it will be one of novel information processing techniques.

Claims (12)

What is claimed is:
1. A memory capable of reading and writing information, provided with an information search function,
(1) wherein the memory has one word bit width of n and N word addresses, that is a storage cell structure of N×n bits,
(2) the memory comprising: a logical operation unit having a set of n bits for the entirety of the memory;
(3) a function to enter (substitute) storage cell information having the one word bit width of n into the logical operation unit having the set of n bits, and logically operate the storage cell information, wherein the storage cell information is repeatedly selected/specified from among the N word addresses; and
(4) a function to output contents of the logical operation unit.
2. The memory provided with the information search function of claim 1, wherein the logical operation unit is configured to be capable of logical storage, logical product, logical sum, logical negation, exclusive logic or any combinatorial operation thereof between single bits of the entered (substituted) memory storage cell information having the one word bit width of n.
3. The memory provided with the information search function of claim 1, wherein the logical operation unit is provided with a shift register function.
4. The memory provided with the information search function of claim 1, wherein the memory performs an exact match search for a data value and a range search for data values using the logical storage, logical product, logical sum, logical negation, exclusive logic or any combinatorial operation thereof.
5. The memory provided with the information search function of claim 1, wherein the memory performs an addition or a subtraction of data values using the logical storage, logical product, logical sum, logical negation, exclusive logic or any combinatorial operation thereof.
6. The memory provided with the information search function of claim 1, wherein the memory is integrated with a semiconductor device of another type such as a CPU.
7. The memory provided with the information search function of claim 1, wherein the memory is implemented in an FPGA.
8. A method for using the memory provided with the information search function of claim 1, comprising the step of assigning a database record to any one of columns within the one word bit width of n to generate a database, wherein a total bit number of fields per record of the database equals the number of word addresses N.
9. A method for using the memory provided with the information search function of claim 1, comprising the step of:
connecting for use, memories each provided with the information search function
(1) in series, in parallel or in series and parallel;
(2) hierarchically; or
(1) or (2).
10. A device including the memory of claim 1.
11. An information processing method comprising the step of repeating logical product, logical sum, logical negation, exclusive logic or any combinatorial logical operation thereof between single bits of the memory storage cell information within the memory to thereby obtain a prescribed operation result.
12. The information processing method of claim 11, further comprising the step of performing the logical operation in parallel between the single bits of the information.
US15/425,555 2013-12-23 2017-02-06 Memory equipped with information retrieval function, method for using same, device, and information processing method Abandoned US20170147712A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/425,555 US20170147712A1 (en) 2013-12-23 2017-02-06 Memory equipped with information retrieval function, method for using same, device, and information processing method

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
JP2013264763 2013-12-23
JP2013-264763 2013-12-23
PCT/JP2014/083607 WO2015098707A1 (en) 2013-12-23 2014-12-18 Memory equipped with information retrieval function, method for using same, device, and information processing method
US201615107565A 2016-06-23 2016-06-23
US15/425,555 US20170147712A1 (en) 2013-12-23 2017-02-06 Memory equipped with information retrieval function, method for using same, device, and information processing method

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
PCT/JP2014/083607 Continuation WO2015098707A1 (en) 2013-12-23 2014-12-18 Memory equipped with information retrieval function, method for using same, device, and information processing method
US15/107,565 Continuation US9627065B2 (en) 2013-12-23 2014-12-18 Memory equipped with information retrieval function, method for using same, device, and information processing method

Publications (1)

Publication Number Publication Date
US20170147712A1 true US20170147712A1 (en) 2017-05-25

Family

ID=53478566

Family Applications (2)

Application Number Title Priority Date Filing Date
US15/107,565 Active US9627065B2 (en) 2013-12-23 2014-12-18 Memory equipped with information retrieval function, method for using same, device, and information processing method
US15/425,555 Abandoned US20170147712A1 (en) 2013-12-23 2017-02-06 Memory equipped with information retrieval function, method for using same, device, and information processing method

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US15/107,565 Active US9627065B2 (en) 2013-12-23 2014-12-18 Memory equipped with information retrieval function, method for using same, device, and information processing method

Country Status (5)

Country Link
US (2) US9627065B2 (en)
EP (1) EP3091448A4 (en)
JP (2) JP5981666B2 (en)
CN (1) CN105900085B (en)
WO (1) WO2015098707A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
RU2652501C1 (en) * 2017-06-14 2018-04-26 Виктория Владимировна Олевская Module for searching the information unit on input data

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2018169636A (en) * 2015-07-23 2018-11-01 井上 克己 Data retrieval apparatus, and data retrieval lsi
US20200410039A1 (en) * 2016-11-28 2020-12-31 Katsumi Inoue Data comparison arithmetic processor and method of computation using same
KR102631380B1 (en) * 2018-05-17 2024-02-01 에스케이하이닉스 주식회사 Semiconductor system including various memory deivces capable of processing data
JP6986309B1 (en) * 2021-09-22 2021-12-22 明男 三水 Data processing equipment, data processing methods, and programs

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2718062B2 (en) * 1988-05-25 1998-02-25 日本電気株式会社 Learning type character string search device and control method of the device
US5325500A (en) * 1990-12-14 1994-06-28 Xerox Corporation Parallel processing units on a substrate, each including a column of memory
JP3595565B2 (en) * 1993-12-24 2004-12-02 シャープ株式会社 Semiconductor memory cell circuit and memory cell array
JP2000067573A (en) 1998-08-19 2000-03-03 Mitsubishi Electric Corp Memory provided with arithmetic function
EP1182577A1 (en) * 2000-08-18 2002-02-27 SER Systeme AG Produkte und Anwendungen der Datenverarbeitung Associative memory
JP2006127460A (en) * 2004-06-09 2006-05-18 Renesas Technology Corp Semiconductor device, semiconductor signal processing apparatus and crossbar switch
US8341362B2 (en) * 2008-04-02 2012-12-25 Zikbit Ltd. System, method and apparatus for memory with embedded associative section for computations
CN101577757A (en) * 2009-03-05 2009-11-11 上海闻泰电子科技有限公司 Method for realizing short message retrieval function
WO2011022114A1 (en) * 2009-08-20 2011-02-24 Rambus Inc. Atomic memory device
CA2790009C (en) * 2010-02-18 2017-01-17 Katsumi Inoue Memory having information refinement detection function, information detection method using memory, device including memory, information detection method, method for using memory, and memory address comparison circuit
JP4588114B1 (en) 2010-02-18 2010-11-24 克己 井上 Memory having information narrowing detection function, method of using the same, and device including the memory.
JP6014120B2 (en) 2012-03-28 2016-10-25 井上 克己 Memory having set operation function and set operation processing method using the same

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
RU2652501C1 (en) * 2017-06-14 2018-04-26 Виктория Владимировна Олевская Module for searching the information unit on input data

Also Published As

Publication number Publication date
JPWO2015098707A1 (en) 2017-03-23
WO2015098707A1 (en) 2015-07-02
CN105900085A (en) 2016-08-24
US20160322105A1 (en) 2016-11-03
JP5981666B2 (en) 2016-08-31
CN105900085B (en) 2019-08-09
JP2017016668A (en) 2017-01-19
EP3091448A1 (en) 2016-11-09
EP3091448A4 (en) 2017-12-06
US9627065B2 (en) 2017-04-18
JP6229024B2 (en) 2017-11-08

Similar Documents

Publication Publication Date Title
US20170147712A1 (en) Memory equipped with information retrieval function, method for using same, device, and information processing method
US8711638B2 (en) Using storage cells to perform computation
US11604834B2 (en) Technologies for performing stochastic similarity searches in an online clustering space
US7565482B1 (en) Method and device for scalable multiple match extraction from search data
US20140122509A1 (en) System, method, and computer program product for performing a string search
Chowdhury et al. A DNA read alignment accelerator based on computational RAM
US11327881B2 (en) Technologies for column-based data layouts for clustered data systems
KR20210135920A (en) Technologies for refining stochastic similarity search candidates
CN103914483A (en) File storage method and device and file reading method and device
US20200264874A1 (en) Technologies for performing random sparse lifting and procrustean orthogonal sparse hashing using column read-enabled memory
US11392494B2 (en) Technologies for performant column read operations on clustered data in a dimm architecture
US11258539B2 (en) Technologies for performing encoding of data symbols for column read operations
Imani et al. CAP: Configurable resistive associative processor for near-data computing
JP2003256265A (en) Search memory, controller for memory search, and memory search method
US20200410039A1 (en) Data comparison arithmetic processor and method of computation using same
CN102663051B (en) Method and system for searching content addressable memory
WO2016199808A1 (en) Memory type processor, device including memory type processor, and method for using same
CN102937969A (en) Method for quickly searching CAM (Central Address Memory)
Ipek et al. Resistive memories in associative computing
Imani et al. In-Memory Processing to Support Search-Based and Bitwise Computation
JP2018169636A (en) Data retrieval apparatus, and data retrieval lsi
Hilimire Strategies for Targeting the Frameshift Stimulatory RNA of HIV-1 with Synthetic Molecules
JPH0272481A (en) Character string retrieving device by logical expression and control system for such device
US20040168022A1 (en) Pipeline scalable architecture for high density and high speed content addressable memory (CAM)
CN103839581A (en) Intelligent storage apparatus

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: EXPRESSLY ABANDONED -- DURING EXAMINATION