US20070250593A1 - Reconfigurable semantic processor - Google Patents
Reconfigurable semantic processor Download PDFInfo
- Publication number
- US20070250593A1 US20070250593A1 US11/766,733 US76673307A US2007250593A1 US 20070250593 A1 US20070250593 A1 US 20070250593A1 US 76673307 A US76673307 A US 76673307A US 2007250593 A1 US2007250593 A1 US 2007250593A1
- Authority
- US
- United States
- Prior art keywords
- client
- parsing
- data
- server
- storage server
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/10—File systems; File servers
- G06F16/18—File system types
- G06F16/182—Distributed file systems
- G06F16/1824—Distributed file systems implemented using Network-attached Storage [NAS] architecture
- G06F16/1827—Management specifically adapted to NAS
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/10—File systems; File servers
- G06F16/14—Details of searching files based on file metadata
- G06F16/148—File search processing
Definitions
- This invention relates generally to storage servers, and more specifically to storage server implementations using digital semantic processors.
- a storage server is a networked computer that provides media and file access to other computers and computing devices, referred to herein as “clients.”
- the storage server may be accessible on a Wide-Area Network (WAN), a Local-Area Network (LAN), and/or share a dedicated point-to-point connection with another computing device.
- WAN Wide-Area Network
- LAN Local-Area Network
- Storage servers can be of several types, including a Network Attached Storage (NAS) server, a Storage Area Network (SAN) server, and an application-style server.
- NAS Network Attached Storage
- SAN Storage Area Network
- application-style server an application-style server.
- a NAS server comprehends directory structures. Requests to a NAS server generally specify a file path/filename and action such as create, read, write, append, etc. A NAS server converts the requests to one or more disk sector/block disk access transactions, accesses the disk (or other storage media), and performs the requested file-level transaction. Such a server maintains and utilizes a directory structure to locate the file starting locations, and a File Allocation Table to reconstruct segmented files and locate free sectors on a disk or other writeable media.
- a SAN server generally receives direct requests for disk sector/block-based transactions, and may have no knowledge of a file system or relationships between data blocks. Although block requests can relate to physical block locations on a media device, more typically a SAN server supports logical blocks—thus allowing the server to make one physical disk drive appear as several logical drives, or several physical disk drives appear as one logical drive, as in a RAID (Redundant Array of Independent Disks) scheme.
- RAID Redundant Array of Independent Disks
- An application-style server generally provides access to media in a media-specific format.
- an application-style server could provide access to streaming audio or video without the explicit knowledge of a filepath/filename by the client (who may not know the filename or have the authority to access a file directly), while reading the streaming media from files and encapsulating it in a streaming packet format.
- FIG. 1 shows a block diagram for a typical storage server 20 .
- a Central Processing Unit (CPU) 22 typically one or more microprocessors, operates the server according to stored program instructions.
- This type of processor is often called a von Neumann (VN) processor or machine, after its innovator, who proposed its execution style of processing sequential instructions.
- VN von Neumann
- CPU 22 connects through a Frontside Bus (FSB) 26 to a Memory Controller/Hub (MCH) 24 , which is responsible for interfacing CPU 22 to other system components that store microprocessor program instructions and store, supply, and/or consume data.
- MCH 24 manages system memory 30 through memory bus 32 .
- MCH 24 also communicates with PCI (Peripheral Component Interconnect) bridge 40 across hub bus 42 to move data between PCI-connected devices and either memory 30 or CPU 22 .
- PCI Peripheral Component Interconnect
- NIC network interface card
- ATA AT Attachment
- NIC 50 allows the server to attach directly to a Local Area Network (LAN) or other network, a Fibre Channel switch fabric, or other point-to-point connection to a client or another device that provides network visibility for the server.
- NIC 50 thereby provides a communication path for the server to receive data requests from clients and respond to those requests.
- the actual network physical communication protocol can be varied by inserting a different NIC, e.g., for whatever wired or wireless communication protocol is to be used, and loading a software device driver for that NIC onto CPU 22 .
- ATA controller 60 provides at least one ATA bus 62 for attaching media devices 64 such as hard disks, optical disks, and/or tape drives. Each ATA bus 62 allows either one or two media devices 64 to attach to storage server 20 . Some servers may employ other controllers, such as a SATA (serial ATA) controller or a SCSI (Small Computer System Interface) host adapter.
- SATA serial ATA
- SCSI Serial Computer System Interface
- NFS Networked File System
- CIFS Common Internet File System
- NFS Network File System
- TCP/IP Transmission Control Protocol
- RPC Remote Procedure Call
- XDR Extra Data Representation
- VFS Virtual File System
- NFS server software For each NFS data transaction.
- CPU 22 must execute a process for each of these software entities, switching context between them as required.
- storage server 20 operates at relatively high speed and power, requiring forced air cooling and a substantial power supply to operate the microprocessor, chipset, network interface, memory, other peripherals, and cooling fans.
- the current state of the art with a 1 Gbps (Gigabit per second) full duplex network connection, requires about 300 watts of power, occupies a volume of 800 cubic inches, and costs about $1500 without storage media.
- the present invention provides a different architecture for a storage server, using in exemplary embodiments what is referred to generally as a semantic processor.
- a semantic processor Such a device is configurable, and preferably reconfigurable like a VN machine, as its processing depends on its “programming.”
- a “semantic processor” contains at least two components: a direct execution parser to parse an input stream according to a defined grammar; and an execution engine capable of manipulating data (e.g., data movement, mathematical, and logical operations) in response to requests from the direct execution parser.
- the programming of a semantic processor differs from the conventional machine code used by a VN machine.
- the protocol input parameters of a data packet are sequentially compared to all possible inputs using sequential machine instructions, with branching, looping, and thread-switching inefficiencies.
- the semantic processor responds directly to the semantics of an input stream.
- the “code” segments that the semantic processor executes are driven directly by the input data. For instance, of the 75 or so possible CIFS commands that could appear in a packet, a single parsing cycle of an embodiment described below is all that is necessary to allow the semantic processor to load grammar and microinstructions pertinent to the actual CIFS command sent in the packet.
- a semantic processor can perform many of the functions of the prior art VN processor, chipset, and attached devices.
- the semantic processor receives datagrams from a network port or datagram interface. Some of these datagrams will include client requests for data operations.
- the semantic processor parses elements of the received datagrams using a parser table designed for the server protocol grammar(s) supported by the server. Based on what is parsed, the semantic processor transacts responsive data operations with a data storage device or devices. Generally, the data operations are performed by launching microinstruction code segments on a simple execution unit or units. As the data operations are performed, the semantic processor generates response datagrams to send back to the client.
- the resulting operational efficiency allows an entire storage server to be collapsed in some embodiments into a few relatively small integrated circuits that can be placed on a media device's printed circuit board, with the semantic processor itself drawing perhaps a few Watts of power.
- FIG. 1 contains a block diagram for a typical von Neumann machine storage server
- FIG. 2 contains a block diagram for a generic storage server embodiment of the present invention
- FIG. 3 illustrates, in block form, a more-detailed storage server embodiment of the invention
- FIG. 4 shows the general structure of an Ethernet/IP/TCP/CIFS storage server client request frame
- FIG. 5 shows details of a Storage Message Block portion of the CIFS frame structure of FIG. 4 ;
- FIG. 6 illustrates, in block form, one semantic processor implementation useful with embodiments of the present invention
- FIG. 7 shows a block diagram for a storage server embodiment with several physical data storage devices
- FIG. 8 illustrates a block diagram for an embodiment of the invention that can be implemented on the printed circuit board of a data storage device
- FIG. 9 shows a block diagram for am implementation with a semantic processor that communicates directly with the drive electronics of a disk
- FIG. 10 shows a translator-storage server block diagram according to an embodiment of the invention.
- FIG. 11 illustrates a system embodiment wherein multiple physical data storage devices are accessed by the semantic processor over an external storage interface
- FIG. 13 illustrates yet another semantic processor implementation useful with embodiments of the present invention.
- FIG. 2 illustrates a high-level diagram representative of many embodiments of the present invention.
- a storage server comprises a semantic processor 100 with a datagram interface 90 and a storage interface 110 .
- the datagram interface 90 provides client connectivity to the server, e.g., over network 80 as shown or via a point-to-point connection to the client.
- the storage interface 110 provides a path for the semantic processor to initiate data transactions with data storage device 120 in response to client requests.
- the data storage device may be local.
- the data storage device alternately may be network-connected to the illustrated server, e.g., if the semantic processor 100 emulates a NAS server to clients, while at the same time using remote SAN servers for physical storage.
- FIG. 3 contains a more-detailed block diagram of a storage server 200 , including a semantic processor 100 .
- the datagram interface 90 connects buffer 130 of semantic processor 100 to a physical interface device (PHY) 92 , e.g., an optical, electrical, or radio frequency driver/receiver pair for an Ethernet, Fibre Channel, 802.11x, Universal Serial Bus, Firewire, or other physical layer interface.
- PHY physical interface device
- the datagram interface supplies an input digital data stream to buffer 130 , and receives an output digital data stream from buffer 130 .
- the storage interface 110 is implemented as a block I/O bus between a data storage device 120 and the semantic processor 100 .
- this bus can be a cabled connection (iSCSI, Fibre Channel, SATA) or circuit board bus (SATA), depending on the configured capability of the semantic processor and the data storage device.
- FIG. 3 shows major elements of a hard disk type of media device, e.g., a disk controller 122 , drive electronics 124 , and the disk platter(s), motor, and read/write heads 126 .
- Semantic processor 100 contains a direct execution parser (DXP) 140 that controls the processing of input packets or frames received at buffer 130 (e.g., the input “stream”).
- DXP 140 maintains an internal parser stack of terminal and non-terminal symbols, based on parsing of the current frame up to the current symbol. When the symbol at the top of the parser stack is a terminal symbol, DXP 140 compares data at the head of the input stream to the terminal symbol and expects a match in order to continue.
- DXP direct execution parser
- DXP 140 uses the non-terminal symbol and current input data to expand the grammar production on the stack. As parsing continues, DXP 140 instructs semantic code execution engine (SEE) 150 to process segments of the input, or perform other operations.
- SEE semantic code execution engine
- the semantic processor is reconfigurable by modifying its tables, and thus has the flexibility appeal of a VN machine. Because the semantic processor responds to the input it is given, it can generally operate efficiently with a smaller instruction set than a VN machine.
- the instruction set also benefits because the semantic processor allows data processing in a machine context, as will be explained below.
- Semantic processor 100 uses at least three tables to perform a given function. Codes for retrieving production rules are stored in a parser table (PT) 170 . Grammatical production rules are stored in a production rule table (PRT) 180 . Code segments for SEE 150 are stored in semantic code table (SCT) 190 .
- PT parser table
- PRT production rule table
- SCT semantic code table
- parser table codes are stored, e.g., in a row-column format or a content-addressable format.
- a row-column format the rows of the table are indexed by a non-terminal code on the internal parser stack, and the columns of the table are indexed by an input data value at the head of the input (e.g., the symbol currently on the Si-Bus).
- a content-addressable format a concatenation of the non-terminal code and the input data value can provide the input to the table.
- the production rule table 180 is indexed by the codes in parser table 170 .
- the tables can be linked as shown in FIG. 3 , such that a query to the parser table will directly return the production rule applicable to the non-terminal code and input data value.
- the direct execution parser replaces the non-terminal code at the top of its stack with the production rule returned from the PRT, and continues to parse its input.
- codes for many different grammars can exist at the same time in a production rule table.
- one set of codes can pertain to MAC (Media Access Control) packet header format parsing
- other sets of codes can pertain to Address Resolution Protocol (ARP) packet processing, Internet Protocol (IP) packet processing, Transmission Control Protocol (TCP) packet processing, Real-time Transport Protocol (RTP) packet processing, etc.
- ARP Address Resolution Protocol
- IP Internet Protocol
- TCP Transmission Control Protocol
- RTP Real-time Transport Protocol
- Non-terminal codes need not be assigned in any particular order in production rule code memory 122 , nor in blocks pertaining to a particular protocol.
- the semantic code table 190 can be indexed by parser table codes, and/or from production rule table 180 . Generally, parsing results allow DXP 140 to detect whether, for a given production rule, a code segment from semantic code table 190 should be loaded and executed by SEE 150 .
- the semantic code execution engine 150 has an access path to machine context 160 , which is a structured memory interface, addressable by contextual symbols.
- Machine context 160 , parser table 170 , production rule table 180 , and semantic code table 190 may use on-chip memory, external memory devices such as synchronous DRAMs and CAMs, or a combination of such resources.
- Each table or context may merely provide a contextual interface to a shared physical memory space with one or more of the other tables or contexts.
- FIG. 4 shows pertinent header/data blocks of an Ethernet/IP/TCP/CIFS frame 250 (ignoring trailers and checksums).
- the MAC header will contain, among other things, the MAC address of server 200 for any frame intended for server 200 .
- Server 200 can support several network and transport protocols, but this example uses an Internet Protocol (IP) network header and Transmission Control Protocol (TCP) transport protocol header.
- IP Internet Protocol
- TCP Transmission Control Protocol
- Following the TCP header are a CIFS Storage Message Block (SMB) header HEADER 1 and a SMB data buffer BUFFER 1 containing data related to HEADER 1 .
- SMB Storage Message Block
- Many CIFS opcodes can be combined with other CIFS opcodes in the same SMB if desired, as long as the maximum frame length is not exceeded.
- the additional headers for second, third, etc. opcodes contain only the last few fields of the first header, with all other fields implied from the first header. As shown, the last SMB
- FIG. 5 shows additional detail for the first SMB header and buffer of a frame.
- a full SMB header first indicates its protocol, i.e., a character 0xFF indicates that this is an SMB header.
- a command character follows the protocol character, and indicates an operation code (opcode) for an operation that is either being requested or responded to.
- opcode operation code
- the status and flags field determine how other fields of the SMB should be interpreted, and/or whether an error occurred.
- the MAC signature (in this instance MAC stands for Media Authentication Code), when valid, can be used to authenticate messages.
- TID Telession Identifier
- PID Process Identifier
- UID User Identifier
- MID Multiplex Identifier
- the SMB ends with a byte count and buffer of byte count length.
- a buffer is used, e.g., to send a filename, transfer write data to a server, or transfer read data to a client.
- each new frame received at buffer 130 starts with a MAC header.
- the MAC header contains a destination address, a source address, a payload type field, a payload, and a checksum.
- DXP 140 parses the destination address for a match to the address(es) assigned to storage server 200 , and to a broadcast address. If the destination address fails to parse, the DXP launches a SEE code segment on SEE 150 to flush the frame. Otherwise, the source address is consumed by a SEE and saved for use in reply, and the payload type field is parsed.
- the MAC payload type parses to IP.
- the DXP consequently loads production rules for IP header parsing onto its parser stack, and works through the IP header. This may include, for instance, parsing the destination IP address to ensure that the packet was intended for the storage server, dispatching a SEE to check the IP header checksum, update addressing tables, save the destination and source addresses for return packet and other processing, and/or reassemble IP fragmented packets.
- the protocol indicated in the IP header is parsed to load production rules for the next header (in this case a TCP header) onto the DXP parser stack.
- the TCP header is then parsed.
- a SEE is dispatched to load the TCP connection context associated with the destination IP address, source IP address, destination port, and source port (if a valid connection exists—a condition that is assumed for this example).
- the SEE then consumes the sequence number, acknowledgment number, window, checksum, and other fields necessary to update the connection context.
- the connection context is saved for use in reply.
- the next symbol on the input stream will indicate that the data contains an SMB request.
- DXP 140 parses this symbol and loads CIFS grammar production rules onto its stack.
- the next input symbol will be matched with a non-terminal symbol for a CIFS command (CIFS opcode).
- CIFS opcode a CIFS command
- the parser table can contain an entry for each possible combination of CIFS opcode and this non-terminal symbol.
- the status and flags fields of the CIFS header can be parsed, but preferably are consumed by SEE 150 and saved in machine context 160 for use as necessary in interpreting packet contents.
- the MAC signature, TID, PID, UID, and MID fields are also directed to a SEE, which saves the fields for use in constructing a return frame, and performs lookups in machine context 160 to authenticate the packet as originating from a valid source and directed to a proper resource.
- the Parameters field format varies with CIFS opcode. Depending on the opcode, it may be preferable to parse elements of the Parameters field, or instruct SEE 150 to consume the Parameters using a segment of microcode. Several examples for common CIFS commands will be given.
- a CIFS NEGOTIATE client request is used by a new client to identify the CIFS dialects that the client can understand.
- the Parameters field for NEGOTIATE include a ByteCount, followed by ByteCount symbols that list the acceptable dialects as null-terminated strings.
- the production rule for NEGOTIATE prompts DXP 140 to cause SEE 150 to save the ByteCount, and then DXP 140 parses the first input string up until a NULL character. If the string parses to a known dialect, DXP 140 causes SEE 150 to save a code for that dialect to the current frame context. SEE 150 also determines whether ByteCount symbols have been parsed, indicating that all dialects have been parsed.
- SEE 150 pushes a non-terminal symbol onto the top of the parser stack, causing DXP 140 to parse the remaining input for another dialect. This process continues until ByteCount symbols have been parsed. At that time, SEE 150 pushes a symbol onto the head of the input stream to cause DXP 140 to stop parsing for dialects. DXP 140 then finishes parsing the packet and instructs SEE 150 to select a dialect (e.g., according to a pre-programmed hierarchical preference) and send a response packet back to the client with parameters pertaining to a new session. SEE 150 also sets up a new session context within machine context 160 .
- a dialect e.g., according to a pre-programmed hierarchical preference
- DXP 140 When a client receives a NEGOTIATE reply, normally the client will send an SMB for SESSION_SETUP_ANDX to complete the session. DXP 140 , upon receiving this SMB, can parse the first Parameter (WordCount) and verify that it is correct for the opcode. The second Parameter, AndXCommand, indicates whether (and which) secondary command X also appears in this frame, following this command (CIFS uses “AndX” to identify opcodes that can be concatenated with a following opcode to form a multi-opcode SMB). If AndXCommand indicates that no second opcode is present (0xFF), DXP 140 can parse this and continue. If a separate opcode is present, processing is more complicated.
- the second opcode can be handled in several ways.
- One way is to write a separate grammar variant for each possible combination of first and second opcodes. This is feasible, but potentially inefficient, depending on parser table and production rule table constraints.
- Another way is to use a multi-level grammar, with a higher level grammar parsing opcodes and a lower level grammar further processing each parsed opcode.
- a third method is to use a pushback mechanism from SEE 150 . In this method, for instance, the AndXCommand Parameter loads a production rule that causes SEE 150 to save the AndXCommand from the input stream.
- SEE 150 When the first opcode has been completely parsed, SEE 150 is prompted to run a microcode segment that pushes the AndXCommand Parameter back onto the head of the input stream. DXP 140 then parses the new opcode and continues from that point, loading Parameter field grammar for the new opcode if a second opcode is present. Additional AndX commands in the SMB can be handled in the same manner.
- SESSION_SETUP_ANDX The other Parameters of SESSION_SETUP_ANDX likely will be consumed by SEE 150 without parsing, as most contain parameters to be saved in the session context, or verified, in the case of a password. Null-terminated string parameters can be parsed to locate the null termination symbols, followed by an instruction to SEE 150 to save the symbols whose length is now determined.
- SEE 150 can be commanded to build a response packet. If a second opcode was included, however, the packet will not be finalized until each opcode has been processed.
- a LOGOFF_ANDX command is used to end a session.
- the primary function performed by the semantic processor in response to this opcode is to cause SEE 150 to remove the session context for that session from machine context 160 .
- a TREE_CONNECT_ANDX command is used to connect the session with a shared server resource indicated by a parameter string Path.
- the Path string included with this command could be parsed only for length, and/or parsed for correct server name.
- the remainder of the Path name could be parsed, although since valid paths can be created and destroyed frequently on a writeable resource, keeping the production rule codes correct for each directory may be challenging. Accordingly, the path would typically be passed to SEE 150 for opening. Alternately, DXP 140 could parse for “/” characters and pass the path to SEE 150 one level at a time.
- SEE 150 traverses the specified path by reading directories stored on data storage device 120 across block I/O bus 110 , starting with the root directory, and verifying that the requested path exists and is not restricted from the client. Assuming the path is valid and available, SEE 150 builds a response packet and saves the path in the session context.
- the NT_CREATE_ANDX command creates or opens a file or directory. Like with the tree connect command, DXP 140 may hand off the bulk of this command to SEE 150 for block I/O transactions with data storage device 120 . SEE 150 will open and/or create the file, if possible, by modifying the appropriate directory, assigning a file identifier (FID) to the open file, and creating a file context for the open file in machine context 160 . SEE 150 then formats the appropriate response frame indicating the results of the file create/open request.
- FID file identifier
- the READ_ANDX and WRITE_ANDX commands are used by a client to read and write data from/to an open FID.
- the DXP parses down to the FID parameter, it signals SEE 150 to take FID off of the Si-Bus and locate the corresponding file context in machine context 160 .
- SEE 150 then performs the appropriate block I/O transactions to read or write data on data storage device 120 , and constructs a return packet to the client. It is noted that a write operation return frame could optionally be generated and sent before all block I/O transactions with the data storage device are completed.
- the commands presented above are a subset of possible CIFS commands. Those skilled in the art can appreciate, from the examples above, how a semantic processor can implement full CIFS functionality. Further, the concepts exemplified by the semantic processor performing these CIFS commands are applicable in implementing other storage server protocols on a semantic processor.
- FIG. 6 shows another semantic processor embodiment 300 .
- Semantic processor 300 contains four semantic code execution engines 152 , 154 , 156 , and 158 .
- Semantic code execution engine 158 communicates with a block I/O circuit 112 .
- Machine context 160 contains two functional units: a variable machine context data memory (VMCD) 162 and an array machine context data memory (AMCD) 164 .
- VMCD variable machine context data memory
- AMCD array machine context data memory
- DXP 140 when DXP 140 determines that a SEE task is to be launched at a specific point in its parsing, DXP 140 signals one of the SEEs to load microinstructions from SCT 140 .
- the handle for some instruction segments may indicate to DXP 140 that it may choose any available SEE, or the handle may indicate that a specific SEE (for instance SEE 158 , which has sole access to block I/O 112 ) should receive that code segment.
- SEE 158 which has sole access to block I/O 112
- SEE 152 can be a designated input protocol SEE, responsible for handling the input side of IP, TCP, and other protocols, and updating client, session, and file contexts with data from incoming CIFS frames
- SEE 154 can be designated to perform file system operations, such as comprehending and updating directories, file allocation tables, and user/password lists, authenticating users and requests, etc.
- SEE 156 can be designated to handle the output side of protocols, e.g., building response frames.
- SEE 158 can be designated to handle transactions with the data storage device(s).
- one SEE can launch microcode on another SEE without having to go through DXP 140 , which may have moved on to a different parsing task.
- SEE 154 can launch tasks on block I/O SEE 158 to retrieve or update directories.
- output protocol SEE 156 can have a semaphore that can be set by another SEE when data is ready for a response packet.
- Each SEE contains pipeline registers to allow machine-context access to data.
- the preferred SEE embodiments have no notion of the physical data storage structure used for the data that they operate on. Instead, accesses to data take a machine-context transactional form.
- Variable (e.g., scalar) data is accessed on the V-bus; array data is accessed on the A-bus; and input stream data is accessed on the Si-bus.
- a SEE uses an instruction decoder to prompt the V-bus interface to issue a bus request ⁇ read, ct, offset, m ⁇ .
- the context mct refers to the master context of the semantic processor; other sub-contexts will usually be created and destroyed as the RSP processes input data, such as a sub-context for each active CIFS session, each open file, each open transaction, etc.
- a SEE pipeline register Once a SEE pipeline register has been issued a command, it handles the data transfer process. If multiple bus transfers are required to read or write m octets, the pipeline register tracks the transaction to completion. As an example, a six-octet field can be transferred from the stream input to a machine-context variable using two microinstructions: a first instruction reads six octets from the Si-bus to a pipeline register; a second instruction then writes the six octets from the register to the machine-context variable across the V-bus.
- the register interfaces perform however many bus data cycles are required to effect the transfer.
- VMCD 162 serves the requests initiated on the V-bus.
- VMCD 162 has the capability to translate machine-context variable data requests to physical memory transactions.
- VMCD 162 preferably maintains a translation table referencing machine context identifiers to physical memory starting addresses, contains a mechanism for allocating and deallocating contexts, allows contexts to be locked by a given SEE, and ensures that requested transactions do not fall outside of the requested context's boundaries.
- the actual storage mechanism employed can vary based on application: the memory could be completely internal, completely external, a mix of the two, a cache with a large external memory, etc.
- An external memory can be shared with external memory for other memory sections, such as the AMCD, input and output buffers, parser table, production rule table, and semantic code table, in a given implementation.
- the A-bus interface and AMCD 164 operate similarly, but with an array machine context organization.
- different types of arrays and tables can be allocated, resized, deallocated, written to, read from, searched, and possibly even hashed or sorted using simple bus requests.
- the actual underlying physical memory can differ for different types of arrays and tables, including for example fast onboard RAM, external RAM or ROM, content-addressable memory, etc.
- FIG. 3 The storage server organization shown in FIG. 3 is one of many possible functional partitions of a server according to an embodiment of the present invention. Some other possible configurations are shown in FIGS. 7-10 , as described below.
- FIG. 7 illustrates a storage server 500 with interfaces for one or more conventional SATA drives.
- a network connection 502 (such as a standard electrical or optical connector port or antenna) allows clients to communicate with a PHY 510 supporting a desired protocol, such as Fibre Channel, Ethernet, etc.
- a desired protocol such as Fibre Channel, Ethernet, etc.
- a commercially available PHY such as a Broadcom BCM5421 or a Marvell 88E1011 can be used.
- PHY 510 supplies input frames to, and drives output frames from, an RSP 520 .
- RSP 520 can be configured in one of the configurations described above, one of the configurations described in copending application Ser. No. 10/351,030, or any other functionally similar semantic processing configuration.
- RAM 530 provides physical storage, e.g., for machine context and buffers, to RSP 520 .
- RAM 530 may comprise several types of memory (DRAM, Flash, CAM, etc.), or a single type such as Synchronous DRAM.
- a boot ROM or boot Flash memory can be used to initialize RSP 520 when parser, production rule, and semantic code tables are stored in volatile memory during operation. Part of the non-volatile table storage could also exist on the data storage device(s), as long as enough code is available in boot memory to allow RSP 520 to communicate with the data storage device(s).
- a SATA controller 540 connects to a block I/O port of RSP 520 to serve the disk access requests of RSP 520 .
- SATA controller 540 can be, e.g., a commercially available SATA controller such as a SII 3114 or a Marvell 88SX5080.
- Serial ATA controller 540 connects to one or more SATA data storage devices via SATA bus(es). As shown, storage server 500 supports drives 550 - 0 through 550 -N, respectively, through SATA buses BUS 0 through BUSN.
- PHY 510 , RSP 520 , RAM 530 , and SATA controller 540 are preferably interconnected on a common printed circuit board (not shown).
- the circuit board can be arranged with the drives in a common enclosure, with SATA cabling providing BUS 0 through BUSN. Alternately, SATA bus signaling can be routed on the printed circuit board to connectors for each drive, or through a connector to a backplane.
- FIG. 8 Another storage server implementation 600 is shown in FIG. 8 .
- the entire server is implemented on the printed circuit board of a disk drive, with a network connection 602 for providing a connection to storage server clients.
- storage server architecture 500 could possibly be packaged in the same way, storage server 600 achieves space and cost savings by integrating a PHY 610 and an SATA controller 640 within an RSP 620 .
- SATA controller 640 preferably is at least partially implemented using a SEE.
- the SATA controller section of RSP 620 interfaces through an SATA cable or circuit board bus to drive electronics 660 , which include an SATA interface, disk cache, and drive control electronics.
- disk controller functionality is implemented within an RSP 740 .
- the control devices such as a Marvell 88C7500, motors, and servos shown as block 770 interface directly with RSP 740 .
- FIG. 10 illustrates a storage server 800 that is essentially a “translation” gateway.
- Storage server 800 comprises a first network physical interface PHY 1 (block 810 ) to connect to a network 802 , and consequently to storage server clients.
- Storage server interface 800 also contains a second network physical interface PHY 2 (block 840 ) to connect, across a network or point-to-point connection, to one or more physical servers, such as the illustrated SAN server 850 .
- An RSP 820 with attached memory 830 , attaches to PHY 1 and PHY 2 .
- PHY 1 can be, e.g., an Ethernet PHY
- PHY 2 can be, e.g., a Fibre Channel PHY.
- RSP 820 responds to client requests, such as NAS-style or application-style requests, by initiating requests on the remotely-attached server 850 .
- client requests such as NAS-style or application-style requests
- server 800 appears as a server to clients on network 802 , and as a client to server 850 .
- server 850 is illustrated as a SAN server, server 850 could be a NAS or even an application-style server. If PHY 1 and PHY 2 support the same server protocol, storage server 800 can still serve a useful function as an aggregation point for a scalable server farm, an encryption point, and/or firewall.
- RSP 820 can provide parsing for both input streams. For instance, both PHY 1 and PHY 2 input streams can be saved to a common input buffer, and a DXP (not shown) in RSP 820 can alternately parse datagrams in both input streams, thus coordinating two-way translation tasks.
- RSP 820 could also be configured with two DXPs, one to serve each physical port, sharing a common bank of SEEs and other RSP resources.
- a storage server 900 contains a storage interface connecting RSP 100 to multiple physical data storage devices, e.g., four devices 120 - 1 , 120 - 2 , 120 - 3 , and 120 - 4 as shown. These data storage devices can be accessed in some embodiments as a RAID (Redundant Array of Independent Disks) array, with either the RSP acting as a RAID controller, or with a separate RAID controller (not shown) implemented as part of storage interface 110 .
- the four data storage devices can alternately be configured as a JBOD (Just a Bunch Of Disks) array.
- RSP 100 can perform other forms of storage server virtualization for its clients.
- disk sector addresses used by a client application are mapped by the RSP into C-H-S (Cylinder-Head-Sector) addresses of a specific device from one of many physical devices.
- data storage devices 120 - 1 through 120 - 4 (and potentially many more such devices) need not be geographically co-located, allowing sophisticated control of physical resources and disk allocation for clients.
- the RSP can also function within a virtualization configuration where a separate intermediate device provides all or part of the virtualization functionality for clients.
- FIG. 12 illustrates the use of an RSP 100 in a storage server system 1000 that incorporates a port extender 950 .
- Port extender 950 connects to a single SATA controller port associated with RSP 100 (either internal or external), but communicates in turn with multiple physical data storage devices (e.g., devices 120 - 1 to 120 - 4 as shown).
- Such a configuration can increase overall system bandwidth while using only one port on the RSP for communication with data storage devices.
- FIG. 13 shows yet another semantic processor implementation 1100 useful with embodiments of the present invention.
- Semantic processor 1100 communicates with a Port 0 PHY 93 , a Port 1 PHY 94 , and a PCI-X interface 95 , each of which can be integrated with processor 1100 or connected externally to processor 1100 , as desired.
- Buffer 130 of FIG. 3 is replaced with a Port Input Buffer (PIB) 132 and a Port Output Buffer (POB) 134 , which preferably, but not necessarily, reside in a common memory section integrated as part of processor 1100 .
- PIB 132 connects to and maintains at least one input queue for each of Port 0 , Port 1 , and the PCI-X interface, and may contains other queues as well.
- POB 134 connects to and maintains at least one output queue for each of Port 0 , Port 1 , and the PCI-X interface, and may contain other queues as well.
- Port 0 and Port 1 typically represent datagram interfaces such as Gigabit Ethernet, while the PCI-X interface couples to a familiar PXI-X bus.
- storage resources can couple to any of these ports, and clients can couple to either Port 1 , Port 2 , or both.
- direct execution parser 140 can receive input from each of the multiple input queues in PIB 132 , through a parser source selector 136 .
- direct execution parser 140 can maintain separate parser stacks for each input source, and can signal parser source selector to switch input sources each time a packet has finished parsing, or when parsing of one packet is stalled, e.g., while a SEE performs some calculation on a packet field.
- SEE cluster 152 comprises N semantic code execution engines 150 - 1 to 150 -N.
- the SEE cluster preferably includes a single dispatcher (not shown) to communicate with DXP 140 over the Sx-Bus, and to distribute tasks to the individual SEEs.
- a PIB read controller 133 and a POB write controller block 135 provide access, respectively, to PIB 132 and POB 134 for SEE cluster 152 .
- parser source selector 136 and PIB read controller 133 allow DXP 140 to work with data from one input while the SEE cluster accesses data from another input.
- PIB read controller 133 and POB write controller 135 preferably allow non-blocking access to the input and output buffers.
- a machine context port controller 154 provides access to machine context 160 for SEE cluster 152 . Like the port buffer controllers, machine context port controller 154 provides non-blocking access to machine context 160 for SEE cluster 152 .
- Machine context 160 prioritizes and executes memory tasks for SEE cluster 152 and for a management microprocessor 195 , and typically contains one or more specialized caches that each depend on the attributes of the target data access.
- Machine context 160 can also contains one or more encryption and/or authentication engines that can be used to perform inline encryption/authentication.
- One or more traditional memory bus interfaces connect machine context 160 with physical memory 165 , which can consist, e.g., of DRAM (Dynamic Random Access Memory), CAM (Content Addressable Memory), and/or any other type of desired storage.
- Physical memory 165 can be located on processor 1100 , external to processor 1100 , or split between these two locations.
- Management microprocessor 195 performs any desired functions for semantic processor 1100 that can reasonably be accomplished with traditional software.
- microprocessor 195 can interface with its own instruction space and data space in physical memory 165 , through machine context 160 , and execute traditional software to boot the processor, load or modify the parser table, production rule table, and semantic code table, gather statistics, perform logging, manage client access, perform error recovery, etc.
- microprocessor 195 also has the capability to communicate with the dispatcher in SEE cluster 152 in order to request that a SEE perform tasks on the microprocessor's behalf.
- the management microprocessor is preferably integrated into semantic processor 1100 .
- the semantic units can encrypt the blocks of data as the blocks are written to disk and/or decrypt the blocks of data as the blocks are read from disk. This provides security to the data “at rest” on the disk drive or drives.
- the semantic units receive data payloads for writing to disk, one operation in preparing the data payloads for writing can include encrypting the packets. The reverse of this process can be used for reading encrypted data from the disk.
- parser as a front-end datagram processor for a general purpose processor.
- SAN-style server e.g., connected by Fibre Channel to an enterprise network
- NAS-style servers include more sophisticated features, such as those described in the detailed CIFS embodiment described above.
- Translation-gateway servers provide clients with access to a potentially large and changing underlying array of physical servers without allowing those clients visibility to the physical servers.
- Both translation-gateway servers and physical servers can be structured as application style-servers.
- As an application-style server data is served in an application-style format.
- a video-on-demand server can store videos for usage by multiple clients, and stream different parts of a video to different clients, allowing each to have independent navigation.
- a music-on-demand server can operate similarly.
- Application-style servers can also supply storage space for applications, like a wired or wireless server that stores digital camera pictures off of a digital camera, allowing the camera to operate with a relatively small internal buffer and/or flash card.
- a server could be relatively small, battery-powered, and portable for field use.
- Semantic processor storage server implementations can also implement wireless protocols, e.g., for use in a home computer data and/or audio/video network.
- wireless protocols e.g., for use in a home computer data and/or audio/video network.
- the detailed embodiments have focused on traditional “storage” devices, printers, scanners, multi-function printers, and other data translation devices can also be attached to a semantic processor storage server.
- semantic processor storage server can be made to serve different client types by changing the protocols that can be parsed by the server. If desirable, the storage server could even parse multiple storage server protocols simultaneously to allow access by different classes of clients (e.g., SAN access by trusted clients and NAS access by others).
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Library & Information Science (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
- This application is a continuation of U.S. patent application Ser. No. 10/843,727, filed on May 11, 2004, now pending, the disclosure of which is herein incorporated by reference
- This invention relates generally to storage servers, and more specifically to storage server implementations using digital semantic processors.
- Traditionally, a storage server is a networked computer that provides media and file access to other computers and computing devices, referred to herein as “clients.” The storage server may be accessible on a Wide-Area Network (WAN), a Local-Area Network (LAN), and/or share a dedicated point-to-point connection with another computing device.
- Storage servers can be of several types, including a Network Attached Storage (NAS) server, a Storage Area Network (SAN) server, and an application-style server.
- A NAS server comprehends directory structures. Requests to a NAS server generally specify a file path/filename and action such as create, read, write, append, etc. A NAS server converts the requests to one or more disk sector/block disk access transactions, accesses the disk (or other storage media), and performs the requested file-level transaction. Such a server maintains and utilizes a directory structure to locate the file starting locations, and a File Allocation Table to reconstruct segmented files and locate free sectors on a disk or other writeable media.
- A SAN server generally receives direct requests for disk sector/block-based transactions, and may have no knowledge of a file system or relationships between data blocks. Although block requests can relate to physical block locations on a media device, more typically a SAN server supports logical blocks—thus allowing the server to make one physical disk drive appear as several logical drives, or several physical disk drives appear as one logical drive, as in a RAID (Redundant Array of Independent Disks) scheme.
- An application-style server generally provides access to media in a media-specific format. For instance, an application-style server could provide access to streaming audio or video without the explicit knowledge of a filepath/filename by the client (who may not know the filename or have the authority to access a file directly), while reading the streaming media from files and encapsulating it in a streaming packet format.
- Architecturally, most storage servers differ little from general-purpose computers (other than in some cases including more and/or faster than usual disk drives and disk bus controllers).
FIG. 1 shows a block diagram for atypical storage server 20. A Central Processing Unit (CPU) 22, typically one or more microprocessors, operates the server according to stored program instructions. This type of processor is often called a von Neumann (VN) processor or machine, after its innovator, who proposed its execution style of processing sequential instructions. -
CPU 22 connects through a Frontside Bus (FSB) 26 to a Memory Controller/Hub (MCH) 24, which is responsible for interfacingCPU 22 to other system components that store microprocessor program instructions and store, supply, and/or consume data. MCH 24 managessystem memory 30 throughmemory bus 32. MCH 24 also communicates with PCI (Peripheral Component Interconnect)bridge 40 acrosshub bus 42 to move data between PCI-connected devices and eithermemory 30 orCPU 22. - Many different types of data source/sink devices can connect to
server 20 throughPCI bridge 40 andPCI bus 44. For storage server purposes, the two necessary devices are a network interface card (NIC) 50 and a media controller, such as ATA (AT Attachment)controller 60. - NIC 50 allows the server to attach directly to a Local Area Network (LAN) or other network, a Fibre Channel switch fabric, or other point-to-point connection to a client or another device that provides network visibility for the server. NIC 50 thereby provides a communication path for the server to receive data requests from clients and respond to those requests. The actual network physical communication protocol can be varied by inserting a different NIC, e.g., for whatever wired or wireless communication protocol is to be used, and loading a software device driver for that NIC onto
CPU 22. - ATA
controller 60 provides at least one ATAbus 62 for attachingmedia devices 64 such as hard disks, optical disks, and/or tape drives. Each ATAbus 62 allows either one or twomedia devices 64 to attach tostorage server 20. Some servers may employ other controllers, such as a SATA (serial ATA) controller or a SCSI (Small Computer System Interface) host adapter. - In order to operate as a storage server,
CPU 22 must execute a variety of software programs. With NAS, the two common formats are NFS (Networked File System) and CIFS (Common Internet File System); the former protocol is found primarily in UNIX environments, and the latter protocol in Microsoft operating system environments. For an NFS server, the following software processes help implement NFS: a network device driver for NIC 50; TCP/IP drivers; RPC (Remote Procedure Call) and XDR (External Data Representation) drivers to present data from TCP/IP to NFS server software; the NFS server software itself; a local file system; a VFS (Virtual File System) driver to interface the NFS server software with the local file system; data buffering software; and device drivers for the storage controller/host adapter. For each NFS data transaction.CPU 22 must execute a process for each of these software entities, switching context between them as required. To provide adequate performance,storage server 20 operates at relatively high speed and power, requiring forced air cooling and a substantial power supply to operate the microprocessor, chipset, network interface, memory, other peripherals, and cooling fans. The current state of the art, with a 1 Gbps (Gigabit per second) full duplex network connection, requires about 300 watts of power, occupies a volume of 800 cubic inches, and costs about $1500 without storage media. - Some vendors have attempted to build custom integrated circuit hardware explicitly for processing NAS requests. Design costs for such circuits generally cannot be recaptured in the fluid world of telecommunication protocols, where a specific complex circuit built to handle a narrow combination of specific protocols and storage interfaces may become rapidly obsolete.
- It has now been recognized that in many storage server applications a microprocessor/sequential program approach is inefficient and bulky. Most of the storage server's existence is spent executing storage server functions in response to client requests received at the NIC. The requests themselves are styled in a protocol-specific format with a limited number of requested functions and options. In response to these requests, well-defined storage system commands are supplied to a media device, and the results of those commands are used to return protocol-specific datagrams to the client.
- Storage server programmability and code commonality are highly desirable traits. Accordingly, high-volume processors and chipsets (such as those built by Intel Corporation and Advanced Micro Devices) have been used to allow programmability and the use of standard operating systems, and thereby reduce costs. Unfortunately, the flexibility and power of a general-purpose microprocessor-based system go largely unused, or used inefficiently, in such an operational configuration.
- The present invention provides a different architecture for a storage server, using in exemplary embodiments what is referred to generally as a semantic processor. Such a device is configurable, and preferably reconfigurable like a VN machine, as its processing depends on its “programming.” As used herein, a “semantic processor” contains at least two components: a direct execution parser to parse an input stream according to a defined grammar; and an execution engine capable of manipulating data (e.g., data movement, mathematical, and logical operations) in response to requests from the direct execution parser.
- The programming of a semantic processor differs from the conventional machine code used by a VN machine. In a VN machine the protocol input parameters of a data packet are sequentially compared to all possible inputs using sequential machine instructions, with branching, looping, and thread-switching inefficiencies. In contrast, the semantic processor responds directly to the semantics of an input stream. In other words, the “code” segments that the semantic processor executes are driven directly by the input data. For instance, of the 75 or so possible CIFS commands that could appear in a packet, a single parsing cycle of an embodiment described below is all that is necessary to allow the semantic processor to load grammar and microinstructions pertinent to the actual CIFS command sent in the packet.
- In a storage server application as described in the embodiments below, a semantic processor can perform many of the functions of the prior art VN processor, chipset, and attached devices. The semantic processor receives datagrams from a network port or datagram interface. Some of these datagrams will include client requests for data operations. The semantic processor parses elements of the received datagrams using a parser table designed for the server protocol grammar(s) supported by the server. Based on what is parsed, the semantic processor transacts responsive data operations with a data storage device or devices. Generally, the data operations are performed by launching microinstruction code segments on a simple execution unit or units. As the data operations are performed, the semantic processor generates response datagrams to send back to the client. The resulting operational efficiency allows an entire storage server to be collapsed in some embodiments into a few relatively small integrated circuits that can be placed on a media device's printed circuit board, with the semantic processor itself drawing perhaps a few Watts of power.
- The invention may be best understood by reading the disclosure with reference to the drawing, wherein:
-
FIG. 1 contains a block diagram for a typical von Neumann machine storage server; -
FIG. 2 contains a block diagram for a generic storage server embodiment of the present invention; -
FIG. 3 illustrates, in block form, a more-detailed storage server embodiment of the invention; -
FIG. 4 shows the general structure of an Ethernet/IP/TCP/CIFS storage server client request frame; -
FIG. 5 shows details of a Storage Message Block portion of the CIFS frame structure ofFIG. 4 ; -
FIG. 6 illustrates, in block form, one semantic processor implementation useful with embodiments of the present invention; -
FIG. 7 shows a block diagram for a storage server embodiment with several physical data storage devices; -
FIG. 8 illustrates a block diagram for an embodiment of the invention that can be implemented on the printed circuit board of a data storage device; -
FIG. 9 shows a block diagram for am implementation with a semantic processor that communicates directly with the drive electronics of a disk; -
FIG. 10 shows a translator-storage server block diagram according to an embodiment of the invention; -
FIG. 11 illustrates a system embodiment wherein multiple physical data storage devices are accessed by the semantic processor over an external storage interface; -
FIG. 12 illustrates a system embodiment wherein multiple physical data storage devices are coupled to the semantic processor through a port extender; and -
FIG. 13 illustrates yet another semantic processor implementation useful with embodiments of the present invention. -
FIG. 2 illustrates a high-level diagram representative of many embodiments of the present invention. A storage server comprises asemantic processor 100 with adatagram interface 90 and astorage interface 110. Thedatagram interface 90 provides client connectivity to the server, e.g., overnetwork 80 as shown or via a point-to-point connection to the client. Thestorage interface 110 provides a path for the semantic processor to initiate data transactions withdata storage device 120 in response to client requests. The data storage device may be local. The data storage device alternately may be network-connected to the illustrated server, e.g., if thesemantic processor 100 emulates a NAS server to clients, while at the same time using remote SAN servers for physical storage. -
FIG. 3 contains a more-detailed block diagram of astorage server 200, including asemantic processor 100. Thedatagram interface 90 connectsbuffer 130 ofsemantic processor 100 to a physical interface device (PHY) 92, e.g., an optical, electrical, or radio frequency driver/receiver pair for an Ethernet, Fibre Channel, 802.11x, Universal Serial Bus, Firewire, or other physical layer interface. The datagram interface supplies an input digital data stream to buffer 130, and receives an output digital data stream frombuffer 130. - In
FIG. 3 , thestorage interface 110 is implemented as a block I/O bus between adata storage device 120 and thesemantic processor 100. In various embodiments, this bus can be a cabled connection (iSCSI, Fibre Channel, SATA) or circuit board bus (SATA), depending on the configured capability of the semantic processor and the data storage device.FIG. 3 shows major elements of a hard disk type of media device, e.g., adisk controller 122, driveelectronics 124, and the disk platter(s), motor, and read/write heads 126. - One configuration for a
semantic processor 100 is illustrated in the storage server ofFIG. 3 .Semantic processor 100 contains a direct execution parser (DXP) 140 that controls the processing of input packets or frames received at buffer 130 (e.g., the input “stream”).DXP 140 maintains an internal parser stack of terminal and non-terminal symbols, based on parsing of the current frame up to the current symbol. When the symbol at the top of the parser stack is a terminal symbol,DXP 140 compares data at the head of the input stream to the terminal symbol and expects a match in order to continue. When the symbol at the top of the parser stack is a non-terminal symbol,DXP 140 uses the non-terminal symbol and current input data to expand the grammar production on the stack. As parsing continues,DXP 140 instructs semantic code execution engine (SEE) 150 to process segments of the input, or perform other operations. - This structure, with a sophisticated grammar parser that assigns tasks to an execution engine, as the data requires, is both flexible and powerful for highly structured input such as datagram protocols. In preferred embodiments, the semantic processor is reconfigurable by modifying its tables, and thus has the flexibility appeal of a VN machine. Because the semantic processor responds to the input it is given, it can generally operate efficiently with a smaller instruction set than a VN machine. The instruction set also benefits because the semantic processor allows data processing in a machine context, as will be explained below.
-
Semantic processor 100 uses at least three tables to perform a given function. Codes for retrieving production rules are stored in a parser table (PT) 170. Grammatical production rules are stored in a production rule table (PRT) 180. Code segments forSEE 150 are stored in semantic code table (SCT) 190. - The codes in parser table 170 point to production rules in table 180. Parser table codes are stored, e.g., in a row-column format or a content-addressable format. In a row-column format, the rows of the table are indexed by a non-terminal code on the internal parser stack, and the columns of the table are indexed by an input data value at the head of the input (e.g., the symbol currently on the Si-Bus). In a content-addressable format, a concatenation of the non-terminal code and the input data value can provide the input to the table.
- The production rule table 180 is indexed by the codes in parser table 170. The tables can be linked as shown in
FIG. 3 , such that a query to the parser table will directly return the production rule applicable to the non-terminal code and input data value. The direct execution parser replaces the non-terminal code at the top of its stack with the production rule returned from the PRT, and continues to parse its input. - Practically, codes for many different grammars can exist at the same time in a production rule table. For instance, one set of codes can pertain to MAC (Media Access Control) packet header format parsing, and other sets of codes can pertain to Address Resolution Protocol (ARP) packet processing, Internet Protocol (IP) packet processing, Transmission Control Protocol (TCP) packet processing, Real-time Transport Protocol (RTP) packet processing, etc. Grammars for CIFS, NFS, and/or other storage server protocols are added to the production rule code memory in order to add storage server capability. Non-terminal codes need not be assigned in any particular order in production
rule code memory 122, nor in blocks pertaining to a particular protocol. - The semantic code table 190 can be indexed by parser table codes, and/or from production rule table 180. Generally, parsing results allow
DXP 140 to detect whether, for a given production rule, a code segment from semantic code table 190 should be loaded and executed bySEE 150. - The semantic
code execution engine 150 has an access path tomachine context 160, which is a structured memory interface, addressable by contextual symbols.Machine context 160, parser table 170, production rule table 180, and semantic code table 190 may use on-chip memory, external memory devices such as synchronous DRAMs and CAMs, or a combination of such resources. Each table or context may merely provide a contextual interface to a shared physical memory space with one or more of the other tables or contexts. - Detailed design optimizations for the functional blocks of
semantic processor 100 are not within the scope of the present invention. For some examples of the detailed architecture of applicable semantic processor functional blocks, the reader is referred to copending application Ser. No. 10/351,030, which has been incorporated herein by reference. - The function of
semantic processor 100 in a storage server context can be better understood with a specific example. In this example, CIFS commands and data structures are used with an Ethernet/IP/TCP supporting structure. Those skilled in the art will recognize that the concepts illustrated readily apply to other server communication protocols as well. -
FIG. 4 shows pertinent header/data blocks of an Ethernet/IP/TCP/CIFS frame 250 (ignoring trailers and checksums). The MAC header will contain, among other things, the MAC address ofserver 200 for any frame intended forserver 200.Server 200 can support several network and transport protocols, but this example uses an Internet Protocol (IP) network header and Transmission Control Protocol (TCP) transport protocol header. Following the TCP header are a CIFS Storage Message Block (SMB) header HEADER1 and a SMB data buffer BUFFER1 containing data related to HEADER1, Many CIFS opcodes can be combined with other CIFS opcodes in the same SMB if desired, as long as the maximum frame length is not exceeded. The additional headers for second, third, etc. opcodes contain only the last few fields of the first header, with all other fields implied from the first header. As shown, the last SMB header contains HEADERN and BUFFERN. -
FIG. 5 shows additional detail for the first SMB header and buffer of a frame. A full SMB header first indicates its protocol, i.e., a character 0xFF indicates that this is an SMB header. A command character follows the protocol character, and indicates an operation code (opcode) for an operation that is either being requested or responded to. The status and flags field determine how other fields of the SMB should be interpreted, and/or whether an error occurred. The MAC signature (in this instance MAC stands for Media Authentication Code), when valid, can be used to authenticate messages. - The next four fields in the SMB are the TID, PID, UID, and MID fields. A TID (Tree Identifier) is assigned by a server to a client when the client successfully connects to a server resource. The client uses the assigned TID to refer to that resource in later requests. A PID (Process Identifier) is assigned by a client—although the server notes it and uses it to reply to the client, its use is primarily up to the client. A UID (User Identifer) is assigned by a server to identify the user that authenticated the connection. A MID (Multiplex Identifier) allows a client to use the same PID for multiple outstanding transactions, by using a different MID for each transaction.
- The Parameters field is of variable length and structure, with its first character generally indicating the length of following parameters, followed by the parameter words appropriate for the SMB opcode. If the opcode is one that indicates a second SMB header and buffer follows, some of the parameter words indicate the follow-on opcode and offset to the follow-on shortened header.
- The SMB ends with a byte count and buffer of byte count length. Such a buffer is used, e.g., to send a filename, transfer write data to a server, or transfer read data to a client.
- With this background, an explanation of basic CIFS functionality for semantic processor 100 (
FIG. 3 ) can proceed. - Referring to
FIGS. 3 and 4 , each new frame received atbuffer 130 starts with a MAC header. The MAC header contains a destination address, a source address, a payload type field, a payload, and a checksum.DXP 140 parses the destination address for a match to the address(es) assigned tostorage server 200, and to a broadcast address. If the destination address fails to parse, the DXP launches a SEE code segment onSEE 150 to flush the frame. Otherwise, the source address is consumed by a SEE and saved for use in reply, and the payload type field is parsed. - In this example, the MAC payload type parses to IP. The DXP consequently loads production rules for IP header parsing onto its parser stack, and works through the IP header. This may include, for instance, parsing the destination IP address to ensure that the packet was intended for the storage server, dispatching a SEE to check the IP header checksum, update addressing tables, save the destination and source addresses for return packet and other processing, and/or reassemble IP fragmented packets. The protocol indicated in the IP header is parsed to load production rules for the next header (in this case a TCP header) onto the DXP parser stack.
- The TCP header is then parsed. A SEE is dispatched to load the TCP connection context associated with the destination IP address, source IP address, destination port, and source port (if a valid connection exists—a condition that is assumed for this example). The SEE then consumes the sequence number, acknowledgment number, window, checksum, and other fields necessary to update the connection context. The connection context is saved for use in reply.
- Assuming that the MAC, IP, and TCP headers have parsed to a valid connection to the storage server, the next symbol on the input stream will indicate that the data contains an SMB request.
DXP 140 parses this symbol and loads CIFS grammar production rules onto its stack. - The next input symbol will be matched with a non-terminal symbol for a CIFS command (CIFS opcode). For instance, the parser table can contain an entry for each possible combination of CIFS opcode and this non-terminal symbol. When the CIFS command is parsed, grammar pertinent to the CIFS opcode in the command field is loaded on to the parser stack.
- The status and flags fields of the CIFS header can be parsed, but preferably are consumed by
SEE 150 and saved inmachine context 160 for use as necessary in interpreting packet contents. - The MAC signature, TID, PID, UID, and MID fields are also directed to a SEE, which saves the fields for use in constructing a return frame, and performs lookups in
machine context 160 to authenticate the packet as originating from a valid source and directed to a proper resource. - The Parameters field format varies with CIFS opcode. Depending on the opcode, it may be preferable to parse elements of the Parameters field, or instruct
SEE 150 to consume the Parameters using a segment of microcode. Several examples for common CIFS commands will be given. - A CIFS NEGOTIATE client request is used by a new client to identify the CIFS dialects that the client can understand. The Parameters field for NEGOTIATE include a ByteCount, followed by ByteCount symbols that list the acceptable dialects as null-terminated strings. The production rule for NEGOTIATE prompts DXP 140 to cause
SEE 150 to save the ByteCount, and thenDXP 140 parses the first input string up until a NULL character. If the string parses to a known dialect,DXP 140 causes SEE 150 to save a code for that dialect to the current frame context.SEE 150 also determines whether ByteCount symbols have been parsed, indicating that all dialects have been parsed. If not,SEE 150 pushes a non-terminal symbol onto the top of the parser stack, causingDXP 140 to parse the remaining input for another dialect. This process continues until ByteCount symbols have been parsed. At that time,SEE 150 pushes a symbol onto the head of the input stream to causeDXP 140 to stop parsing for dialects.DXP 140 then finishes parsing the packet and instructsSEE 150 to select a dialect (e.g., according to a pre-programmed hierarchical preference) and send a response packet back to the client with parameters pertaining to a new session.SEE 150 also sets up a new session context withinmachine context 160. - When a client receives a NEGOTIATE reply, normally the client will send an SMB for SESSION_SETUP_ANDX to complete the session.
DXP 140, upon receiving this SMB, can parse the first Parameter (WordCount) and verify that it is correct for the opcode. The second Parameter, AndXCommand, indicates whether (and which) secondary command X also appears in this frame, following this command (CIFS uses “AndX” to identify opcodes that can be concatenated with a following opcode to form a multi-opcode SMB). If AndXCommand indicates that no second opcode is present (0xFF),DXP 140 can parse this and continue. If a separate opcode is present, processing is more complicated. - The second opcode can be handled in several ways. One way is to write a separate grammar variant for each possible combination of first and second opcodes. This is feasible, but potentially inefficient, depending on parser table and production rule table constraints. Another way is to use a multi-level grammar, with a higher level grammar parsing opcodes and a lower level grammar further processing each parsed opcode. A third method is to use a pushback mechanism from
SEE 150. In this method, for instance, the AndXCommand Parameter loads a production rule that causesSEE 150 to save the AndXCommand from the input stream. When the first opcode has been completely parsed,SEE 150 is prompted to run a microcode segment that pushes the AndXCommand Parameter back onto the head of the input stream.DXP 140 then parses the new opcode and continues from that point, loading Parameter field grammar for the new opcode if a second opcode is present. Additional AndX commands in the SMB can be handled in the same manner. - The other Parameters of SESSION_SETUP_ANDX likely will be consumed by
SEE 150 without parsing, as most contain parameters to be saved in the session context, or verified, in the case of a password. Null-terminated string parameters can be parsed to locate the null termination symbols, followed by an instruction to SEE 150 to save the symbols whose length is now determined. - Once the SESSION_SETUP_ANDX command has been parsed, SEE 150 can be commanded to build a response packet. If a second opcode was included, however, the packet will not be finalized until each opcode has been processed.
- A LOGOFF_ANDX command is used to end a session. The primary function performed by the semantic processor in response to this opcode is to cause
SEE 150 to remove the session context for that session frommachine context 160. - A TREE_CONNECT_ANDX command is used to connect the session with a shared server resource indicated by a parameter string Path. The Path string included with this command could be parsed only for length, and/or parsed for correct server name. The remainder of the Path name could be parsed, although since valid paths can be created and destroyed frequently on a writeable resource, keeping the production rule codes correct for each directory may be challenging. Accordingly, the path would typically be passed to SEE 150 for opening. Alternately,
DXP 140 could parse for “/” characters and pass the path to SEE 150 one level at a time. -
SEE 150 traverses the specified path by reading directories stored ondata storage device 120 across block I/O bus 110, starting with the root directory, and verifying that the requested path exists and is not restricted from the client. Assuming the path is valid and available,SEE 150 builds a response packet and saves the path in the session context. - The NT_CREATE_ANDX command creates or opens a file or directory. Like with the tree connect command,
DXP 140 may hand off the bulk of this command to SEE 150 for block I/O transactions withdata storage device 120.SEE 150 will open and/or create the file, if possible, by modifying the appropriate directory, assigning a file identifier (FID) to the open file, and creating a file context for the open file inmachine context 160.SEE 150 then formats the appropriate response frame indicating the results of the file create/open request. - The READ_ANDX and WRITE_ANDX commands are used by a client to read and write data from/to an open FID. When the DXP parses down to the FID parameter, it signals
SEE 150 to take FID off of the Si-Bus and locate the corresponding file context inmachine context 160.SEE 150 then performs the appropriate block I/O transactions to read or write data ondata storage device 120, and constructs a return packet to the client. It is noted that a write operation return frame could optionally be generated and sent before all block I/O transactions with the data storage device are completed. - The commands presented above are a subset of possible CIFS commands. Those skilled in the art can appreciate, from the examples above, how a semantic processor can implement full CIFS functionality. Further, the concepts exemplified by the semantic processor performing these CIFS commands are applicable in implementing other storage server protocols on a semantic processor.
-
FIG. 6 shows anothersemantic processor embodiment 300.Semantic processor 300 contains four semanticcode execution engines code execution engine 158 communicates with a block I/O circuit 112.Machine context 160 contains two functional units: a variable machine context data memory (VMCD) 162 and an array machine context data memory (AMCD) 164. Each SEE can transact withVMCD 162 across V-Bus and withAMCD 164 across A-Bus. - In
semantic processor 300, whenDXP 140 determines that a SEE task is to be launched at a specific point in its parsing,DXP 140 signals one of the SEEs to load microinstructions fromSCT 140. The handle for some instruction segments may indicate toDXP 140 that it may choose any available SEE, or the handle may indicate that a specific SEE (forinstance SEE 158, which has sole access to block I/O 112) should receive that code segment. The availability of multiple SEEs allows many tasks to proceed in parallel, without some slow tasks (like block I/O) blocking all processing. - Although not strictly necessary, specific kinds of tasks also can be assigned to specific SEEs. For instance, SEE 152 can be a designated input protocol SEE, responsible for handling the input side of IP, TCP, and other protocols, and updating client, session, and file contexts with data from incoming CIFS frames, SEE 154 can be designated to perform file system operations, such as comprehending and updating directories, file allocation tables, and user/password lists, authenticating users and requests, etc.
SEE 156 can be designated to handle the output side of protocols, e.g., building response frames. And SEE 158 can be designated to handle transactions with the data storage device(s). With such partitioning, one SEE can launch microcode on another SEE without having to go throughDXP 140, which may have moved on to a different parsing task. For instance, SEE 154 can launch tasks on block I/O SEE 158 to retrieve or update directories. As another, example,output protocol SEE 156 can have a semaphore that can be set by another SEE when data is ready for a response packet. - Each SEE contains pipeline registers to allow machine-context access to data. As opposed to a standard CPU, the preferred SEE embodiments have no notion of the physical data storage structure used for the data that they operate on. Instead, accesses to data take a machine-context transactional form. Variable (e.g., scalar) data is accessed on the V-bus; array data is accessed on the A-bus; and input stream data is accessed on the Si-bus. For instance, to read a scalar data element of length m octets located at a given location offset within a data context ct, a SEE uses an instruction decoder to prompt the V-bus interface to issue a bus request {read, ct, offset, m}. The context mct refers to the master context of the semantic processor; other sub-contexts will usually be created and destroyed as the RSP processes input data, such as a sub-context for each active CIFS session, each open file, each open transaction, etc.
- Once a SEE pipeline register has been issued a command, it handles the data transfer process. If multiple bus transfers are required to read or write m octets, the pipeline register tracks the transaction to completion. As an example, a six-octet field can be transferred from the stream input to a machine-context variable using two microinstructions: a first instruction reads six octets from the Si-bus to a pipeline register; a second instruction then writes the six octets from the register to the machine-context variable across the V-bus. The register interfaces perform however many bus data cycles are required to effect the transfer.
-
VMCD 162 serves the requests initiated on the V-bus.VMCD 162 has the capability to translate machine-context variable data requests to physical memory transactions. Thus VMCD 162 preferably maintains a translation table referencing machine context identifiers to physical memory starting addresses, contains a mechanism for allocating and deallocating contexts, allows contexts to be locked by a given SEE, and ensures that requested transactions do not fall outside of the requested context's boundaries. The actual storage mechanism employed can vary based on application: the memory could be completely internal, completely external, a mix of the two, a cache with a large external memory, etc. An external memory can be shared with external memory for other memory sections, such as the AMCD, input and output buffers, parser table, production rule table, and semantic code table, in a given implementation. - The A-bus interface and
AMCD 164 operate similarly, but with an array machine context organization. Preferably, different types of arrays and tables can be allocated, resized, deallocated, written to, read from, searched, and possibly even hashed or sorted using simple bus requests. The actual underlying physical memory can differ for different types of arrays and tables, including for example fast onboard RAM, external RAM or ROM, content-addressable memory, etc. - The storage server organization shown in
FIG. 3 is one of many possible functional partitions of a server according to an embodiment of the present invention. Some other possible configurations are shown inFIGS. 7-10 , as described below. -
FIG. 7 illustrates astorage server 500 with interfaces for one or more conventional SATA drives. A network connection 502 (such as a standard electrical or optical connector port or antenna) allows clients to communicate with aPHY 510 supporting a desired protocol, such as Fibre Channel, Ethernet, etc. For Ethernet, a commercially available PHY such as a Broadcom BCM5421 or a Marvell 88E1011 can be used. -
PHY 510 supplies input frames to, and drives output frames from, anRSP 520.RSP 520 can be configured in one of the configurations described above, one of the configurations described in copending application Ser. No. 10/351,030, or any other functionally similar semantic processing configuration. -
RAM 530 provides physical storage, e.g., for machine context and buffers, toRSP 520.RAM 530 may comprise several types of memory (DRAM, Flash, CAM, etc.), or a single type such as Synchronous DRAM. A boot ROM or boot Flash memory can be used to initializeRSP 520 when parser, production rule, and semantic code tables are stored in volatile memory during operation. Part of the non-volatile table storage could also exist on the data storage device(s), as long as enough code is available in boot memory to allowRSP 520 to communicate with the data storage device(s). - A
SATA controller 540 connects to a block I/O port ofRSP 520 to serve the disk access requests ofRSP 520.SATA controller 540 can be, e.g., a commercially available SATA controller such as a SII 3114 or a Marvell 88SX5080. -
Serial ATA controller 540 connects to one or more SATA data storage devices via SATA bus(es). As shown,storage server 500 supports drives 550-0 through 550-N, respectively, through SATA buses BUS0 through BUSN. -
PHY 510,RSP 520,RAM 530, andSATA controller 540 are preferably interconnected on a common printed circuit board (not shown). The circuit board can be arranged with the drives in a common enclosure, with SATA cabling providing BUS0 through BUSN. Alternately, SATA bus signaling can be routed on the printed circuit board to connectors for each drive, or through a connector to a backplane. - Another
storage server implementation 600 is shown inFIG. 8 . Preferably, in this implementation the entire server is implemented on the printed circuit board of a disk drive, with anetwork connection 602 for providing a connection to storage server clients. Althoughstorage server architecture 500 could possibly be packaged in the same way,storage server 600 achieves space and cost savings by integrating aPHY 610 and anSATA controller 640 within anRSP 620.SATA controller 640 preferably is at least partially implemented using a SEE. - The SATA controller section of
RSP 620 interfaces through an SATA cable or circuit board bus to driveelectronics 660, which include an SATA interface, disk cache, and drive control electronics. - With an entire storage server implemented on a common circuit board within a media device, it is possible to remove the SATA interface altogether, as shown in
storage server 700 ofFIG. 9 . InFIG. 9 , disk controller functionality is implemented within anRSP 740. The control devices (such as a Marvell 88C7500), motors, and servos shown asblock 770 interface directly withRSP 740. -
FIG. 10 illustrates astorage server 800 that is essentially a “translation” gateway.Storage server 800 comprises a first network physical interface PHY1 (block 810) to connect to anetwork 802, and consequently to storage server clients.Storage server interface 800 also contains a second network physical interface PHY2 (block 840) to connect, across a network or point-to-point connection, to one or more physical servers, such as the illustratedSAN server 850. AnRSP 820, with attachedmemory 830, attaches to PHY1 and PHY2.PHY 1 can be, e.g., an Ethernet PHY, and PHY2 can be, e.g., a Fibre Channel PHY. - In operation,
RSP 820 responds to client requests, such as NAS-style or application-style requests, by initiating requests on the remotely-attachedserver 850. Thusserver 800 appears as a server to clients onnetwork 802, and as a client toserver 850. Althoughserver 850 is illustrated as a SAN server,server 850 could be a NAS or even an application-style server. If PHY1 and PHY2 support the same server protocol,storage server 800 can still serve a useful function as an aggregation point for a scalable server farm, an encryption point, and/or firewall. - When both PHY1 and PHY2 are supplying datagrams to
RSP 820,RSP 820 can provide parsing for both input streams. For instance, both PHY1 and PHY2 input streams can be saved to a common input buffer, and a DXP (not shown) inRSP 820 can alternately parse datagrams in both input streams, thus coordinating two-way translation tasks.RSP 820 could also be configured with two DXPs, one to serve each physical port, sharing a common bank of SEEs and other RSP resources. - The preceding embodiments can be adapted to other system architectures that provide access to multiple physical data storage devices, such as those shown in
FIGS. 11 and 12 . InFIG. 11 , astorage server 900 contains a storageinterface connecting RSP 100 to multiple physical data storage devices, e.g., four devices 120-1, 120-2, 120-3, and 120-4 as shown. These data storage devices can be accessed in some embodiments as a RAID (Redundant Array of Independent Disks) array, with either the RSP acting as a RAID controller, or with a separate RAID controller (not shown) implemented as part ofstorage interface 110. The four data storage devices can alternately be configured as a JBOD (Just a Bunch Of Disks) array. - Generally,
RSP 100 can perform other forms of storage server virtualization for its clients. In one alternate form, disk sector addresses used by a client application are mapped by the RSP into C-H-S (Cylinder-Head-Sector) addresses of a specific device from one of many physical devices. In such a configuration, data storage devices 120-1 through 120-4 (and potentially many more such devices) need not be geographically co-located, allowing sophisticated control of physical resources and disk allocation for clients. The RSP can also function within a virtualization configuration where a separate intermediate device provides all or part of the virtualization functionality for clients. -
FIG. 12 illustrates the use of anRSP 100 in astorage server system 1000 that incorporates aport extender 950.Port extender 950 connects to a single SATA controller port associated with RSP 100 (either internal or external), but communicates in turn with multiple physical data storage devices (e.g., devices 120-1 to 120-4 as shown). Such a configuration can increase overall system bandwidth while using only one port on the RSP for communication with data storage devices. -
FIG. 13 shows yet anothersemantic processor implementation 1100 useful with embodiments of the present invention.Semantic processor 1100 communicates with aPort 0PHY 93, aPort 1PHY 94, and a PCI-X interface 95, each of which can be integrated withprocessor 1100 or connected externally toprocessor 1100, as desired. Buffer 130 ofFIG. 3 is replaced with a Port Input Buffer (PIB) 132 and a Port Output Buffer (POB) 134, which preferably, but not necessarily, reside in a common memory section integrated as part ofprocessor 1100.PIB 132 connects to and maintains at least one input queue for each ofPort 0,Port 1, and the PCI-X interface, and may contains other queues as well.POB 134 connects to and maintains at least one output queue for each ofPort 0,Port 1, and the PCI-X interface, and may contain other queues as well.Port 0 andPort 1 typically represent datagram interfaces such as Gigabit Ethernet, while the PCI-X interface couples to a familiar PXI-X bus. Depending on the storage server design, storage resources can couple to any of these ports, and clients can couple to eitherPort 1, Port 2, or both. - Unlike
FIG. 3 , inFIG. 13 direct execution parser 140 can receive input from each of the multiple input queues inPIB 132, through aparser source selector 136. For instance,direct execution parser 140 can maintain separate parser stacks for each input source, and can signal parser source selector to switch input sources each time a packet has finished parsing, or when parsing of one packet is stalled, e.g., while a SEE performs some calculation on a packet field. -
SEE cluster 152 comprises N semantic code execution engines 150-1 to 150-N. The SEE cluster preferably includes a single dispatcher (not shown) to communicate withDXP 140 over the Sx-Bus, and to distribute tasks to the individual SEEs. - A PIB read
controller 133 and a POBwrite controller block 135 provide access, respectively, toPIB 132 andPOB 134 forSEE cluster 152. As shown,parser source selector 136 and PIB readcontroller 133 allowDXP 140 to work with data from one input while the SEE cluster accesses data from another input. PIB readcontroller 133 and POB writecontroller 135 preferably allow non-blocking access to the input and output buffers. - A machine
context port controller 154 provides access tomachine context 160 forSEE cluster 152. Like the port buffer controllers, machinecontext port controller 154 provides non-blocking access tomachine context 160 forSEE cluster 152. -
Machine context 160 prioritizes and executes memory tasks forSEE cluster 152 and for amanagement microprocessor 195, and typically contains one or more specialized caches that each depend on the attributes of the target data access.Machine context 160 can also contains one or more encryption and/or authentication engines that can be used to perform inline encryption/authentication. One or more traditional memory bus interfaces connectmachine context 160 withphysical memory 165, which can consist, e.g., of DRAM (Dynamic Random Access Memory), CAM (Content Addressable Memory), and/or any other type of desired storage.Physical memory 165 can be located onprocessor 1100, external toprocessor 1100, or split between these two locations. -
Management microprocessor 195 performs any desired functions forsemantic processor 1100 that can reasonably be accomplished with traditional software. For instance,microprocessor 195 can interface with its own instruction space and data space inphysical memory 165, throughmachine context 160, and execute traditional software to boot the processor, load or modify the parser table, production rule table, and semantic code table, gather statistics, perform logging, manage client access, perform error recovery, etc. Preferably,microprocessor 195 also has the capability to communicate with the dispatcher inSEE cluster 152 in order to request that a SEE perform tasks on the microprocessor's behalf. The management microprocessor is preferably integrated intosemantic processor 1100. - Within the preceding embodiments, the semantic units can encrypt the blocks of data as the blocks are written to disk and/or decrypt the blocks of data as the blocks are read from disk. This provides security to the data “at rest” on the disk drive or drives. For instance, as the semantic units receive data payloads for writing to disk, one operation in preparing the data payloads for writing can include encrypting the packets. The reverse of this process can be used for reading encrypted data from the disk.
- Although special-purpose execution engines have been shown and described, an alternative implementations can use the parser as a front-end datagram processor for a general purpose processor.
- As exemplified by the preceding embodiments, many different semantic-processor based storage servers fall within the scope of the present invention. At the low-functionality end, a SAN-style server, e.g., connected by Fibre Channel to an enterprise network, is possible. NAS-style servers include more sophisticated features, such as those described in the detailed CIFS embodiment described above. Translation-gateway servers provide clients with access to a potentially large and changing underlying array of physical servers without allowing those clients visibility to the physical servers.
- Both translation-gateway servers and physical servers can be structured as application style-servers. As an application-style server, data is served in an application-style format. For instance, a video-on-demand server can store videos for usage by multiple clients, and stream different parts of a video to different clients, allowing each to have independent navigation. A music-on-demand server can operate similarly.
- Application-style servers can also supply storage space for applications, like a wired or wireless server that stores digital camera pictures off of a digital camera, allowing the camera to operate with a relatively small internal buffer and/or flash card. Such a server could be relatively small, battery-powered, and portable for field use.
- Semantic processor storage server implementations can also implement wireless protocols, e.g., for use in a home computer data and/or audio/video network. Although the detailed embodiments have focused on traditional “storage” devices, printers, scanners, multi-function printers, and other data translation devices can also be attached to a semantic processor storage server.
- One of ordinary skill in the art will recognize that the concepts taught herein can be tailored to a particular application in many other advantageous ways. It is readily apparent that a semantic processor storage server can be made to serve different client types by changing the protocols that can be parsed by the server. If desirable, the storage server could even parse multiple storage server protocols simultaneously to allow access by different classes of clients (e.g., SAN access by trusted clients and NAS access by others).
- Those skilled in the art recognize that other functional partitions are possible within the scope of the invention. Further, what functions are and are not implemented on a common integrated circuit can vary depending on application.
- Finally, although the specification may refer to “an”, “one”, “another”, or “some” embodiment(s) in several locations, this does not necessarily mean that each such reference is to the same embodiment(s), or that the feature only applies to a single embodiment.
Claims (10)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/766,733 US20070250593A1 (en) | 2004-05-11 | 2007-06-21 | Reconfigurable semantic processor |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/843,727 US7251722B2 (en) | 2004-05-11 | 2004-05-11 | Semantic processor storage server architecture |
US11/766,733 US20070250593A1 (en) | 2004-05-11 | 2007-06-21 | Reconfigurable semantic processor |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/843,727 Continuation US7251722B2 (en) | 2004-05-11 | 2004-05-11 | Semantic processor storage server architecture |
Publications (1)
Publication Number | Publication Date |
---|---|
US20070250593A1 true US20070250593A1 (en) | 2007-10-25 |
Family
ID=35426734
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/843,727 Expired - Fee Related US7251722B2 (en) | 2004-05-11 | 2004-05-11 | Semantic processor storage server architecture |
US11/766,733 Abandoned US20070250593A1 (en) | 2004-05-11 | 2007-06-21 | Reconfigurable semantic processor |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/843,727 Expired - Fee Related US7251722B2 (en) | 2004-05-11 | 2004-05-11 | Semantic processor storage server architecture |
Country Status (2)
Country | Link |
---|---|
US (2) | US7251722B2 (en) |
CN (1) | CN101128795A (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080114974A1 (en) * | 2006-11-13 | 2008-05-15 | Shao Yi Chien | Reconfigurable image processor and the application architecture thereof |
US20090142068A1 (en) * | 2005-09-13 | 2009-06-04 | Nikon Corporation | Data Communication Apparatus, Electronic Camera, and Data Communication System |
US20100161759A1 (en) * | 2008-12-22 | 2010-06-24 | Ctera Networks Ltd. | Storage device and method thereof for integrating network attached storage with cloud storage services |
US9473419B2 (en) | 2008-12-22 | 2016-10-18 | Ctera Networks, Ltd. | Multi-tenant cloud storage system |
US9521217B2 (en) | 2011-08-08 | 2016-12-13 | Ctera Networks, Ltd. | System and method for remote access to cloud-enabled network devices |
US10521423B2 (en) | 2008-12-22 | 2019-12-31 | Ctera Networks, Ltd. | Apparatus and methods for scanning data in a cloud storage service |
US10783121B2 (en) | 2008-12-22 | 2020-09-22 | Ctera Networks, Ltd. | Techniques for optimizing data flows in hybrid cloud storage systems |
Families Citing this family (40)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030172295A1 (en) * | 2002-03-01 | 2003-09-11 | Onspec Electronics, Inc. | Device and system for allowing secure identification of an individual when accessing information and a method of use |
US7424571B2 (en) * | 2004-07-27 | 2008-09-09 | Gigafin Networks, Inc. | Array machine context data memory |
US7251722B2 (en) * | 2004-05-11 | 2007-07-31 | Mistletoe Technologies, Inc. | Semantic processor storage server architecture |
EP1617341A1 (en) * | 2004-07-16 | 2006-01-18 | International Business Machines Corporation | File operation management device |
US20060026377A1 (en) * | 2004-07-27 | 2006-02-02 | Somsubhra Sikdar | Lookup interface for array machine context data memory |
US20070027991A1 (en) * | 2005-07-14 | 2007-02-01 | Mistletoe Technologies, Inc. | TCP isolation with semantic processor TCP state machine |
US20070019661A1 (en) * | 2005-07-20 | 2007-01-25 | Mistletoe Technologies, Inc. | Packet output buffer for semantic processor |
US7680962B2 (en) * | 2004-12-22 | 2010-03-16 | Nec Electronics Corporation | Stream processor and information processing apparatus |
US8073899B2 (en) * | 2005-04-29 | 2011-12-06 | Netapp, Inc. | System and method for proxying data access commands in a storage system cluster |
US20070162626A1 (en) * | 2005-11-02 | 2007-07-12 | Iyer Sree M | System and method for enhancing external storage |
US20080016266A1 (en) * | 2006-07-14 | 2008-01-17 | Yi-Chun Liu | Method And Apparatus For Smart Storage Devices |
US7876894B2 (en) * | 2006-11-14 | 2011-01-25 | Mcm Portfolio Llc | Method and system to provide security implementation for storage devices |
US7711213B2 (en) * | 2007-01-29 | 2010-05-04 | Hewlett-Packard Development Company, L.P. | Nanowire-based modulators |
US20090046858A1 (en) * | 2007-03-21 | 2009-02-19 | Technology Properties Limited | System and Method of Data Encryption and Data Access of a Set of Storage Devices via a Hardware Key |
US20080288782A1 (en) * | 2007-05-18 | 2008-11-20 | Technology Properties Limited | Method and Apparatus of Providing Security to an External Attachment Device |
US20080181406A1 (en) * | 2007-01-30 | 2008-07-31 | Technology Properties Limited | System and Method of Storage Device Data Encryption and Data Access Via a Hardware Key |
US20080184035A1 (en) * | 2007-01-30 | 2008-07-31 | Technology Properties Limited | System and Method of Storage Device Data Encryption and Data Access |
US20080288703A1 (en) * | 2007-05-18 | 2008-11-20 | Technology Properties Limited | Method and Apparatus of Providing Power to an External Attachment Device via a Computing Device |
US7882202B2 (en) * | 2008-04-01 | 2011-02-01 | International Business Machines Corporation | System to delegate virtual storage access method related file operations to a storage server using an in-band RPC mechanism |
US9477727B2 (en) * | 2008-08-01 | 2016-10-25 | Sybase, Inc. | Abstracting data for use by a mobile device having occasional connectivity |
US8434097B2 (en) * | 2009-12-30 | 2013-04-30 | Sybase, Inc. | Dynamic data binding for MBOs for container based application |
US8321392B2 (en) * | 2009-12-30 | 2012-11-27 | Sybase, Inc. | Pending state management for mobile business objects |
US8909662B2 (en) * | 2009-12-30 | 2014-12-09 | Sybase, Inc. | Message based mobile object with native PIM integration |
US9336291B2 (en) * | 2009-12-30 | 2016-05-10 | Sybase, Inc. | Message based synchronization for mobile business objects |
US8788458B2 (en) * | 2009-12-30 | 2014-07-22 | Sybase, Inc. | Data caching for mobile applications |
US10102242B2 (en) | 2010-12-21 | 2018-10-16 | Sybase, Inc. | Bulk initial download of mobile databases |
US8892569B2 (en) | 2010-12-23 | 2014-11-18 | Ianywhere Solutions, Inc. | Indexing spatial data with a quadtree index having cost-based query decomposition |
US8375031B2 (en) * | 2011-02-10 | 2013-02-12 | Tektronix, Inc. | Lossless real-time line-rate filtering using PCAP style filters and hardware assisted patricia trees |
JP2014522011A (en) * | 2011-05-17 | 2014-08-28 | データボード・インコーポレーテツド | Providing access to mainframe data objects in a heterogeneous computing environment |
US8874682B2 (en) | 2012-05-23 | 2014-10-28 | Sybase, Inc. | Composite graph cache management |
US9110807B2 (en) | 2012-05-23 | 2015-08-18 | Sybase, Inc. | Cache conflict detection |
US9477467B2 (en) * | 2013-03-30 | 2016-10-25 | Intel Corporation | Processors, methods, and systems to implement partial register accesses with masked full register accesses |
CN105408875B (en) * | 2014-02-23 | 2019-06-21 | 拉姆伯斯公司 | Distributed process on memory interface executes and file system |
CN111427534B (en) * | 2014-12-11 | 2023-07-25 | 微软技术许可有限责任公司 | Virtual assistant system capable of actionable messaging |
CN105791126B (en) * | 2014-12-26 | 2020-04-24 | 中兴通讯股份有限公司 | Ternary Content Addressable Memory (TCAM) table look-up method and device |
US9954681B2 (en) * | 2015-06-10 | 2018-04-24 | Nxp Usa, Inc. | Systems and methods for data encryption |
CN105047029B (en) * | 2015-07-20 | 2018-02-13 | 北京大学 | A kind of courseware bringing method of automatic Display geometry topic course of solving questions |
CN109918673B (en) * | 2019-03-14 | 2021-08-03 | 湖北亿咖通科技有限公司 | Semantic arbitration method and device, electronic equipment and computer-readable storage medium |
CN110011807B (en) * | 2019-03-25 | 2021-12-24 | 创新先进技术有限公司 | Key information maintenance method and system |
CN111061482B (en) * | 2019-10-24 | 2023-12-08 | 贝壳技术有限公司 | Method and device for analyzing parameters in character string, storage medium and electronic equipment |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4994966A (en) * | 1988-03-31 | 1991-02-19 | Emerson & Stern Associates, Inc. | System and method for natural language parsing by initiating processing prior to entry of complete sentences |
US5644686A (en) * | 1994-04-29 | 1997-07-01 | International Business Machines Corporation | Expert system and method employing hierarchical knowledge base, and interactive multimedia/hypermedia applications |
US5715468A (en) * | 1994-09-30 | 1998-02-03 | Budzinski; Robert Lucius | Memory system for storing and retrieving experience and knowledge with natural language |
US5916305A (en) * | 1996-11-05 | 1999-06-29 | Shomiti Systems, Inc. | Pattern recognition in data communications using predictive parsers |
US7251722B2 (en) * | 2004-05-11 | 2007-07-31 | Mistletoe Technologies, Inc. | Semantic processor storage server architecture |
Family Cites Families (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5487147A (en) | 1991-09-05 | 1996-01-23 | International Business Machines Corporation | Generation of error messages and error recovery for an LL(1) parser |
US5793954A (en) | 1995-12-20 | 1998-08-11 | Nb Networks | System and method for general purpose network analysis |
US6493761B1 (en) | 1995-12-20 | 2002-12-10 | Nb Networks | Systems and methods for data processing using a protocol parsing engine |
US6034963A (en) | 1996-10-31 | 2000-03-07 | Iready Corporation | Multiple network protocol encoder/decoder and data processor |
US6330659B1 (en) | 1997-11-06 | 2001-12-11 | Iready Corporation | Hardware accelerator for an object-oriented programming language |
US6145073A (en) | 1998-10-16 | 2000-11-07 | Quintessence Architectures, Inc. | Data flow integrated circuit architecture |
US6356950B1 (en) | 1999-01-11 | 2002-03-12 | Novilit, Inc. | Method for encoding and decoding data according to a protocol specification |
US7421586B2 (en) * | 1999-05-12 | 2008-09-02 | Fraunhofer Gesselschaft | Protecting mobile code against malicious hosts |
US6772413B2 (en) | 1999-12-21 | 2004-08-03 | Datapower Technology, Inc. | Method and apparatus of data exchange using runtime code generator and translator |
US6985964B1 (en) | 1999-12-22 | 2006-01-10 | Cisco Technology, Inc. | Network processor system including a central processor and at least one peripheral processor |
US6892237B1 (en) | 2000-03-28 | 2005-05-10 | Cisco Technology, Inc. | Method and apparatus for high-speed parsing of network messages |
US7020128B1 (en) * | 2000-11-02 | 2006-03-28 | Efficient Networks, Inc. | System and method for datagram staggering in a communication system |
US20040215665A1 (en) * | 2002-01-09 | 2004-10-28 | Edgar David A. | System, method, and computer program product for providing accelerated and secure wireless data transmission over the internet |
EP1546892B1 (en) * | 2002-09-04 | 2017-03-29 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Protecting mobile code against malicious hosts cross references to related applications |
US7130987B2 (en) * | 2003-01-24 | 2006-10-31 | Mistletoe Technologies, Inc. | Reconfigurable semantic processor |
US7716364B2 (en) * | 2003-06-27 | 2010-05-11 | Broadcom Corporation | Internet protocol multicast replication |
-
2004
- 2004-05-11 US US10/843,727 patent/US7251722B2/en not_active Expired - Fee Related
-
2005
- 2005-05-11 CN CNA2005800227372A patent/CN101128795A/en active Pending
-
2007
- 2007-06-21 US US11/766,733 patent/US20070250593A1/en not_active Abandoned
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4994966A (en) * | 1988-03-31 | 1991-02-19 | Emerson & Stern Associates, Inc. | System and method for natural language parsing by initiating processing prior to entry of complete sentences |
US5644686A (en) * | 1994-04-29 | 1997-07-01 | International Business Machines Corporation | Expert system and method employing hierarchical knowledge base, and interactive multimedia/hypermedia applications |
US5715468A (en) * | 1994-09-30 | 1998-02-03 | Budzinski; Robert Lucius | Memory system for storing and retrieving experience and knowledge with natural language |
US5916305A (en) * | 1996-11-05 | 1999-06-29 | Shomiti Systems, Inc. | Pattern recognition in data communications using predictive parsers |
US7251722B2 (en) * | 2004-05-11 | 2007-07-31 | Mistletoe Technologies, Inc. | Semantic processor storage server architecture |
Cited By (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8582973B2 (en) * | 2005-09-13 | 2013-11-12 | Nikon Corporation | Data communication apparatus, electronic camera, and data communication system |
US20090142068A1 (en) * | 2005-09-13 | 2009-06-04 | Nikon Corporation | Data Communication Apparatus, Electronic Camera, and Data Communication System |
US20080114974A1 (en) * | 2006-11-13 | 2008-05-15 | Shao Yi Chien | Reconfigurable image processor and the application architecture thereof |
US8027551B2 (en) * | 2006-11-13 | 2011-09-27 | Vivotek Inc. | Reconfigurable image processor and the application architecture thereof |
US9473419B2 (en) | 2008-12-22 | 2016-10-18 | Ctera Networks, Ltd. | Multi-tenant cloud storage system |
US8924511B2 (en) | 2008-12-22 | 2014-12-30 | Ctera Networks Ltd. | Cloud connector for interfacing between a network attached storage device and a cloud storage system |
US20100161759A1 (en) * | 2008-12-22 | 2010-06-24 | Ctera Networks Ltd. | Storage device and method thereof for integrating network attached storage with cloud storage services |
US9614924B2 (en) * | 2008-12-22 | 2017-04-04 | Ctera Networks Ltd. | Storage device and method thereof for integrating network attached storage with cloud storage services |
US10375166B2 (en) | 2008-12-22 | 2019-08-06 | Ctera Networks, Ltd. | Caching device and method thereof for integration with a cloud storage system |
US10521423B2 (en) | 2008-12-22 | 2019-12-31 | Ctera Networks, Ltd. | Apparatus and methods for scanning data in a cloud storage service |
US10574753B2 (en) | 2008-12-22 | 2020-02-25 | Ctera Networks, Ltd. | Data files synchronization with cloud storage service |
US10783121B2 (en) | 2008-12-22 | 2020-09-22 | Ctera Networks, Ltd. | Techniques for optimizing data flows in hybrid cloud storage systems |
US11178225B2 (en) | 2008-12-22 | 2021-11-16 | Ctera Networks, Ltd. | Data files synchronization with cloud storage service |
US9521217B2 (en) | 2011-08-08 | 2016-12-13 | Ctera Networks, Ltd. | System and method for remote access to cloud-enabled network devices |
Also Published As
Publication number | Publication date |
---|---|
US7251722B2 (en) | 2007-07-31 |
CN101128795A (en) | 2008-02-20 |
US20050268032A1 (en) | 2005-12-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7251722B2 (en) | Semantic processor storage server architecture | |
KR100201574B1 (en) | Parallel i/o network file server architecture | |
US7664892B2 (en) | Method, system, and program for managing data read operations on network controller with offloading functions | |
US7181439B1 (en) | System and method for transparently accessing a virtual disk using a file-based protocol | |
US8321658B2 (en) | Method and system for iSCSI boot in which an iSCSI client loads boot code from a host bus adapter and/or network interface card | |
US7415596B2 (en) | Parser table/production rule table configuration using CAM and SRAM | |
US7849274B2 (en) | System and method for zero copy block protocol write operations | |
RU2302034C9 (en) | Multi-protocol data storage device realizing integrated support of file access and block access protocols | |
US7249173B2 (en) | Abstracted node discovery | |
US20060230148A1 (en) | TCP forwarding of client requests of high-level file and storage access protocols in a network file server system | |
US20080155051A1 (en) | Direct file transfer system and method for a computer network | |
US20040210584A1 (en) | Method and apparatus for increasing file server performance by offloading data path processing | |
US20090254647A1 (en) | System and method for network interfacing | |
US20050281281A1 (en) | Port input buffer architecture | |
US8185633B1 (en) | Method and apparatus for offloading network processes in a computer storage system | |
US9401921B2 (en) | Insuring integrity of remote procedure calls used in a client and server storage system | |
US9891929B2 (en) | System and method for redirecting input/output (I/O) sequences | |
WO2005111813A2 (en) | Semantic processor storage server architecture | |
US20070174034A1 (en) | Transparent intellectual network storage device | |
JP2004259045A (en) | Storage device | |
Ma | Enhancing functionality in networked storage systems | |
Mathews et al. | Storage and Print Solutions for SOHOs |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: GIGAFIN NETWORKS, INC., CALIFORNIA Free format text: CHANGE OF NAME;ASSIGNOR:MISTLETOE TECHNOLOGIES, INC.;REEL/FRAME:021219/0979 Effective date: 20080708 |
|
AS | Assignment |
Owner name: VENTURE LENDING & LEASING IV, INC, CALIFORNIA Free format text: SECURITY AGREEMENT;ASSIGNOR:GIGAFIN NETWORKS, INC.;REEL/FRAME:021415/0206 Effective date: 20080804 |
|
AS | Assignment |
Owner name: MISTLETOE TECHNOLOGIES, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SIKDAR, SOMSUBHRA;ROWETT, KEVIN;REEL/FRAME:024797/0380 Effective date: 20040506 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |