AU2003277249A1 - Hardware parser accelerator - Google Patents

Hardware parser accelerator Download PDF

Info

Publication number
AU2003277249A1
AU2003277249A1 AU2003277249A AU2003277249A AU2003277249A1 AU 2003277249 A1 AU2003277249 A1 AU 2003277249A1 AU 2003277249 A AU2003277249 A AU 2003277249A AU 2003277249 A AU2003277249 A AU 2003277249A AU 2003277249 A1 AU2003277249 A1 AU 2003277249A1
Authority
AU
Australia
Prior art keywords
state
character
state table
interest
document
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
AU2003277249A
Inventor
Michael C. Dapp
Eric C. Lett
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Lockheed Martin Corp
Original Assignee
Lockheed Corp
Lockheed Martin Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US10/331,315 external-priority patent/US20040083466A1/en
Application filed by Lockheed Corp, Lockheed Martin Corp filed Critical Lockheed Corp
Publication of AU2003277249A1 publication Critical patent/AU2003277249A1/en
Abandoned legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F8/00Arrangements for software engineering
    • G06F8/40Transformation of program code
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Document Processing Apparatus (AREA)
  • Image Generation (AREA)
  • Devices For Executing Special Programs (AREA)

Description

WO 2004/040446 PCT/US2003/031314 1 HARDWARE PARSER ACCELERATOR DESCRIPTION BACKGROUND OF THE INVENTION Field of the Invention 5 The present invention generally relates to processing of applications for controlling the operations of general purpose computers and, more particularly, to performing parsing operations on applications programs, documents and/or other 10 logical sequences of network data packets. Description of the Prior Art The field of digital communications between computers and the linking of computers into networks has developed rapidly in recent years, similar, in 15 many ways to the proliferation of personal computers of a few years earlier. This increase in interconnectivity and the possibility of remote processing has greatly increased the effective capability and functionality of individual computers 20 in such networked systems. Nevertheless, the variety of uses of individual computers and systems, preferences of their users and the state of the art when computers are placed into service has resulted in a substantial degree of variety of capabilities 25 and configurations of individual machines and their operating systems, collectively referred to as "platforms" which are generally incompatible with each other to some degree particularly at the level WO 2004/040446 PCT/US2003/031314 2 of operating system and programming language. This incompatibility of platform characteristics and the simultaneous requirement for the capability of communication and remote 5 processing and a sufficient degree of compatibility to support it has resulted in the development of object oriented programming (which accommodates the concept of assembling an application as well as data as a group of more or less generalized modules 10 through a referencing system of entities, attributes and relationships) and a number of programming languages to embody it. Extensible Markup Language TM (XMLTM) is such a language which has come into widespread use and can be transmitted as a document 15 over a network of arbitrary construction and architecture. In such a language, certain character strings correspond to certain commands or identifications, including special characters and other important 20 data (collectively referred to as control words) which allow data or operations to, in effect, identify themselves so that they may be thereafter treated as "objects" such that associated data and commands can be translated into the appropriate 25 formats and commands of different applications in different languages in order to engender a degree of compatibility of respective connected platforms sufficient to support the desired processing at a given machine. The detection of these character 30 strings is performed by an operation known as parsing, similar to the more conventional usage of resolving the syntax of an expression, such as a sentence, into its component parts and describing them grammatically.
WO 2004/040446 PCT/US2003/031314 3 When parsing an XMLTM document, a large portion and possibly a majority of the central processor unit (CPU) execution time is spent traversing the document searching for control words, special 5 characters and other important data as defined for the particular XML TM standard being processed. This is typically done by software which queries each character and determines if it belongs to the predefined set of strings of interest, for example, 10 a set of character strings comprising the following "<command>", "<data = dataword>", "<endcommand>", etc. If any of the target strings are detected, a token is saved with a pointer to the location in the document for the start of the token and the length 15 of the token. These tokens are accumulated until the entire document has been parsed. The conventional approach is to implement a table-based finite state machine (FSM) in software to search for these strings of interest. The state 20 table resides in memory and is designed to search for the specific patterns in the document. The current state is used as the base address into the state table and the ASCII representation of the input character is an index into the table. For 25 example, assume the state machine is in state 0 (zero) and the first input character is ASCII value 02, the absolute address for the state entry would be the sum/concatenation of the base address (state 0) and the index/ASCII character (02). The FSM 30 begins with the CPU fetching the first character of the input document from memory. The CPU then constructs the absolute address into the state table in memory corresponding to the initialized/current state and the input character and then fetches the WO 2004/040446 PCT/US2003/031314 4 state data from the state table. Based on the state data that is returned, the CPU updates the current state to the new value, if different (indicating that the character corresponds to the first 5 character of a string of interest) and performs any other action indicated in the state data (e.g. issuing a token or an interrupt if the single character is a special character or if the current character is found, upon a further repetition of the 10 foregoing, to be the last character of a string of interest). The above process is repeated and the state is changed as successive characters of a string-of interest are found. That is, if the initial 15 character is of interest as being the initial character of a string of interest, the state of the FSM can be advanced to a new state (e.g. from initial state 0 to state 1). If the character is not of interest, the state machine would (generally) 20 remain the same by specifying the same state (e.g. state 0) or not commanding a state update) in the state table entry that is returned from the state table address. Possible actions include, but are not limited to, setting interrupts, storing tokens 25 and updating pointers. The process is then repeated with the following character. It should be noted that while a string of interest is being followed and the FSM is in a state other than state 0 (or other state indicating that a string of interest has 30 not yet been found or currently being followed) a character may be found which is not consistent with a current string but is an initial character of another string of interest. In such a case, state table entries would indicate appropriate action to WO 2004/040446 PCT/US2003/031314 5 indicate and identify the string fragment or portion previously being followed and to follow the possible new string of interest until the new string is completely identified or found not to be a string of 5 interest. In other words, strings of interest may be nested and the state machine must be able to detect a string of interest within another string of interest, and so on. This may require the CPU to traverse portions of the XMLTM document numerous 10 times to completely parse the XML T M document. The entire XMLTM or other language document is parsed character-by-character in the above-described manner. As potential target strings are recognized, the FSM steps through various states character-by 15 character until a string of interest is fully identified or a character inconsistent with a possible string of interest is encountered (e.g. when the string is completed/fully matched or a character deviates from a target string). In the 20 latter case, no action is generally taken other than returning to the initial state or a state corresponding to the detection of an initial character of another target string. In the former case, the token is stored into memory along with the 25 starting address in the input document and the length of the token. When the parsing is completed, all objects will have been identified and processing in accordance with the local or given platform can be started. 30 Since the search is generally conducted for multiple strings of interest, the state table can provide multiple transitions from any given state. This approach allows the current character to be analyzed for multiple target strings at the same WO 2004/040446 PCT/US2003/031314 6 time while conveniently accommodating nested strings. It can be seen from the foregoing that the parsing of a document such as an XMLTM document 5 requires many repetitions and many memory accesses for each repetition. Therefore, processing time on a general purpose CPU is necessarily substantial. A further major complexity of handling the multiple strings lies in the generation of the large state 10 tables and is handled off-line from the real-time packet processing. However, this requires a large number of CPU cycles to fetch the input character data, fetch the state data and update the various pointers and state addresses for each character in 15 the document. Thus, it is relatively common for the parsing of a document such as an XMLTM document to fully pre-empt other processing on the CPU or platform and to substantially delay the processing requested. 20 It has been recognized in the art that, through programming, general-purpose hardware can be made to emulate the function of special purpose hardware and that special purpose data processing hardware will often function more rapidly than programmed general 25 purpose hardware even if the structure and program precisely correspond to each other since there is less overhead involved in managing and controlling special purpose hardware. Nevertheless, the hardware resources required for certain processing 30 may be prohibitively large for special purpose hardware, particularly where the processing speed gain may be marginal. Further, special purpose hardware necessarily has functional limitations and providing sufficient flexibility for certain WO 2004/040446 PCT/US2003/031314 7 applications such as providing the capability of searching for an arbitrary number of arbitrary combinations of characters may also be prohibitive. Thus, to be feasible, special purpose hardware must 5 provide a large gain in processing speed while providing very substantial hardware economy; requirements which are increasingly difficult to accommodate simultaneously as increasing amounts of functional flexibility or programmability are needed 10 in the processing function required. In this regard, the issue of system security is also raised by both interconnectability and the amount of processing time required for parsing a document such as an XML
T
M document. On the one hand, 15 any process which requires an extreme amount of processing time at relatively high priority is, in some ways, similar to some characteristics of a denial-of-service (DOS) attack on the system or a node thereof or can be a tool that can be used in 20 such an attack. DOS attacks frequently present frivolous or malformed requests for service to a system for the purpose of maliciously consuming and eventually overloading available resources. Proper 25 configuration of hardware accelerators can greatly reduce or eliminate the potential for overloading of available resources. In addition, systems often fail or expose security weaknesses when overloaded. Thus, eliminating overloads is an important security 30 consideration. Further, it is possible for some processing to begin and some commands to be executed before parsing is completed since the state table must be able to contain CPU commands at basic levels which WO 2004/040446 PCT/US2003/031314 8 are difficult or impossible to secure without severe compromise of system performance. In short, the potential for compromise of security is necessarily reduced by reduction of processing time for 5 processes such as XMLTM parsing.
WO 2004/040446 PCT/US2003/031314 9 SUMMARY OF THE INVENTION The invention provides a dedicated processor and associated hardware for accelerating the parsing process for documents such as XMLTM documents while 5 limiting the amount of hardware and memory required. In order to accomplish these and other capabilities of the invention, a hardware parser accelerator is provided including a document memory, a character pallette containing addresses 10 corresponding to characters in the document, a state table containing a plurality of entries corresponding to a character, a next state pallette including a state address or offset, and a token buffer, wherein entries in said state table include 15 at least one of an address into said next state pallette and a token.
WO 2004/040446 PCT/US2003/031314 10 BRIEF DESCRIPTION OF THE DRAWINGS The foregoing and other objects, aspects and advantages will be better understood from the following detailed description of a preferred 5 embodiment of the invention with reference to the drawings, in which: Figure 1 is a representation of a portion of a state table used in parsing a document, Figure 2 is a high level schematic diagram of 10 the parser accelerator in accordance with the invention, Figure 3 illustrates a preferred character palette format as depicted in Figure 2, Figures 4A and 4B illustrate a state table 15 format and a state table control register used in conjunction therewith in a preferred form of the invention as depicted in Figure 2, Figure 5 illustrates a preferred next state palette format as depicted in Figure 2, and 20 Figure 6 is a preferred token format as depicted in Figure 5.
WO 2004/040446 PCT/US2003/031314 11 DETAILED DESCRIPTION OF A PREFERRED EMBODIMENT OF THE INVENTION Referring now to the drawings, and more particularly to Figure 1, there is shown a 5 representation of a portion of a state table useful in understanding the invention. It should be understood that the state table shown in Figure 1 is potentially only a very small portion of a state table useful for parsing an XMLTM document and is 10 intended to be exemplary in nature. It should be noted that an XMLTM document is used herein as an example of one type of logical data sequence which can be processed using an accelerator in accordance with the invention. Other logical data sequences 15 can also be constructed from network data packet contents such as user terminal command strings intended for execution by shared server computers. While the full state table does not physically exist, at least in the form shown, in the invention 20 and Figure 1 can also be used in facilitating an understanding of the operation of known software parsers, no portion of Figure 1 is admitted to be prior art in regard to the present invention. It will also be helpful observe that many 25 entries in the portion of the state table illustrated in Figure 1 are duplicative and it is important to an appreciation of the invention that hardware to accommodate the entirety of the state table represented by Figure 1 is not required. 30 Conversely, while the invention can be implemented in software, possibly using a dedicated processor, the hardware requirements in accordance with the invention are sufficiently limited that the penalty WO 2004/040446 PCT/US2003/031314 12 in increased processing time for parsing by software is not justified by any possible economy in hardware. In Figure 1 the state table is divided into an 5 arbitrary number of rows, each having a base address corresponding to a state. The rows of the base address are divided into a number of columns corresponding to the number of codes which may be used to represent characters in the document to be 10 parsed; in this example, two hundred fifty-six (256) columns corresponding to a basic eight bit byte for a character which is used as an index into the state table. It will be helpful to note several aspects of 15 the state table entries shown, particularly in conveying an understanding of how even the small portion of the exemplary state table illustrated in Figure 1 supports the detection of many words: 1. In the state table shown, only two entries 20 in the row for state 0 include an entry other than "stay in state 0" which maintains the initial state when the character being tested does not match the initial character of any string of interest. The single entry which provides for progress to state 1 25 corresponds to a special case where all strings of interest begin with the same character. Any other character that would provide progress to another state would generally but not necessary progress to a state other than state 1 but a further reference 30 to the same state that could be reached through another character may be useful to, for example, detect nested strings. The inclusion of a command (e.g. "special interrupt") with "stay in state 0" illustrated at {state 0, FD} would be used to detect WO 2004/040446 PCT/US2003/031314 13 and operate on special single characters. 2. In states above state 0, an entry of "stay in state n" provides for the state to be maintained through potentially long runs of one or more 5 characters such as might be encountered, for example, in numerical arguments of commands, as is commonly encountered. The invention provides special handling of this type of character string to provide enhanced acceleration, as will be discussed 10 in detail below. 3. In states above state 0, an entry of "go to state 0" signifies detection of a character which distinguishes the string from any string of interest, regardless of how many matching characters 15 have previously been detected and returns the parsing process to the initial/default state to begin searching for another string of interest. (For this reason, the "go to state 0" entry will generally be, by far, the most frequent or numerous 20 entry in the state table.) Returning to state 0 may require the parsing operation to return to a character in the document subsequent to the character which began the string being followed at the time the distinguishing character was detected. 25 4. An entry including a command with "go to state 0 indicates completion of detection of a complete string of interest. In general, the command will be to store a token (with an address and length of the token) which thereafter allows the 30 string to be treated as an object. However, a command with "go to state n" provides for launching of an operation at an intermediate point while continuing to follow a string which could potentially match a string of interest.
WO 2004/040446 PCT/US2003/031314 14 5. To avoid ambiguity at any point where the search branches between two strings of interest (e.g. strings having n-1 identical initial characters but different n-th characters, or 5 different initial characters), it is generally necessary to proceed to different (e.g. non consecutive) states, as illustrated at {state 1, 01) and {statel, FD}. Complete identification of a string of arbitrary length n will require n-1 states 10 except for the special circumstances of included strings of special characters and strings of interest which have common initial characters. For these reason, the number of states and rows of the state table must usually be extremely large, even 15 for relatively modest numbers of strings of interest. 7. Conversely to the previous paragraph, most states can be fully characterized by one or two unique entries and a default "go to state 0". This 20 feature of the state table of Figure 1 is exploited in the invention to produce a high degree of hardware economy and substantial acceleration of the parsing process for the general case of strings of interest. 25 As alluded to above, the parsing operation, as conventionally performed, begins with the system in a given default/initial state, depicted in Figure 1 as state 0, and then progresses to higher numbered states as matching characters are found upon 30 repetitions of the process. When a string of interest has been completely identified or when a special operation is specified at an intermediate location in a string which is potentially a match, the operation such as storing a token or issuing an WO 2004/040446 PCT/US2003/031314 15 interrupt is performed. At each repetition for each character of the document, however, the character must be fetched from CPU memory, the state table entry must be fetched (again from CPU memory) and 5 various pointers (e.g. to a character of the document and base address in the state table) and registers (e.g. to the initial matched character address and an accumulated length of the string) must be updated in sequential operations. 10 Therefore, it can be readily appreciated that the parsing operation can consume large amounts of processing time. A high-level schematic block diagram of the parser accelerator 100 in accordance with the 15 invention is illustrated in Figure 2. As will be appreciated by those skilled in the art, Figure 2 can also be understood as a flow diagram illustrating the steps performed in accordance with the invention to perform parsing. As will be 20 discussed in greater detail below in connection with Figures 3, 4A, 4B, 5 and 6, the invention exploits some hardware economies in representing the state table such that a plurality of hardware pipelines are developed which operate essentially in parallel 25 although slightly skewed in time. Thus, the updating of pointers and registers can be performed substantially in parallel and concurrently with other operations while the time required for memory accesses is much reduced through both faster access 30 hardware operated in parallel and prefetching from CPU memory in regard to the state table and the document. As a general overview, the document such as an
XML'
M
document is stored externally in DRAM 120 which WO 2004/040446 PCT/US2003/031314 16 is indexed by registers 112, 114 and transferred by, preferably, thirty-two bit words to and input buffer 130 which serves as a multiplexer for the pipelines. Each pipeline includes a copy of a character palette 5 140, state table 160 and a next state palette 170; each accommodating a compressed form of part of the state table. The output of the next state palette 170 contains both the next state address portion of the address into entries in the state table 160 and 10 the token value to be stored, if any. Operations in the character palette 140 and the next state palette 170 are simple memory accesses into high speed internal SRAM which may be performed in parallel with each other as well as in parallel with simple 15 memory accesses into the high speed external DRAM forming the state table 160 (which may also be implemented as a cache). Therefore, only a relatively few clock cycles of the CPU initially controlling these hardware elements (but which, once 20 started, can function autonomously with only occasional CPU memory operation calls to refresh the document data and to store tokens) are required for an evaluation of each character in the document. The basic acceleration gain is the reduction of the 25 sum of all memory operation durations per character in the CPU plus the CPU overhead to the duration of a single autonomously performed memory operation in high-speed SRAM or DRAM. It should be understood that memory structures 30 referred to herein as "external" is intended to connote a configuration of memories 120, 140, which is preferred by the inventors at the present time in view of the amount of storage required and access from the hardware parser accelerator and/or the host WO 2004/040446 PCT/US2003/031314 17 CPU. In other words, it may be advantageous for handling of tokens and some other operations to provide an architecture of the parser accelerator in accordance with the invention to facilitate sharing 5 of the memory or at least access to the memory by the host CPU as well as the hardware accelerator. No other connotation intended and a wide variety of hardware alternatives such as synchronous DRAM (SDRAM) will be recognized as suitable by those 10 skilled in the art in view of this discussion. Referring now to Figures 3 - 6, the formats of the character palette 140, the state table 160, next state palette 170 and next state and token will be discussed as exemplary of the hardware economies 15 which support the preferred implementation of Figure 2. Other techniques/formats can be employed, as well, and the illustrated formats should be understood as exemplary although currently preferred. 20 Figure 3 illustrates the preferred form of a character palette which corresponds to the characters which are or may be included in the strings of interest. This format preferably provides entries numbered 0 - 255, corresponding to 25 the number of columns in the state table of Figure 1. (The term "palette" is used in much the same sense as in the term "color palette" containing data for each color supported and collectively referred to as a gamut. Use of a pallette reduces 30 entries/columns in the state table.) For example, a character referred to as a "null character" which does not result in a change of state can be expressed in one column of the state table rather than many such columns. It is desirable to test for WO 2004/040446 PCT/US2003/031314 18 a null character output at 144 which can substantially accelerate processing for parsing since it allows immediate processing of the next character without a further memory operation for 5 state table access. The format can be accommodated by a single register or memory locations configured as such by, for example, data in base address register 142 which points to a particular character palette (schematically illustrated by overlapping 10 memory planes in Figure 2). The current eight bit character from the document (e.g. XMLTM document), one of four provided from the input buffer 130 as received as a four byte word from the external DRAM 120, addresses an entry in the character palette 15 which then outputs an address as an index or partial pointer into the state memory. Thus by providing a palette in such a format a portion of the functionality of Figure 1 can be provided in the form of a single register of relatively limited 20 capacity; thus allowing a plurality thereof to be formed and operated in parallel while maintaining substantial hardware economy and supporting others in the state table 160. Figure 4A shows the preferred state table 25 format which is constituted or-configured similarly to the character palette (e.g. substantially as a register). The principal difference from the character palette of Figure 3 is that the length of the register is dependent on the number of responses 30 to characters desired and the number and length of strings of interest. Therefore, it is considered desirable to provide for the possibility of implementing this memory in CPU or other external DRAM (possibly with an internal or external cache) WO 2004/040446 PCT/US2003/031314 19 if the amount of internal memory which can be economically provided is insufficient in particular instances. Nevertheless, it is clear that a substantial hardware economy is provided since 5 highly duplicative entries in the state table of Figure 1 can be reduced to a single entry; the address of which is accommodated by the data provided as described above in accordance with the character palette of Figure 3. The output of the 10 state table 160 is preferably one, two or four bits but provision for as much as thirty-two bits may provide increased flexibility, as will be discussed below in connection with Figure 4B. In any case, the output of the state table provides an address or 15 pointer into the next state palette 170. Referring now to Figure 4B, as a perfecting feature of the invention in this latter regard, a preferred implementation feature of the invention includes a state table control register 162 which 20 allows a further substantial hardware economy, particularly if a thirty-two bit output of state table 160 is to be provided. Essentially, the state table control register provides for compression of the state table information by allowing a variable 25 length word to be stored in and read out of the state table. More specifically, the state table control register 162 stores and provides the length of each entry in the state table 160 of Figure 4A. Since 30 some state table entries in Figure 1 are highly duplicative (e.g. "go to state 0", "stay in state n"), these entries not only can be represented by a single entry in state table 160 or at least much fewer than in Figure 1 but may also be represented WO 2004/040446 PCT/US2003/031314 20 by fewer bits, possibly as few as one which will yield substantial hardware economies even if most or all duplicative entries are included in the state table, as may be found convenient in some state 5 tables. The principle of this reduction will be recognized by those skilled in the art as similar to so-called entropy coding. Referring now to Figure 5, the preferred format of the next state palette 170 will now be discussed. 10 The next state pallette 170 is preferably implemented in much the same manner as the character palette 140 discussed above. However, as with the state memory 160, the number of entries that may be required is not, a priori, known and the length of 15 individual entries is preferably much longer (e.g. two thirty-two bit words). On the other hand, the next state palette 170 can be operated as a cache (e.g. using next state palette base address register 172) since only relatively small and predictable 20 ranges of addresses need be contained at any given time. Further, if thirty-two bit outputs of the state table 160 is provided, some of that data can be used to supplement the data in entries of the next state palette 170, possibly allowing shorter 25 entries in the latter or possibly bypassing the next state pallette altogether, as indicated by dashed line 175. As shown in Figure 5, the lower address thirty two bit word output from the next state palette 170 30 is the token to be saved. This token preferably is formed as a token value of sixteen bits, eight bits of token flags, both of which are stored in token buffer 190 at an address provided by pointer 192 to the beginning of the string and together with the WO 2004/040446 PCT/US2003/031314 21 length accumulated by counting successful character comparisons, and eight bits of control flags. The control flags set interrupts to the host CPU or control processing in the parser accelerator. One 5 of these latter control flags is preferably used to set a skip enable function for characters which do not cause a change of state at a state other than state 0 such as a string of the same or related characters of arbitrary length which may occur in a 10 string of interest, as alluded to above. In such a case, the next state table entry can be reused without fetching it from SRAM/SDRAM. The input buffer address 112 is incremented without additional processing; allowing substantial addition 15 acceleration of parsing for certain strings of characters. The second thirty-two bit word is an address offset fed back to register 180 and adder 150 to be concatenated with the index output from the character palette to form a pointer into the 20 state table for the next character. The initial address corresponding to state 0 is supplied by register 182. Thus, it is seen that the use of a character palette, a state memory in an abbreviated form and a 25 next state memory articulate the function of the conventional state memory operations into separate stages; each of which can be performed extremely rapidly with relatively little high speed memory which can thus be duplicated to form parallel 30 pipelines operating on respective characters of a document in turn and in parallel with other operations and storage of tokens. Therefore, the parsing process can be greatly accelerated relative to even a dedicated processor which must perform all WO 2004/040446 PCT/US2003/031314 22 of these functions in sequence before processing of another character can be started. In summary, the accelerator has access to the program memory of the host CPU where the character 5 data (sometimes referred to as packet data connoting transmission of a network) and state table are located. The accelerator 100 is under control of the main CPU via memory-mapped registers. The accelerator can interrupt the main CPU to indicate 10 exceptions, alarms and terminations. When parsing is to be started, pointers (112, 114) are set to the beginning an end of the input buffer 130 data to be analyzed, the state table to be used (as indicated by base address 182 and other control information 15 (e.g. 142) is set up within the accelerator. To initiate operation of the accelerator, the CPU issues a command to the accelerator which, in response, fetches a first thirty-two bit word of data from the CPU program memory (e.g. 120 or a 20 cache) and places it into the input buffer 130 from which the first byte/ASCII character is selected. The accelerator fetches the state information corresponding to the input character (i.e. Figure 4A corresponds to a single character or a single column 25 of the full state table of Figure 1) and the current state. The state information includes the next state address and any special actions to be performed such as interrupting the CPU or terminating the processing. The advancing of the 30 state information thus supports detection not only of single strings of interest but also nested strings, alluded to above, and sequences of strings or corresponding tokens such as words or phrases of text in a document. The interrupts and or WO 2004/040446 PCT/US2003/031314 23 exceptions which can be issued in response thereto are not limited to internal control of the parser and the issuance of tokens but may generate alerts or other initiate other processing to provide other 5 functions such as intercepting unwanted electronic mail or blocking objectionable subject matter or content-based routing, possibly through issuance of special tokens for. The accelerator next selects the next byte to 10 be analyzed from input buffer 130 and repeats the process with the new state information which will already be available to adder 150. The operation or token information storage can be performed concurrently. This continues until all four 15 characters of the input word have been analyzed. Then (or concurrently with the analysis of the fourth character by prefetching) buffers 112, 114 are compared to determine if the end of the document buffer 120 is reached and, if so, an interrupt is 20 sent back to the CPU. If not, a new word is fetched, the buffer 112 is updated and the processing is repeated. Since the pointers and counters are implemented in dedicated hardware they can be updated in 25 parallel rather than serially as would be required if implemented in software. This reduces the time to analyze a byte of data to the time required to fetch the character from a local input buffer, generate the state table address from high speed 30 local character palette memory, fetch the corresponding state table entry from memory and to fetch the next state information, again from local high speed memory. Some of these operations can be performed concurrently in separate parallel WO 2004/040446 PCT/US2003/031314 24 pipelines and other operations specified in the state table information (partially or entirely provided through the next state palette) may be carried out while analysis of further characters 5 continues. Thus, it is clearly seen that the invention provides substantial acceleration of the parsing process through a small and economical amount of dedicated hardware. While the parser accelerator 10 can interrupt the CPU, the processing operation is entirely removed therefrom after the initial command to the parser accelerator. While the invention has been described in terms of a single preferred embodiment, those skilled in 15 the art will recognize that the invention can be practiced with modification within the spirit and scope of the appended claims.

Claims (5)

10. A method as recited in claim 9, including the further step of detecting sequences of strings of interest, and issuing a special token responsive to said step of detecting sequences for controlling processing.
11. A method as recited in claim 10, wherein a said sequence of strings of interest includes a nested string.
12. A method as recited in claim 10, wherein a said sequence of strings of interest correspond to words or phrases of text in a document.
13. A method as recited in claim 10, wherein said further processing performs blocking of a message.
14. A method as recited in claim 10, wherein said further processing performs content based routing.
AU2003277249A 2002-10-29 2003-10-03 Hardware parser accelerator Abandoned AU2003277249A1 (en)

Applications Claiming Priority (9)

Application Number Priority Date Filing Date Title
US42177402P 2002-10-29 2002-10-29
US42177502P 2002-10-29 2002-10-29
US42177302P 2002-10-29 2002-10-29
US60/421,775 2002-10-29
US60/421,773 2002-10-29
US60/421,774 2002-10-29
US10/331,315 2002-12-31
US10/331,315 US20040083466A1 (en) 2002-10-29 2002-12-31 Hardware parser accelerator
PCT/US2003/031314 WO2004040446A2 (en) 2002-10-29 2003-10-03 Hardware parser accelerator

Publications (1)

Publication Number Publication Date
AU2003277249A1 true AU2003277249A1 (en) 2004-05-25

Family

ID=32234359

Family Applications (1)

Application Number Title Priority Date Filing Date
AU2003277249A Abandoned AU2003277249A1 (en) 2002-10-29 2003-10-03 Hardware parser accelerator

Country Status (6)

Country Link
EP (1) EP1579320A2 (en)
JP (1) JP2006505043A (en)
KR (1) KR20050072128A (en)
AU (1) AU2003277249A1 (en)
CA (1) CA2504652A1 (en)
WO (1) WO2004040446A2 (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7596745B2 (en) 2005-11-14 2009-09-29 Sun Microsystems, Inc. Programmable hardware finite state machine for facilitating tokenization of an XML document
US7716577B2 (en) 2005-11-14 2010-05-11 Oracle America, Inc. Method and apparatus for hardware XML acceleration
US7665016B2 (en) 2005-11-14 2010-02-16 Sun Microsystems, Inc. Method and apparatus for virtualized XML parsing
US7665015B2 (en) 2005-11-14 2010-02-16 Sun Microsystems, Inc. Hardware unit for parsing an XML document
KR101231054B1 (en) * 2008-12-19 2013-02-07 한국전자통신연구원 Parser Accelerator having hardware engine and method

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS60105039A (en) * 1983-11-12 1985-06-10 Nippon Telegr & Teleph Corp <Ntt> Collation system of character string
JPH0746362B2 (en) * 1986-02-03 1995-05-17 日本電信電話株式会社 String matching method
JP2713962B2 (en) * 1988-03-16 1998-02-16 株式会社日立製作所 Character string search method and apparatus
US5193192A (en) * 1989-12-29 1993-03-09 Supercomputer Systems Limited Partnership Vectorized LR parsing of computer programs
JPH04348469A (en) * 1990-07-23 1992-12-03 Hitachi Ltd Character string retrieving device and its method
US5511213A (en) * 1992-05-08 1996-04-23 Correa; Nelson Associative memory processor architecture for the efficient execution of parsing algorithms for natural language processing and pattern recognition
US5414833A (en) * 1993-10-27 1995-05-09 International Business Machines Corporation Network security system and method using a parallel finite state machine adaptive active monitor and responder
US5995963A (en) * 1996-06-27 1999-11-30 Fujitsu Limited Apparatus and method of multi-string matching based on sparse state transition list
JP3062119B2 (en) * 1997-06-20 2000-07-10 日本電気アイシーマイコンシステム株式会社 Character string search table, method for creating the same, and character string search method
JP3368883B2 (en) * 2000-02-04 2003-01-20 インターナショナル・ビジネス・マシーンズ・コーポレーション Data compression device, database system, data communication system, data compression method, storage medium, and program transmission device

Also Published As

Publication number Publication date
WO2004040446A3 (en) 2004-09-02
JP2006505043A (en) 2006-02-09
EP1579320A2 (en) 2005-09-28
CA2504652A1 (en) 2004-05-13
KR20050072128A (en) 2005-07-08
WO2004040446A2 (en) 2004-05-13

Similar Documents

Publication Publication Date Title
US20040083466A1 (en) Hardware parser accelerator
AU2003277248B2 (en) Intrusion detection accelerator
US20070061884A1 (en) Intrusion detection accelerator
KR101648235B1 (en) Pattern-recognition processor with matching-data reporting module
US7080094B2 (en) Hardware accelerated validating parser
US7254632B2 (en) Apparatus and method for pattern matching in text based protocol
KR102004290B1 (en) Methods and apparatuses for reducing power consumption in a pattern recognition processor
US7644080B2 (en) Method and apparatus for managing multiple data flows in a content search system
US7529746B2 (en) Search circuit having individually selectable search engines
US7624105B2 (en) Search engine having multiple co-processors for performing inexact pattern search operations
JP4855400B2 (en) Method and system for multi-pattern search
US20060259508A1 (en) Method and apparatus for detecting semantic elements using a push down automaton
EP1604277A2 (en) Hardware accelerator personality compiler
AU2003277249A1 (en) Hardware parser accelerator
CN100430896C (en) Hardware parser accelerator
CA2504491A1 (en) Hardware accelerated validating parser
WO2008036501A2 (en) Regular expression searching of packet contents using dedicated search circuits
JP4061283B2 (en) Apparatus, method and program for converting lexical data to data
CN113868485A (en) Data loading method and device

Legal Events

Date Code Title Description
MK5 Application lapsed section 142(2)(e) - patent request and compl. specification not accepted