CA2393764A1 - Method and apparatus for longest match address lookup - Google Patents

Method and apparatus for longest match address lookup Download PDF

Info

Publication number
CA2393764A1
CA2393764A1 CA 2393764 CA2393764A CA2393764A1 CA 2393764 A1 CA2393764 A1 CA 2393764A1 CA 2393764 CA2393764 CA 2393764 CA 2393764 A CA2393764 A CA 2393764A CA 2393764 A1 CA2393764 A1 CA 2393764A1
Authority
CA
Canada
Prior art keywords
subtree
mapper
stored
entry
level
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
CA 2393764
Other languages
French (fr)
Inventor
David A. Brown
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Mosaid Technologies Inc
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Publication of CA2393764A1 publication Critical patent/CA2393764A1/en
Abandoned legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/28Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/74Address processing for routing
    • H04L45/742Route cache; Operation thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/901Indexing; Data structures therefor; Storage structures
    • G06F16/9017Indexing; Data structures therefor; Storage structures using directory or table look-up
    • G06F16/902Indexing; Data structures therefor; Storage structures using directory or table look-up using more than one table in sequence, i.e. systems with three or more layers
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11CSTATIC STORES
    • G11C15/00Digital stores in which information comprising one or more characteristic parts is written into the store and in which information is read-out by searching for one or more of these characteristic parts, i.e. associative or content-addressed stores
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/02Topology update or discovery
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/02Topology update or discovery
    • H04L45/04Interdomain routing, e.g. hierarchical routing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/54Organization of routing tables
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/74Address processing for routing
    • H04L45/745Address table lookup; Address filtering
    • H04L45/7453Address table lookup; Address filtering using hashing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/74Address processing for routing
    • H04L45/745Address table lookup; Address filtering
    • H04L45/74591Address table lookup; Address filtering using content-addressable memories [CAM]
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99931Database or file accessing
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99931Database or file accessing
    • Y10S707/99933Query processing, i.e. searching
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99941Database schema or data structure
    • Y10S707/99943Generating database or data structure, e.g. via user interface

Abstract

We present a method and apparatus for increasing the depth of a lookup table . The lookup table includes at least two lookup units. A search request for a final result value corresponding to the key is issued to each of the lookup units in parallel. Each of the lookup units performs a multi-level search in parallel for the final result value. The final result value stored in only o ne of the lookup units is provided by one of the lookup units.

Description

METHOD AND APPARATUS FOR LONGEST MATCH ADDRESS LOOKUP
BACKGROUND OF THE INVENTION
The Internet is a set of networks connected by routers. A muter maintains a routing table that indicates for each possible destination network, the next hop to which a received data packet should be forwarded. The next hop may be another router or the final destination.
An Internet Protocol ("IP") data packet received at a port in a router includes an IP destination address. The IP destination address is the final destination of the IP data packet. Currently there are two versions of IP, IP version 4 ("IPv4") and IP
version 6 ("IPv6"). IPv4 provides a 32-bit field in an IP header included in the data packeted for storing the IP destination address. The router forwards a received data packet connected to a next-loop router or the final destination if the destination is the local network, dependent on the IP destination address stored in the IP
header.
A 32-bit IPv4 destination address provides 4 billion possible routes. An Internet router typically stores 50,000 of the 4 billion possible routes.
However, the number of stored routes will increase with the growth of the Internet and the widespread use of IPv6.
Originally, the IP address space was divided into three classes of IP
addresses; A, B and C. Each IP address space was divided into a network address and a host address. Class A allowed for 126 networks and 16 million hosts per network. Class B allowed for 16382 networks with 64,000 hosts per network and class C allowed for 2 million networks with 256 hosts per network. However, dividing the IP address space into different classes reduced the number of available IP addresses. Class C only allowed a maximum of 256 hosts per network which is too small for most organizations. Therefore, most organizations were assigned a Class B address, taking up 64,000 host addresses which could not be used by other organizations even if they were not used by the organization to which they were assigned. Hosts in an organization with a Class B IP address all store the same network address in the 16 Most Significant Bits ("MBSs"), for example, 27.32.xx.xx.
Classless InterDomain Routing ("CIDR") was introduced to free up unused IP host addresses. The remaining unused networks are allocated to organization in variable sized blocks. An organization requiring 500 addresses gets 500 continuous addresses. For example, an organization can be assigned 500 available addresses starting at 128.32.xx. The number of routes stored by a router has increased since the introduction of Classless InterDomain Routing. Classless InterDomain Routing requires longest prefix matching to find the corresponding route instead of searching for a matching network address in order to find the corresponding next hop for the IP destination address. For example, a search can no longer stop after the 16 MSBs of a Class B IP address, for example, 128.xx.xx because 128.32.4.xx may be assigned to another organization requiring a different next hop.
One method for searching for a longest prefix match for a key is through the use of a binary tree search. A binary tree search matches a 32-bit input bit by bit down to 32 levels, requiring 32 searches to finding the entry matching the 32-bit key. Another method for searching for a match is through the use of a Patricia tree.
A Patricia tree reduces the number of searches required if there are no entries down a leaf of the binary tree.
Yet another method for efficiently searching for a next hop associated with an IP destination address is described in PCT application Serial Number PCT/SE98/00854 entitled "Method and System for Fast Routing Lookups" by Brodnick et al. filed on May 1 l, 1998. The method described by Brodnick reduces the number of next hops stored by not storing duplicate routes. By reducing the number of next hops, the memory requirement'is reduced so that a route lookup table can be stored in fast cache memory.
Brodnick et al. divides the binary tree into 3-levels. Dividing the binary tree into 3-levels reduces the number of searches to three. The indexed entry in the first level indicates whether the search can end at the first level with the route taken from the entry, or the search must continue to a subsequent level using a further portion of the IP destination address.
Fig. 1A illustrates a prior art 64K (65536) bit map representing the first level of a binary tree. A 64K bit map 30 represents the leaves or nodes 44 of the binary tree at depth 16, with one bit per node 44. The bit map is divided into bit-masks of length 16. There are 2'Z = 4096 bit masks in the 64k bit map. One bit mask is shown in Fig. 1A. A bit in the bit map 30 is set to '1' if there is a subtree or a route index stored in an array of pointers corresponding to the node 44. A bit in the bit.
map 30 is set to '0' if the node shares a route entry with a previous node 44.
Fig. 1B illustrates prior art lookup table implemented in cache memory. The lookup table includes an array of code words 36, an array of base indices 34 and a map table 40. A 32-bit IP address 38 is also shown in Fig. 1B. A codeword 46 is stored in the array of code words 36 for each bit mask in the bit map 30 (Fig.
1A).
The code word 46 includes a six-bit value 46a and a 10-bit offset 46b. A base index 42 is stored in the array of base indices 34 for every four code words 46 in the array of code words 36.
1 S The array of code words 36, array of base indices 34 and map table 40 are used to select a pointer in an array of pointers (not shown). The pointer stores a route index or an index to perform a further search.
A group of pointers in the array of pointers is selected by selecting a code word 46 in the array of code words 36 and a base index 42 in the array of base indices 34. The code word 46 is selected using the first 12 bits 50 of the IP
address 38. The base index 42 is selected using the first 10 bits 48 of the IP address 38. The correct pointer in the group of pointers is selected using the map table 32.
The 10-bit value 46b in the selected code word 36 is an index into the map table 32. The map table 32 maps bit numbers within a bit-mask to 4-bit offsets. The offset specifies the pointer within the selected group of pointers in the array of pointers. The 10-bit value 46b selects the row in the map table 32 and bits 19:16 of the IP address 52 selects the 4-bit offset 54.
Thus, a search for a pointer requires the following cache memory accesses:
(1) read a 16 bit code word 46; (2) read a 16-bit base address 42; (3) read a 4 bit offset 54 from the map table 32; (4) read a pointer at a pointer index where the pointer index is the sum of the base address 42, the code word offset 46a and the 4-bit offset 54.

The same memory accesses are required for each level of the binary tree.
Thus, a search of three levels requires 12 memory accesses.
SUMMARY OF THE INVENTION
Attorney Docket Number 2037.2004-001 entitled "Method and Apparatus S for Longest Match Address Lookup," by David A. Brown describes a lookup table for performing multiple level searches for a value corresponding to a key. The number of route indices which can be stored in the lookup table is limited by the number of memory locations in the lookup table.
In accordance with the principles of the invention, a lookup table including a first lookup unit and a second lookup unit is presented. The first lookup unit receives a key and provides a first result through a mufti-level search in the first lookup unit. The second lookup unit receives the key and provides a second result through a multi-level search in the second lookup unit. Both the first lookup unit and the second lookup unit receive the key in parallel and only one of the units provides a final result for the key stored in only one of the lookup units dependent on the first result and the second result.
The final results are distributed amongst the tables dependent on the total number of final results to be stored in the lookup table. The key may be an IP
address and the final result is a route index for the IP address.
A first memory location in the first lookup unit indexed by the key stores the final result and a second memory location in the second lookup unit indexed by the key stores an indication that the final result is stored the first lookup unit. Each of the first lookup unit and the second lookup unit includes a plurality of mapper levels and each of the mapper levels includes memory locations for storing results.
In one embodiment, the final result corresponding to the key stored in a location in any one of the plurality of mapper levels in the first lookup unit is forwarded to the second lookup unit and provided by the lookup unit.
In an alternative embodiment, the final result corresponding to the key is provided by the lookup unit in which the final result is stored.
BRIEF DESCRIPTION OF THE DRAWINGS
Fig. 1A illustrates a prior art bit map for representing the first level of a binary tree;
Fig. 1B illustrates a prior art lookup table implemented in cache memory;
Fig. 2A illustrates a longest match prefix lookup table according to the principles of the present invention;
Fig. 2B illustrates a binary tree representation of the route indexes stored in the lookup table shown in Fig. 2A;
Fig. 3 illustrates a longest match prefix lookup table for a 40-bit key according to the principles of the present invention;
Fig. 4 illustrates the types of mapper entries which can be stored in the direct mapped mapper shown in Fig. 2A;
Fig. 5 illustrates the mapper corresponding to the nodes in mapper level 2 112b shown in Fig. 2B;
Fig. 6A is a binary tree representation of a subtree;
Fig. 6B illustrates the dense subtree descriptor stored in the data field shown in Fig. 5 corresponding to the nodes in the bottom level of the subtree shown in Fig.
6A;
Fig. 7 illustrates the ptr field shown in Fig. 5;
Fig. 8 illustrates the mapper address logic shown in Fig. S;
Fig. 9 is a flowchart illustrating the steps for searching for the longest match;
Fig. 10A illustrates an embodiment for providing depth expansion;
Fig. l OB illustrates one of lookup tables in the embodiment shown in Fig.
1 OA;
Fig. l OC illustrates another embodiment for providing depth expansion to increase the number of mapper entries available for storing values;
Fig. l OD illustrates the slave lookup table in the embodiment shown in Fig.
IOC;
Figs. 1 lA-B illustrate a binary tree representation of the distribution of the nodes in the binary tree representation of route indices shown in Fig. 2B
amongst the mapper entries in the lookup tables shown in Figs. 10A and 10C;
Fig. 12 is a flowchart illustrating a method for distributing values amongst the mapper entries in the lookup tables shown in Figs. 10A and l OC;
Fig. 13 is a flowchart illustrating a method for searching for a value corresponding to a search key stored in one of mapper entries in the lookup tables shown in Fig. l OC;
Fig. 14 is a flowchart illustrating a method for searching for a value corresponding to a search key stored in one of the mapper entries in the lookup tables shown in Fig. 10A;
Fig. 15 illustrates a binary tree representation of a sparse subtree and a dense subtree in the second mapper level indexed by the first mapper level;
Fig. 16A-C illustrate the modification to the data field and the pointer field in the subtree entry shown in Fig. 5 and the subtree entry shown in Fig 4 to permit storage of a plurality of sparse subtree descriptors in the subtree entry;
Fig. 17 illustrates the sparse mode subtree logic in the offset logic shown in Fig. 8 to select a block offset for a node in a sparsely populated subtree;
Fig. 18 illustrates the sparse mode logic shown in the offset logic in Fig.
17;
20. Fig. 19A-D illustrate the selection of a block offset for a node in a sparsely populated subtree;
Fig. 20 is a block diagram illustrating the sparse mode base select logic in the pointer logic shown in Fig. 8;
Fig. 21 illustrates a dense subtree descriptor and a sparse subtree descriptor stored in the subtree memory;
Fig. 22 is a flow chart illustrating a method for providing a mapper address for a mapper entry in a subtree mapper storing a route for a node in a sparsely populated subtree and a densely populated subtree;
Fig. 23 illustrates a binary tree representation of a new route to be added to the lookup table;
Fig. 24 illustrates update routes stored in a processor memory;
Fig. 25 illustrates the new route shown in Fig. 23 stored in the lookup table;

_'7_ hig. 26 is a flowchart tlluslratlog the steps for adding the new route to the lookup table shown in 1~ig. 25, The foregoing and other objects, features and advantages of the invention will be apparent from the followinb more particular description of preferred embodiments of the invention, as illustrated. in the accompanying drawings in which like reference characters refer to the same parts throughout the di ffcrcnt views. The drawings are not necessarily to scale, emphasis instead being placed upon illustrating the principles of the invention.
Dr,TAT(,F.D AESCR(fTlON OF'rHL INVL?N'fION
A description of preferred etnbodimenls of the invention follows, Dcnsc Modc hig. 2A illustrates a longest match prefix lookup table 100 according to the principles of the present invention. The lookup table 100 provides a route index 102 for a key 104. The route index 10? is used to access the next hop for the TP
(1C51111at1011 it(lClrt~SS. (n the emhodimcnt showmin I~ig. 2A the key 104 is 32 bits wide;, however, the key 104 is not limited to 32 bits. The lookup table 100 includes three mappcrs IOGa~c. Each mapper lOGa-c includes a separately addressed memory. 'fhe route index 102 or a default route index cowespontling to the ke;y 104 is stored in a localioo in one of the mappers IOGa-c. 'l he mapper outputs 11 Oa-c from each mappcr arc stored in delay memory 150a-c until all mappcrs IOGa-c have been accessed for the key, if nmltiple mappers are required to he searched.
The multiplcxcr 108 selects one of the mappcr outputs 1 l0a-c forwarded to the inputs of the multiplexer 108 as the route index 102. The mapper output 1 l0a-c is selectt:il dependent on the Most Significant bit ("MSB") of the roappcr output 11 Oa-c. The MSB ofthe mappcr output 110a-c is'1' only if the a~~apper output 11 Oa-c includes a route index 102, hi g. 2R illustrates a binary tree repr~scntation of the entries stored in the u~ahpers 1 OGa-c in the lookup table 100 shown in Fi g. 2A. Fig, 2R is described in conjunction with FiS. 2A. The 32-b'it key 104 can be represented as a 32-level binary tree_ A binary tree impletncntation requires 32 starches to search bit by bit _$-down to 32 levels. To reduce the number of searches, the 32 levels of the binary tree are divided into three mapper levels 112a-c with each mapper level 112a-c corresponding to a mapper 106a-c (Fig. 2A). Mapper level_1 112a includes the first 16 of the 32 levels of the binary tree. However, for simplicity only 5 of the 16 levels S are shown in Fig. 2B. Mapper level 2 112b includes the next 8 levels of the level binary tree, with three of the eight levels shown in Fig. 2B. Mapper level 3 includes the last 8 levels of the 32-level binary tree, with three of the eight levels shown in Fig. 1B. Each mapper level 112a-c includes a plurality of nodes.
Dividing the 32-levels such that 16-levels ( 16 MSBs of the key 104) are in mapper level-1 112a, 8-levels in mapper level 2 112b and 8-levels in mapper level 3 appears to be optimal in the current memory technology; however, the invention is not limited to this configuration.
Instead of performing 16 separate bit by bit searches for the first 16 bits of the key 104, the route indices 102 associated with the first 16-bits of the key 104 are stored in mapper 106a (Fig. 2A). The mapper 106a (Fig. 2A) is directly indexed with the first 16-bits MSBs of the key 104. The next mapper 106b is searched dependent on whether the previous mapper 106a stores the route index 102 used to access the next hop information associated with the key 104.
As shown in Fig. 2B, the nodes or leaves shown in mapper level-1 112a include two routes 114, 116 labeled r0 and r1 respectively and two pointers to mapper level 2 112b 1304 and 13023 labeled s0 and s1, respectively. A route index 102 for each route 114, 116 is stored in the L1 mapper 106a. Also, an address pointer 120 for L2 mapper 106b is stored for subtree index 1304 and an address pointer (not shown) for L2 mapper 106b is stored for subtree 13023. An address pointer 120 stored for subtree index 1304, in mapper entry 1404 in mapper 106a indicates that a search of the next level is required in order to find a route index 102 associated with the key 104.
The value of any node in the tree can be determined by tracing a path from the root 114. Each node in the binary tree is shown with two children, a right child and a left child. The right child is chosen if the parent node is ' 1.' The left child is chosen if the parent node is '0'. Tracing the path from the root 114 to node 116, r1 is stored as the route index 102 in the L1 mapper 106a for all keys with MSBs set to '010' . Tracing the path from the root node 114 to s0 node 1304, s0 is stored in the L1 mapper 106a for all keys with MSBs set to '00011'.
The L1 mapper 106a is a direct mapped mapper and stores a route index 102 for each bottom-level node or leaf in the bottom level of mapper level_1 112a.
The bottom level of mapper level_1 112a is the sixteenth level of the 32-level binary tree. The sixteenth level has 64K nodes. However, for illustrative purposes, the bottom level of mapper level-1 112a is shown as level-5 of the 32-level binary tree.
The route indices 102 shown in the L1 mapper 106a correspond to level-5 130'-13032 nodes of mapper level_1 112a. Tracing the path from the root node 114 to level 5 nodes 130', 1302, 130' the route index 102 is r0. Thus r0 is stored in locations 140' ,1402, 1403 of L1 mapper 106a; that is, at index 00000, 00001, and 00010. Node 1304 stores a subtree index s0, thus s0 is stored in location 1404 in the L1 mapper 106a at address 00011. Similarly the route index 102 for level-5 nodes 1305-1308 is r0 thus, locations 1405,1406,140' ,1408 at addresses 00100, 00101, 00110, and 00111 in the Ll mapper 106a store r0. The route index 102 for level-nodes 1309-130'2 is r1, thus locations 1409,140'° at addresses 01000 and 010001 in the L1 mapper 106a store r1.
Each location in the L1 mapper 106a stores a route index 102 assigned to the level-5 node 300'-30032 directly or through a parent of the level-5 node 300'-3z or an address pointer to the next mapper 106b-c. Mapper level 3 106c includes two host nodes h0 at node 138 and hl at node 140 at the bottom level of the 32-level binary tree. A search for a host node requires a search of all bits of the key 104.
As shown in Fig. 2A the route index 102 for h0 is stored in location 1464 in L3 mapper 106c.
Unlike the L1 mapper 106a, the L2 mapper 106b and L3 mapper 106c are not directly mapped.
In mappers 106b, and 106c, a route index 102 is not stored for each possible input. A route index 102 is stored only if the route index 102 for the node differs from the previously stored route index 102 in the mapper 106b-c. Looking at the level-3 nodes in the first subtree A shown in mapper level 2 112b, the route index 102 for node 132' and node 1322 is r0, therefore a route index for r0 is stored in location 142' for both node 132' and node 1322 in the L2 mapper 106b. A
subtree index s0 for node 1322 is stored in location 1422. The route index 102 associated with level-3 node 1324 and level-3 nodes 1325 and 1326 is r0, which is different from s0 stored for the previous node 1322, thus r0 is stored in the next location 1423 in L2 mapper 106b. Route r2 is stored in location 142' in L2 mapper 106b for node 132' because node 132' does not share the same route as the previous node 1326.
S Subtree index s3 is stored for the next level-3 node 132', thus s3 is stored in location 1425 in the L2 mapper 106b. By storing a route index only when there is a route change from the previous node, the memory required for storing route indices 102 is reduced. As shown, only five locations are required in the L2 mapper 106b for storing route indices for the eight level-3 nodes 132'-$ in the first subtree A in mapper level 2 112b. The non-direct mapper 106b, 106c is described in more detail later in conjunction with Fig. S.
Fig. 3 illustrates a longest match prefix lookup table 200 for a 40-bit key according to the principles of the present invention. In one embodiment the 40-bit key includes an 8-bit prefix and a 32-bit IP address. The 8-bit prefix can be a Virtual Private Network ("VPN") identifier associated with the 32-bit IP
address.
The lookup table 200 includes four mappers 106a-d. Mapper 106a is a direct mapped mapper as described in conjunction with Fig. 2A. Mappers 106b-d are indirect mappers. Mapper 106a stores a route index 102 or a subtree index for the L2 mapper 106b corresponding to the 16 MSBs of the 40-bit key 210. Thus, the mapper has 64K possible locations, one for each of the 64K nodes in the first mapper level 112a (Fig. 2B). The L1 mapper entry data 220a stored at the corresponding location in the L1 mapper 106a is forwarded to a pipeline 208 and the L2 indirect mapper 106b. If the Ll mapper entry data 220a indicates that a search of the next level is required using the next eight bits of the key 210b, a search is performed in the L2 indirect mapper 106b dependent on the next eight bits of the key 210b, and the L1 mapper entry data 220a.
The result of the second level search is provided on L2 mapper entry data 220b which is forwarded to the pipeline 208 and the L3 indirect mapper 106c. A
third level search is performed in the L3 indirect mapper 106c dependent on the next eight bits of the key 210c and the L2 mapper entry data 220b.
The result of the search is provided on L3 mapper entry data 220c to the pipeline 208 and to the L4 indirect mapper 106d. The L3 mapper entry data 220c determines if a fourth search is performed in the L4 indirect mapper 106d dependent on the last eight bits of the key 210d and the L3 mapper entry data 220c.
The result of the fourth search is provided on L4 mapper entry data 220d.
The route index 102 associated with the longest match prefix for key 210 is stored in only one location in one of the mappers 106a-d. Thus, the route index 102 forwarded to the pipeline 208 is included in only one mapper entry data 220a-d. If the route index 102 is found in one of the mappers 106a-d, for example, mapper 106b a search of the remaining mappers 106c-d is not necessary and mappers 106c-d are not accessed. The pipeline 208 includes a multiplexer 108 (Fig. 2A) for selecting the route index 102 included in one of the mapper entry data 220a-d.
For example, the MSB of the mapper entry data 220a-d can provide an indication of whether a route index is included.
By using a pipeline 208 in conjunction with the mappers 106a-d, multiple searches of the longest match prefix table 200 with different keys 210 can be performed in parallel. The pipeline 208 allows multiple searches of the 40-bit lookup table 200 to take place in parallel by storing the mapper entry data 220a-d for each mapper 106a-d associated with the 40-bit key 210 until a search of each of the other mappers 106x-d has been completed, if required, to find route index corresponding to the 40-bit key 210. Thus, a search request for a route index corresponding to a received IP address is issued to the lookup table 200 by performing a single memory access to the direct mapped mapper 106a. A
subsequent search for a route index corresponding to another key can be issued to the lookup table 200 in the next memory access cycle for the direct mapped mapper 106a.
Fig. 4 illustrates the types of mapper entries which can be stored in the direct mapped mapper 106a shown in Fig. 3. A mapper entry for any node in the binary tree shown in Fig. 2B can store, a no-entry 300, a route entry 302 or a subtree entry descriptor 304. Each type of mapper entry 300, 302, 304 includes a subtree flag 306. The state of the subtree flag 306 indicates whether the mapper entry is a subtree entry descriptor 304. If the subtree flag 306 is set to'1', the mapper entry is a subtree entry descriptor 304 and includes a subtree index 312. The subtree index 312 is the address of the subtree entry descriptor 304 stored in the next non-direct mapped mapper 106b-d. The subtree entry is described later in conjunction with Fig. 4. If the subtree flag 306 is '0', the no-entry flag 314 is checked to determine if the mapper entry is a no-entry 300 or a route entry 302. If the no-entry flag 314 is '0', the entry is a no-entry 300. If the no-entry flag 314 is'1', the entry is a route entry 302 and stores the route index 102 (Fig. 3) associated with the key 104 in the route index field 310. The subtree flag 306 is used by the multiplexer 108 (Fig. 2A) to select the mapper entry data 220a-d including the route index 102 (Fig. 3).
Fig. 5 illustrates the mapper 106b corresponding to the nodes in mapper level 2 112b shown in Fig. 2B. The mapper 106b includes a subtree memory 400, mapper address logic 402 and a subtree mapper 418. The subtree index 312 selected by the first portion of the key 210a stored in mapper 106a is forwarded to the subtree memory 400. The subtree memory 400 includes a subtree entry 404 selected by the subtree index 312. The subtree entry 404 includes a data field 406 and a pointers field 408.
Returning to Fig. 2B, the subtree entry 404 corresponds to the bottom level of one of the subtrees shown in mapper level 2 112b. If mapper level 2 112b has eight levels, the bottom level of each subtree (not shown) has a maximum of routes, one for each node.
Continuing with Fig. 5, the subtree entry 404 provides access to 256 possible route indices 102 (Fig. 3) corresponding to each node on the bottom level of the subtree. The route indices 102 (Fig. 3) are stored in the subtree mapper 418.
To provide access to 256 possible route indices, a dense subtree descriptor is stored in the data field 406. The data field 406 is 256 bits wide, providing one bit for each node at the bottom level of the subtree. The data field 406 is described in more detail later in conjunction with Fig. 6A and Fig. 6B. The pointers field 408 is 256 bits wide to allow for the storage of sixteen 16-bit pointers, with each pointer storing the base address for 16 contiguous mapper entries in the subtree mapper 418 to provide access to the 256 route indices. Thus, the pointers field 408 can indirectly provide. a pointer to a mapper entry in the subtree mapper 418 for each node in the bottom level of the subtree. The pointers field 408 is described in more detail in conjunction with Fig. 6.

The subtree data 412 stored in the dense subtree descriptor in the data field 406 and the subtree pointer 414 stored in the pointers field 408 are forwarded to the mapper address logic 402. The mapper address logic 402 also receives the next portion of the key 210b (the next eight bits).
The mapper address logic 402 determines the mapper address 416 of the mapper entry associated with the node in the bottom level of the subtree dependent on the next eight bits of the key 212b, the subtree data 412 and the subtree pointers 414 associated with the subtree. The mapper address 416 selects the mapper entry in the subtree mapper 418. The subtree mapper 418 includes the same types of mapper entries as described in conjunction with Fig. 4 for the direct mapped mapper 106a.
The contents of mapper data entry 220b determines whether a subsequent search is required. A subsequent search is required if the mapper entry data 220b includes a subtree index 312 (Fig. 4) indicating that there is another subtree entry 404 in the next mapper level 112c (Fig. 2B).
The second portion of the key 210b selects the node in the bottom level of the selected subtree. The subtree pointers 414 selects the base address associated with the node in the subtree and the subtree data 412 selects the offset within the block of mapper entries associated with the base address. The mapper address logic 402 is described later in conjunction with Fig. 7.
Fig. 6A is a binary tree representation of a subtree. The subtree shown includes 5-levels. The subtree includes three route indices r1, r2, and r3 and two subtree indices s0 and s1. There are 32 nodes 500'-50032 on the bottom level of the subtree. The route index or subtree index associated with each node 500'-50032 in the bottom level is shown in Table 1 below.

Subtree Bits Route/Subtree 00000 r0 00001 r0 00010 r0 00011 r0 00100 r1 00101 r1 00110 r0 00111 r0 01000 r2 01001 s0 01010 r2 01011 r2 01100 r2 Ol 110 r2 01111 r2 1 xxxx r3 Table 1 Fig. 6B illustrates the dense subtree descriptor stored the data field 406 shown in Fig. 5 corresponding to the nodes in the bottom level of the subtree shown in Fig. 6A. The data field 406 includes 32 bits, one bit for each node 500 in the bottom level of the subtree shown in Fig. 6A. The bits 502'-50232 in the data field 406 are assigned as follows: a bit in the data field 406 is set to '0' if the route index for the previous node is to be used and set to '1' to increment to the next mapper entry address if the next route index stored in the subtree mapper 418 (Fig.
5) is to be used. The first bit in the data field 402 selects the default route r0 stored in mapper entry 504' unless a route is specified. Thus, bit 502' is set to '0' to select the default route because there is no route specified. The default route r0 stored in mapper entry 504' is selected for the next three nodes 5002 - 500°, therefore, corresponding bits 5022-5024, are set to '0' in the data field 406 to use the previous route index used by 502'. There is a route change at node 5005.
The route r1 used for node 5005 stored in mapper entry 5042 is shared by node 5006. Therefore bit 5025 is ' 1' indicating a route change to select the mapper entry 5042 in the subtree mapper 418 (Fig. 5). Bit 5026 is '0' indicating that the route index stored in 5025 is to be used for this node. No route is provided for node 500', therefore there is a route change and a'1' is stored in bit 502'requiring a mapper entry 504' in the subtree mapper 418 (Fig. 5) storing the default route r0.
Node 5008 shares the same route as the previous node 500', a new mapper entry is not required in the subtree mapper 418 (Fig. 5). Bit 5028 corresponding to node 5008 is set to '0'. Node 5009 has a different route from the previous node 5008, a new mapper entry is required in the subtree mapper 418 (Fig. 5). Bit 5029 corresponding to node 5009 is set to '1' and mapper entry 5044 storing r2 is added to the subtree mapper 418 (Fig. 5) in the next contiguous memory location.
Node 500'° has a different route from the previous node 5009, a new route entry is required in the subtree mapper 418 (Fig. 5). Bit 502'°
corresponding to node 500'° is set to ' 1' and mapper entry 5045 storing s0 is added to the subtree mapper 418 (Fig. 5) in the next contiguous memory location.
Node 500" has a different route from the previous node 500'°, a new mapper entry is required in the subtree mapper 418 (Fig. 5). Bit 502" corresponding to node 500" is set to '1' and mapper entry 5046 storing r2 is added to the subtree mapper 418 (Fig. 5) in the next contiguous memory location.
Nodes 50012 and 500'3 share the same route as previous node 500", a new mapper entry is not required in the subtree mapper 418 (Fig. 5). Bit 50212 corresponding to node 500'2 and bit 502'3 corresponding to node 500'3 are set to'0' in the data field 406.
Node 500'4 has a different route from the previous node 500'3, a new mapper entry is required in the subtree mapper 418 (Fig. 5). Bit 50214 in data field corresponding to node 50014 is set to '1' and mapper entry 504' storing s1 is added to the subtree mapper 418 (Fig. 5). Node 50015 has a different route from the previous node 50014, a new mapper entry is required in the subtree mapper 418 (Fig. 5).
Bit 502'5 in the data field corresponding to node 50015 is set to ' 1' and mapper entry 5048 storing r2 is added to the subtree mapper 418 (Fig. 5). Node 500'6 shares the same route as the previous node 500'5, a new mapper entry is not required in the subtree mapper 418 (Fig. 5).

Node 500" has a different route from the previous node 500'6, a new mapper entry is required in the subtree mapper 418 (Fig. 5). Bit 502" in the.data field 406 corresponding to node 500" is set to '1' and mapper entry 5049 storing r3 is added to the subtree mapper 418 (Fig. 5).
S Nodes 500'8-50032 all share the same route as node 500", a new mapper entry is not required in the subtree mapper 418 (Fig. 5). Corresponding bits 502'8-50232 are set to '0'. Thus nine mapper entries 504'-9 are required to store route entries 302 (Fig. 4)or subtree entries 304 (Fig. 4) for the 32 nodes S00'-50032.
A mapper entry 504'-5049 corresponding to a node 500'-50032 is indexed in the subtree mapper 418 (Fig. 5) by computing the number of'1"s stored in the dense subtree descriptor stored in the data field 406. For example, to find the mapper entry 504'-5049 corresponding to node SOOZB, the number of'1's stored in bits 502'-of the data field 406 are counted. The number of'1's is 8, and the corresponding mapper entry is the eighth location from the default route; that is, mapper entry 5049.
Storing mapper entries only when there is a route change reduces the number of mapper entries 504'-5049 per subtree in the subtree mapper 418 (Fig. 5).
Fig. 7 illustrates the pointers field 408 shown in Fig. 5. The pointers field 408 includes block base address fields 600', 6002 for storing base addresses for blocks of 16 contiguous mapper entry locations 504'-504'6 (Fig. 6B) in the subtree mapper 418 (Fig. 5). Memory is allocated in the subtree mapper 418 (Fig. 5) in blocks 602', 6022 of 16 contiguous mapper entries. An 8-level subtree can have up to 256 different routes requiring 16 blocks 602', 6022 in order to store all 256 routes.
The number of blocks 602 required is dependent on the number of routes for the subtree. A block 602 is allocated to a particular subtree by removing the block base address 602', 6022 from a free list of block base addresses (not shown).
Methods for providing a free list of addresses for a memory are well-known in the art.
By allocating memory blocks of 16 mapper entries 504'''6, the memory in the subtree mapper 418 (Fig. 5) is easier to manage because the allocated 16 locations are contiguous.
Fig. 8 illustrates the mapper address logic 402 shown in Fig. 5. The mapper address logic 402 includes offset logic 700, pointer logic 702 and adder logic 704.

The offset logic 700 includes node select logic 706 and ones count logic 708.
The pointer logic includes base address select logic 710.
The node select logic 706 selects the node 500 (Fig. 6B) in the subtree data 412 corresponding to the eight bits of the key 210b. The corresponding node number is forwarded on node select 718 to the ones count logic 708. The ones count logic 708 counts the number of '1's stored in the subtree data field 406 up to the bit corresponding to the selected node 500. The number of ones is forwarded on block select 712 to the pointer logic 702 and block offset 714 to the adder logic 704.
There can be a maximum of 256 ' 1's stored in a 256 bit subtree data field 406 requiring an 8-bit wide count field. The 8-bit count field is divided into two fields, with the 4 MSBs providing the block select 712 and the 4 Least Significant Bits ("LSBs") providing the block offset 714.
For example, if the eight bit key 210b is '0100 0100', to select node number 68 and there are 27 ' 1's stored in the first 68 bits of subtree data 412, the count is 1 C
Hex (0001 1100), the MSBs (0001); that is, block select 714, select block 602' (Fig.
6) and the LSBs (1100); that is, base block offset select mapper entry 504"
(Fig. 6), that is, the twelfth entry in block 502'.
The base address select logic 710 selects the base address 716 from the subtree pointer 414 dependent on the block select 712 forwarded from the offset logic 700. The adder logic 704 adds the block offset 714 forwarded from the offset logic 700 to the base address 716 and provides the mapper address 416. The mapper address 416 is the index of the mapper entry 504 (Fig. 6B) in the mapper 106b-d.
Fig. 9 is a flowchart illustrating the steps for searching for the longest match prefix for a key 210 (Fig. 3) in lookup table 200 according to the principles of the present invention.
At step 800, the first portion of the key 210a (Fig. 3) is forwarded as the index to mapper 106a. Processing continues with step 802.
At step 802, the mapper entry data 220a (Fig. 3) stored in the mapper entry 504 (Fig. 6B) in the first level mapper indexed by the first portion of the key 210a (Fig. 3) determines whether a further search of the next level is required. If so, processing continues with step 804. If not, the route entry 302 (Fig. 4)in the indexed mapper entry 504 (Fig. 6B) in the first level mapper stores the corresponding longest prefix route for the key and processing continues with step 808.
At step 804, the next level mapper 106b-d is searched. The index for the next level mapper is dependent on the subtree index 312 stored in the subtree entry descriptor304 (Fig. 4) in the indexed mapper entry 504 (Fig. 6B) in the previous level mapper and the next portion of the key 210b-d. Processing continues with step 806.
At step 806, the indexed mapper entry 504 (Fig. 6B) in the next level mapper stores the corresponding longest prefix route index for the key or a subtree index indicating a further search is required. If a further search is required, processing continues with step 804. If not, processing continues with step 808.
At step 808, the route index 102 (Fig. 3) stored in a mapper entry 504 (Fig.
6B) in one of the mappers 106a-d is forwarded from the lookup table 200 as the route index 102 (Fig. 3). Processing is complete.
Depth Expansion The number of route indices 102 (Fig. 3) which can be stored in the lookup table 200 shown in Fig. 3 is limited by the number of available mapper entries (Fig. 6B) in the subtree mapper 418 (Fig. 5). For example, if each subtree mapper 418 (Fig. 5) includes 128K mapper entries and there are two subtree mappers (Fig. 5) in the lookup table, a maximum of 256K route indices 102 (Fig. 3) can be stored in the lookup table 200. A subtree mapper 418 (Fig. 5) with 128K mapper entries requires a 17-bit index. A subtree mapper 418 (Fig. 5) with 512K
mapper entries requires a 19-bit index. Two S 12K subtree mappers 418 (Fig. S) in the lookup table 200 provides storage for 1 million out of a possible 4 billion route indices 102 (Fig. 3) for a 32-bit IPv4 destination address.
The number of mapper entries 504 (Fig. 6B) for storing route indices 102 (Fig. 3) can be increased by providing a plurality of lookup tables 200. The plurality of lookup tables are searched in parallel for the value corresponding to a search key 210 stored in a mapper entry 504 (Fig. 6B) in subtree mapper 418 (Fig. 5) in one of the lookup tables 200.

Fig. 10A illustrates an embodiment for depth expansion. Two lookup tables are shown, a master lookup table 200a and a slave lookup table 200b. However, the number of lookup tables is not limited to the two shown, more than one slave lookup table 200b can be added.
Each of the lookup tables 200a-b is searched in parallel with the same search key 210. The route index 102 (Fig. 3) corresponding to the search key 210 is stored in a subtree mapper 418 (Fig. 5) in one of the lookup tables 200a-b or in direct mapper 106a in both of the lookup tables 200a, 200b. The final route index 900 is found after a search in parallel of both lookup tables 200a, 200b.
Fig. l OB illustrates one of the lookup tables 200a in the embodiment shown in Fig. 10A. Each of the lookup tables 200a-b includes mappers 106a-d and a pipeline 208 as has already been described for lookup table 200 in conjunction with Fig. 3, and driver logic 902. The lookup table 200a performs a mufti-level search in the mappers 106a-d for a route index corresponding to the search key. The result of each level search is forwarded on mapper entry data 220a-d to the pipeline 208. The pipeline 208 forwards the result of the search 904 to driver logic 902. Mapper entries 504 (Fig. 6B) in mapper 106a in each of the lookup tables 200a-b store subtree entries 304 (Fig. 4) but route entries 302 (Fig. 4) are only stored in the mapper 106a in master lookup table 200a. No-entry 300 (Fig. 4) is stored in mapper entries 504 (Fig. 6B) in mapper 106a in slave lookup table 200b instead of route entries 302. Populating only one lookup table with the route indices in mapper 106a avoids selecting a lookup table to provide the final route index 900. This results in 64K of memory in the slave lookup table 200b which can not be used to store route indices but allows the same lookup table to be configured as a master lookup table or a slave lookup table as described in conjunction with Fig. 3. In an alternative embodiment, a slave lookup device without mapper 106a can be provided. A
search ends in mapper 106a in the master lookup table 200a if a route index 102 (Fig.
3) is stored in a route entry 302 (Fig. 4) in mapper 106a.
As shown in Fig. 10A, master lookup table 200a and slave lookup table 200b share the final route index 900. The lookup table 200a, 200b in which the final route index 900 is stored provides the route index 102 (Fig. 3). If each of the lookup tables 200a, 200b is a separate device, sharing the final route index 900 reduces each device's external pin count. Only one of the lookup tables 200a, b drives the final route index 900 at any time.
To avoid an error condition in which the route index associated with the search key 210, is stored in both lookup tables 200a, 200b where both lookup tables would drive the final route index 900 at the same time, each lookup table 200a,b stores a device code 906. A 3-bit device code allows the expanded lookup table to include 8 devices.
The driver logic 902 determines if the result of the search 904 includes a route index 102 (Fig. 3). If so, the driver logic 902 in the lookup table 200a, signals an intention to drive the final route index 900 on a bus request signal (not shown).
If two or more lookup tables 200a, 200b signal the intention to drive the route index signals at the same time, the route index is provided by the lookup table 200a, 200b with the lowest device code. Methods for resolving bus conflicts through the use of a bus request signal are well-known in the art.
Fig. l OC illustrates another embodiment for providing depth expansion to increase the number of mapper entries available for storing a value corresponding to a search key 210. In the embodiment shown in Fig. 10C, two lookup tables 200c-d are provided for storing values, a master lookup table 200c and a slave lookup table 200d. However, the number of lookup tables is not limited to the two shown, the number of mapper entries can be increased by adding more slave lookup tables 200d. A search for the value stored in a mapper entry in one of the lookup tables 200c-d corresponding to the search key [39:0] 210 is performed in parallel in the lookup tables 200c-d.
Fig. lOD illustrates the slave lookup table 200d in the embodiment shown in Fig. 10C. Each lookup table includes mappers 106a-d as described for lookup table 200 in conjunction with Fig. 3. Mapper entries in mapper 106a in each of the lookup tables 200c-d store subtree entries 304 (Fig. 4). Each lookup table 200c-d forwards a subtree index 312 stored in a subtree entry descriptor 304 (Fig. 4) stored in a mapper entry 504 (Fig. 6B) in mapper 106a to the next mapper 106b-d.
However, a route index 102 (Fig. 3) is only stored in mapper 106a in a master lookup table 200c. A no-entry is stored in mapper 106a in a slave lookup table 200d to avoid storing a route index corresponding to the key in more than one lookup table 200b, 200d.
The result of the mufti-level search 904 in the slave lookup 200d is forwarded to the final index logic 1004. The final index logic 1004 forwards the result of the mufti-level search 904 or the incoming result 1000a forwarded from the master lookup table 200c as the outgoing result 1002x. If a route index 102 (Fig. 3) is included in the result of the mufti-level search 904, the result of the mufti-level search is forwarded as the outgoing result 1002x. If the route index 102 (Fig.
3) is included in the incoming result 1000x, the incoming result 1000a is forwarded as the outgoing result 1002x. If the route index 102 (Fig. 3) is included in neither the incoming result 1000a or the result of the mufti-level search 904, the result of the mufti-level search 904 is forwarded as the outgoing result 1002x.
As shown in Fig. l OC, master lookup table 200c and slave lookup table 200d are connected through a common bus labeled incoming result 1000x. The route index 102 (Fig. 3) is only forwarded from slave lookup table 200d on outgoing result 1002x. If there is more than one slave lookup table 200d, the route index 102 (Fig. 3) for the expanded lookup table is provided by the last slave lookup table.
This embodiment avoids the implementation of mufti-driver final route index described in conjunction with Fig. l0A but requires more device external pins for incoming result 1000x.
Figs. 1 lA-B illustrate a binary tree representation of the distribution of the routes shown in Fig. 2B among the lookup tables 200x-b (Fig. l0A) or 200c-d (Fig.
lOC).
Fig. 11A illustrates a binary tree representation of the routes stored in master lookup table 200a (Fig. l0A) or 200c (Fig. lOC). Subtree B shown in the binary tree representation of the routes shown in Fig. 2B is not included in master lookup table 200x. Nodes 130'-13022 and 1302x-3z are coded in mapper 106a in lookup table 200a as described in conjunction with Fig. 3. The node at which subtree B would be indexed, if stored in the master lookup table 200x, is graphically represented with an X, indicating a pruned subtree. The mapper entry 504 (Fig. 6B) corresponding to node 13023 in master lookup table 200a no longer stores a subtree index 312 (Fig. 4) to subtree B. Instead no- entry 300 (Fig. 4) is stored in the mapper entry 504 (Fig.

6B) corresponding to node 130Z' in master lookup table 200a indicating that the mapper entry corresponding to node 13023 is stored in a subtree mapper 418 (Fig. 5) in another slave lookup table 200b.
Fig. 11B illustrates a binary tree representation of the routes stored in mapper entries 504 (Fig. 6B) in the subtree mappers 418 (Fig. 5) in slave lookup table 200b (FIG. 10A) or 200d (Fig. l OC). The binary tree representation of the routes stored in slave lookup table 200b differs from the binary tree representation shown in Fig. 2B
in that subtree A is not included. Thus, nodes 130'-1303 and 1305-13032 are coded as described in conjunction with Fig. 2B. The mapper entry 504 (Fig. 6B) corresponding to node 1304 in slave lookup table 200b no longer stores a subtree index 312 (Fig. 4) to subtree A. Instead, the mapper entry corresponding to node 1304 in slave lookup table 200b stores no-entry 300 (Fig. 4) indicating that the mapper entry corresponding to node 1304 is stored in another lookup table. The subtree index for subtree A and thus the route index for host 138 (Fig. 11A) is stored in master lookup table 200a, and the subtree index for subtree B and thus the route index for host 140 is stored in slave lookup table 200b. The slave lookup table 200b, 200d only stores results for a subtree; that is, a slave lookup table 200b, 200d does not store a result in the first level mapper 106a.
Refernng to Fig. 1 1A and Fig. 11B, a search for any of the nodes 1309-130'Z
in master mapper level_1 1102 (Fig. 3)a or slave mapper level_1 1104a with a first portion of the key 210a results in r1 116 stored in a route entry 302 (Fig. 4) in a mapper entry 504 (Fig. 6B) in mapper 106a in the master lookup table 200a, 200c and a no-entry 300 (Fig. 4) stored in a mapper entry 504 (Fig. 6B) in mapper 106a in the slave lookup table 204b, 200d. The route entry 302 (Fig. 4) stored in the master lookup table 200a, 200c is forwarded on incoming result 1000a to the slave lookup table 200b, 200d and forwarded by the slave lookup table 200b, 200d on outgoing result 1002a.
A search for node 1304 with a first portion of the key 210a results in a subtree index 312 (Fig. 4) for subtree A stored in a subtree entry descriptor 304 (Fig.
4) in a mapper entry 504 (Fig. 6B) in mapper 106a in master lookup table 200a.
The subtree index 312 is forwarded to mapper 106b in master lookup table 200a to continue the search for the route entry 302 (Fig. 4) stored in master lookup table 200a.
A search for node 13023 with a first portion of the key 210 a results in a no-entry 300 (Fig. 4) stored in a mapper entry 504 (Fig. 6B) in mapper 106a in master S lookup table 200a and a subtree entry descriptor 304 (Fig. 4) stored in a mapper entry 504 (Fig. 6B) in mapper 106a in slave lookup table 200b. Thus, the search for the route entry 302 (Fig. 4) continues with the next portion of the key 210b in mapper 106b in slave lookup table 200b.
Fig. 12 is a flowchart illustrating a method for distributing route entries (Fig. 4) to be stored in mapper entries 504 (Fig. 6B) amongst the lookup tables 200a-b shown in Figs. 10A. The same method applies to the lookup tables 200c-d shown in Fig. 10C. The route entries 302 (Fig. 4) to be stored in mapper entries are initially stored in a memory by a processor (not shown) prior to being stored in the lookup tables 200a-b.
While the route entries 302 (Fig. 4) are stored in memory, the number of route entries 302 (Fig. 4) to be stored in each of the lookup tables 200a-b (Fig. 10A) are counted. The route entries 302 (Fig. 4) for mapper level-1 1104a (Fig.
11B) are stored in mapper 106a in lookup table 200a. The subtree entries 304 (Fig. 4) for mapper level_1 1104a (Fig. 11B) are stored in mapper 106a in each of the lookup tables 200a-200b.
At step 1200, the number of route entries 302 (Fig. 4) to be stored for each of the subtree entries 304 (Fig. 4) in mapper 106a in each of the lookup tables 200a-200b (Fig. 10A) are computed in order to determine how to distribute the route entries 302 (Fig. 4) amongst the lookup tables 200a-b (Fig. 10A). After the total number of mapper entries 504 (Fig. 6B) required to store the route entries 302 (Fig.
4) has been determined, processing continues with step 1202.
At step 1202, the total number of mapper entries 504 (Fig. 6B) to be stored for the subtrees is divided by the number of lookup tables 200a-b (Fig. 10A) to determine the number of route entries 302 (Fig. 4) to store in each lookup table 200a-b (Fig. 10A). Processing continues with step 1204.

At step 1204, a route entry 302 (Fig. 4)is stored in a mapper entry 504 (Fig.
6B) in a subtree mapper 418 (Fig. 5) in the selected lookup table 200a-b.
Processing continues with step 1206.
At step 1206, if the number of route entries stored in mapper entries 504 (Fig. 6B) in subtree mapper 418 (Fig. 5) in the selected lookup table 200a-b (Fig.
10A) is less than 1/n where n is the number of available lookup tables 200a-b (Fig.
10A), processing continues with step 1204. If not, the selected lookup table 200a-b already stores 1/n of the total number of mapper entries and processing continues with step 1208.
At step 1208, the selected lookup table 2000a-b stores 1/n of the total number of mapper entries, a no-entry 300 (Fig. 4) is stored for any remaining subtree nodes in the selected lookup table 200a-b because route indices for the respective subtree are not stored in the currently selected lookup table.
Processing continues with step 1210.
1 S At step 1210, if all the route entries have been stored, processing is complete.
If not, processing continues with step 1212.
At step 1212, the next lookup table 200a-b (Fig. 10A) is selected. Processing continues with step 1204.
Route entries are distributed amongst the lookup tables 200a-b (Fig. 10A) before searching for a route index corresponding to an IP address. The search is performed in parallel in each of the lookup tables 200a-b (Fig. 10A). The method for searching implemented in parallel in each of the lookup tables is described for one of the lookup tables 200a-b (Fig. 10A).
Fig. 13 is a flowchart illustrating a method for searching with a search key for a value corresponding to the search key stored in any one of the lookup tables 200c-d shown in Fig. IOC.
At step 1300, each of the lookup tables 200c-d (Fig. 10C) receives a search key 210. Mapper 106a in each of the lookup tables 200c-d is searched for a value corresponding to the first portion of the key 210a. Processing continues with step 1302.
At step 1302, the entry stored in mapper entry 504 (Fig. 6B) in mapper 106a is read. The mapper entry 504 (Fig. 6B) in master lookup table 200c can store no-entry 300 (Fig. 4), a route entry 302 (Fig. 4) or a subtree entry descriptor 304 (Fig.
4). The mapper entry 504 (Fig. 6B) in slave lookup table 200d can store no-entry 300 (Fig. 4) and subtree entry descriptor 304 (Fig. 4). If the mapper entry in the respective lookup table 200 stores a route entry 302 (Fig. 4), the entry is a valid S value, no further searching of subsequent mappers 106b-d in the lookup table 200c-d is required, processing continues with step 1310. If not, processing continues with step 1304.
At step 1304, if the entry stores a subtree entry descriptor 304 (Fig. 4), a further search in the lookup table 200c-d is required and processing continues with step 1306. If not, the entry stores no-entry, indicating that a further search is not required, processing continues with step 1310.
At step 1306, the search continues in the selected subtree. The next level mapper 106b-d (Fig. 3) is searched dependent on a next portion of the key 210b-d and a subtree index 312 (Fig. 4) resulting from a search of the previous level.
Processing continues with step 1308.
At step 1308, a determination is made as to whether to continue the search dependent on the mapper entry resulting from the search in the current level mapper 106b-d. If the mapper entry 504 (Fig. 6B) stores a subtree entry descriptor 304 (Fig.
4), the search continues with the next level mapper 106b-d and processing continues with step 1306. If the mapper entry 504 (Fig. 6B) does not store a subtree entry descriptor 304 (Fig. 4), a further search is not required, processing continues with step 1310.
At step 1310, the result of the search is compared with the incoming result 1000a received from another lookup table. For example, if the lookup table is slave lookup table 200d, the incoming result from the search in master lookup table 200c is forwarded on incoming result 1000a to lookup table 200d and compared with the result of the search in slave lookup table 200d. Processing continues with step 1312.
At step 1312, if the incoming result 1000a and the result of the search in the current lookup table 200d are different, processing continues with step 1314.
If both the incoming result 1000a and the result of the search in the current lookup table 200d are the same, two valid results have been stored in mapper entries 504 (Fig.

6B) in separate lookup tables 200c-d. Two valid results should not be stored for the same key 210, processing continues with step 1316.
At step 1314, the incoming result 1000a is checked to determine if it is valid.
An incoming result 1000a is valid if it is a route entry 302 (Fig. 4). The incoming result 1000a is invalid if it is a no-entry 300 (Fig. 4) or a subtree entry descriptor 304 (Fig. 4). subtree entry descriptor 304 (Fig. 4), route entry 302 (Fig. 4) and no-entry 300 (Fig. 4) have already been described in conjunction with Fig. 4. If the incoming result 1000a is invalid, processing continues with step 1318. If not, processing continues with step 1320.
At step 1318, the incoming result 1000a is valid and the result resulting from the search in the current lookup table 200d is invalid. The incoming result 1000a is forwarded on outgoing result 1002a from the current lookup table 200d. The incoming result 1000a is forwarded as the route index 102 (Fig. 3) if the current lookup table 200d is the last lookup table or is forwarded as the incoming result 1000a to the next lookup table. Processing is complete.
At step 1316, two valid result values are stored for the key in different lookup tables. An error occurred during the storing of the route entries in the lookup tables 200c-d. An error code is generated so that the error can be corrected.
Processing is complete.
At step 1320, neither the result from a search of the current lookup table 200d or the incoming result 1000a is valid. The result of search in the current lookup table 200d, even though invalid, is forwarded as the incoming result 1000a to the next lookup table. Processing is complete.
Fig. 14 is a flowchart illustrating a method for searching for a value corresponding to a search key stored in one of the lookup tables 200a-b shown in Fig. 10A.
At step 1340, the first level mapper 106a in both lookup tables 200a-b is searched for a value corresponding to a first portion of a key 210x.
Processing continues with step 1342.
At step 1342, if a valid result value is found after searching the first level mapper 106a with the first portion of the key 210a, processing continues with step 1352. If not, processing continues with step 1344.

At step 1344, if the value resulting from the search of the first level mapper 106a with the first portion of the key 210a is a subtree entry descriptor 304 (Fig. 4), processing continues with step 1346. If not, the valid value for the key is not stored in the current lookup table, processing is complete.
S At step 1346, the search for a valid value continues in the subtree identified in the subtree entry descriptor 304 (Fig. 4) found during the search of the previous level mapper. The next level mapper is searched for a value dependent on a next portion of the key 210b-c and a subtree select resulting from a search of the next level. Processing continues with step 1348.
At step 1348, the result of the search determines if a search of a next level mapper is required. The entry resulting from the current search can store a route entry 302, a no-entry 300 (Fig. 4) or a subtree entry descriptor 304 (Fig. 4).
If the entry stores a subtree entry descriptor 304 (Fig. 4), a further search is necessary and processing continues with step 1346. If the entry does not store a subtree entry descriptor 304 (Fig. 4), processing continues with step 1350.
At step 1350, if the entry stores a route index 102 (Fig. 3), processing continues with step 1352. If not, the entry is stored in another lookup table.
Processing is complete.
At step 1352, a valid value corresponding to the key is stored in the current lookup table. The valid value is forwarded as the route index 102 (Fig. 3) corresponding to the key. Processing is complete.
Sparse Mode Returning to Fig. 5, the subtree entry 404 provides for the access of up to 256 possible route indices, one per node in the 256 node subtree. The route indices are stored in mapper entries 504'-504" in the subtree mapper 418 (Fig. 5) (Fig. 5).
The mapper address 416 for a mapper entry 504 (Fig. 6B) in the subtree mapper (Fig. 5) is determined dependent on a dense subtree descriptor stored in the data field 406 and a subtree pointer stored in the pointers field 408 in the subtree entry 404. The format of dense subtree descriptor has already been described in conjunction with Figs 6A-6B. The dense subtree descriptor stores a node bit (Fig. 6B) for each node in the 256 node subtree. However, all subtrees have a different route index for each of the 256 nodes, for example, a subtree may only have one route index.
Fig. 15 illustrates a binary tree representation of a sparse subtree A and a dense subtree B in the second mapper level 112b indexed by subtree entries 304 (Fig. 4) in the first mapper level 112a. A subtree entry descriptor 304 (Fig.
4) for s1 in mapper 106a stores a subtree index 312 for the subtree entry 404 for subtree A. A
subtree entry descriptor 304 (Fig. 4) for s0 in mapper 106a stores a subtree index 312 for the subtree entry 404 for subtree B.
Densely populated subtree B has eleven route indices; that is, r6-r16 and six subtree entries; that is, s2-s7. The mapper addresses 416 corresponding to the mapper entries 504 (Fig. 6B) storing route entries 302 (Fig. 4) and subtree entries 304 (Fig. 4) for subtree B are coded in a dense subtree descriptor as has already been described in conjunction with Fig. 6B
Sparsely populated subtree A stores two route indices; that is r1 and r2. If they are stored in a dense subtree descriptor an entire subtree entry 404 is used to provide three mapper addresses 416 for the mapper entries 504 (Fig. 6B); that is, r0, r 1 and r2.
The number of routes stored in the lookup table 200 can be increased by coding a sparse subtree in one of a plurality of sparse subtree descriptors and coding a densely populated subtree in a dense subtree descriptor in a subtree entry 404.
A densely populated subtree has 16 or more mapper entries 504 (Fig. 6B), the data field 406 in the subtree entry 404 stores a dense subtree descriptor as described in conjunction with Figs. 6A-6B. A sparsely populated subtree has 15 or less mapper entries 504 (Fig. 6B); the data field in the subtree entry 404 stores a plurality of sparse subtree descriptors. By providing the ability to store sparsely populated subtrees in sparse subtree descriptors, more subtrees can be stored in the subtree memory 400 and thus more route entries can be stored in the lookup table 200.
Fig. 16A-C illustrate the modifications to the data field 406 and the pointers field 408 in the subtree entry 404 shown in Fig. 5 and the subtree entry descriptor 304 (Fig. 4) shown in Fig. 4 to permit storage of a plurality of sparse subtree descriptors in the subtree entry 404.

_2g_ Turnips to Fis. 16A., the data field 40G in a subtree entry 404 con(igurc<l in sparse mode includes a plurality of sparse subtree descriptors 1400' - 1400"
instead of tlic dense sublrcc descriptor with one bit per node of t1e subtrce described in conjunction with Fig. GB. Each sparse subtrec descriptor 1400' -1400" includes node descriptor 1402 - 1402". A node descriptor 1402' - 1402" is a 9-bit value representing a fully encoded rouse in the subtree. The node descriptor 1402'-1402"
describes a single node or a plurality of nodes in the subirce.
Turning to ~i g. 1 G13, to support storage of sparse sublree desct~ptors, a !node licld 1404 is added to the pointers field 408 in the subtrcc entry 404. The pointers 1 U licld 408 also stores block base address GOU' and block base address 6002 with each bloc!: including 1 G allocated mapher addresses 41G providing a total of 32 mapper addccsscs 41 G per subtrce entry 404. The mode Ccld 1404 stores a mode value.
The mode value stored in tllC I110dC field 1404 indicates the number of'sparse subtree descriptors 1400' - 1400" scored in the subtree entry 404 and the number of node descriptors 1402' - 1402" stored in each sparse subtrcc descriptor 1400' -1400".
Table 2 illustrates the conl7suration of the subtrcc entry 404 for each mode.
nodes/subtrcc bins subrrecsrou~ew/subtrcc unusedmapocr Isubtrccentry entries entry rc<)uircd for a subtrce envy mode max min nrax min max min o l 1 9 1G 16 16 112 3z 32 ?0 1 2 2 18 10 20 20 76 30 30 5 15 8 ~ 135 1 ~ 30 1 1G I ~ 32 ~ 18 'fable 2 Referring to Table 2, for example, if the mode value stored in mode field 1404 in the pointers field 408 in the subtree entry 404 is set to '4', each sparse subtree descriptor 1400 in the subtree entry 404 stores between 5 and 7 node descriptors 1402 ' - 1402". Each node descriptor 1402' - 1402" stores 9 bits.
The total number of bits stored in the sparse subtree descriptor 1400 is calculated by multiplying the number of node descriptors 1402' - 1402" per sparse subtree descriptor 1400' - 1400" by 9 (the number of bits per node descriptor 1402' -1402") Computing the number of bits per sparse subtree descriptor 1400 for mode 4, a sparse subtree descriptor 1400 with 7 node descriptors 1402 stores 63 bits (7 node descriptors * 9 bits = 63).
The number of sparse subtree descriptors 1400 per subtree entry 404 is computed by dividing the number of bits in the data field 406 by the number of bits in the sparse subtree descriptor 1400' - 1400". For mode 4, the number of bits in the data field 406 is 256 and the number of bits in the sparse subtree descriptor is 63.
Thus, the number of sparse subtree descriptors 1400' - 1400" is 4 ( int(256/63) = 4).
The total number of node descriptors 1402' - 1402" per subtree entry 404 is the number of nodes per subtree multiplied by the number of subtrees per subtree entry 404. Computing for mode 4, the total number of node descriptors 1402 per subtree entry 404 is 28, if there are 7 node descriptors 1402' - 1402" stored in a sparse subtree descriptor 1400' - 1400" (7*4=28), and 20 if there are 5 node descriptors 1402 per sparse subtree descriptor 1400' - 1400" (5*4=20).
The mapper entries column in Table 2 indicates how many mapper entries 504 (Fig. 6B) in the subtree mapper 418 (Fig. 5) are used by the subtree entry 404.
The mapper value is computed by incrementing the nodes per subtree by one and by multiplying by the number of subtrees in the sparse subtree descriptor. The nodes per subtree is incremented by one because one more mapper entry than the number of nodes per subtree is required to store the default entry for the subtree.
Referring to the mode 4 row in Table 2, 32 ((7+1) * 4 = 32) mapper entries are required per subtree entry 404 if there are seven node descriptors 1402 per sparse subtree descriptor 1400, and 24 ((5+1) * 4 = 24) node descriptors 1402 are required per sparse subtree descriptor 1400 if there are five node descriptors 1402 per sparse subtree descriptor 1400. The number of nodes per subtree and subtrees per subtree entry 404 are selected so that the maximum number of node descriptors per subtree entry 404 does not exceed 30 because mapper addresses 416 in subtree mapper (Fig. 5) are allocated in 16 block increments. The 32 mapper addresses 416 are provided by storing two block base addresses 600', 6002 stored in the pointers field 408.
Turning to Fig. 16C, each subtree entry 404 in the subtree memory 400 can be configured in dense mode as described in conjunction with Fig. 6B or in sparse mode. The subtree entry descriptor 304 (Fig.4) stored in subtree mapper 418 (Fig. S) which has been described in conjunction with Fig. 4 for dense mode is modified to allow sparse mode by providing an indication of whether the subtree entry 404 is coded in dense mode or in sparse mode. The indicator is provided by the type field 1406.
The state of the type field 1406 indicates whether the subtree entry 404 is configured in dense mode or sparse mode. If the subtree entry 404 is configured in 1 S sparse mode, the values stored in the sparse subtree descriptor select field 1408 and subtree index 312 are used to select a sparse subtree descriptor 1400. The sparse subtree descriptor select 1408 will be described in more detail later in conjunction with Fig. 16.
Fig. 17 illustrates the sparse mode logic 1502 in the offset logic 700 shown in Fig. 8 for providing a block offset 714 to select a mapper entry 504 (Fig.
6B) for a node in a sparsely populated subtree. The sparse mode logic 1502 provides the block offset 714 dependent on a node descriptor 1402 stored in a sparse subtree descriptor 1400 in a subtree entry 404. The offset logic 700 also includes dense mode logic 1500. Dense mode logic 1500 includes node select 706 and ones count logic 708 for providing a block offset 714 for a route in a densely populated subtree.
Dense mode logic 1500 has already been described in conjunction with Fig. 8.
If the state of the type field 1406 indicates that the subtree entry 404 is configured in sparse mode, the subtree data 412 from the subtree entry 404 is forwarded to the sparse mode logic 1502. The sparse mode subtree logic 1502 is described in conjunction with Fig. 18.
Fig. 18 illustrates the sparse mode logic 1502 shown in the offset logic 700 in Fig. 17. The sparse mode logic 1502 includes subtree select logic 1600, a multiplexer 1602, a Content Addressable Memory ("CAM") 1606 and conversion logic 1604. The sparse subtree descriptors 1400' -1400" stored in the data field 406 in the selected subtree entry 404 are forwarded on subtree data 412 to the offset logic 700. The offset logic 700 forwards the sparse subtree descriptors 1400' -1400"
S to the multiplexor 1602 in the sparse mode logic 1502. One of the sparse subtree descriptors 1400' in the subtree data 412 is selected by select 1614 generated by the subtree select logic 1600.
The subtree select logic 1600 generates select 1614 to select the sparse subtree descriptor 1400' dependent on the state of sparse subtree descriptor select 1408 forwarded from the mapper entry selected in the previous mapper level and mode 1404 stored in the pointers field 408 in the selected subtree entry 404.
Table 3 illustrates the selected sparse subtree descriptor 1400' and the respective subtree data bits 412 forwarded on selected sparse subtree descriptor 1610 from the multiplexer 1602 for a mode 4 subtree entry 404. Referring to the mode 4 row in Table 2, four sparse subtree descriptors can be stored in the mode 4 subtree entry 404. Each of the four sparse subtree descriptors 1400 is 63 bits and can store from seven to five node descriptors 1402. Thus, each of the four sparse subtree descriptors 1400 starts~on a 63 bit boundary. The first sparse subtree descriptor 1400' is stored at bits 62:0 in the data field 406. The second sparse subtree descriptor 14002 is stored at bits 125:63 in the data field 406. The third sparse subtree descriptor 14003 is stored at bits 188:126 in the data field 406 and the fourth sparse subtree descriptor 14004 is stored at bits 251:189 in the data field. The respective bits in the data field 406 are selected by the sparse subtree descriptor select 1408. For example, looking at Table 3, if sparse subtree descriptor select 1408 is '0001' , the second sparse subtree descriptor 14002 is selected and Bits 125:63 of the 256 bit subtree data 412 are forwarded through the multiplexor 1602 on selected sparse subtree descriptor 1610 to the conversion logic 1604.

Sparse subtreeSubtrce data Sparse Subtrcc Descriptor descriptor select 0000 Bits G2:0 1 S 0001 l3ils 125;63 2 OU1.0 Bits 188;126 3 001 I lids 251:189 4 Table 3 L?ach sublre;c-: entry 404 in the subtree memory 400 can he configured in sparse mode or dcnsu rnodc. L;ach subtree catty 404 configured in sparse ntodc can he ronligurcd to store a different number of mode descriptors 1402 per sparse subirec descriptor 1400 tl~rougli mode 1404. All sparse subtTee descriptors 1400 in subtrcc catty 404 con figured in sparse mode store the same number of node descriptors J 402 per sparse subtree descriptor 1400.
1 S A node descriptor 1402 can be coded to represent multiple nodes in the subtrec. Multiple ci~;ht bit nodes represcrttcd by the node descriptor 1402 arc identified by masking some ol'thc eight bits. Titstead of storing mask bits with each node descriptor 1402, a nine bit node descriptor 1402 is used to fully encode the ciglO bit wide nodes represented by the node descriptor 1402. The eight bit wide node is encoded in a nine bit wide node descriptor 1402 using mn bit length encoding. Run bit length encoding permits identification oCwhich of the tight bits of the node arc rtiaskcd, The conversion logic 1 G04 converts the nine bit wide node descriptors 1402'-1402" stored in the s~:lecled sparse subtrcc descriptor 1400 into eight bit CAM
values 1612 including bits set to 'X' (don't care) and loads the eight bit CAM
values 1612 into the CAM 1606. Au example of the conversion of. the 9-hit node descriptors 1402 into 8-bit CAM values 1 G12 by the conversion logic 1 G04 is shown t n '1'ahle 4 below.

8-hit value 9-bit code 1 ooxxxxx a ool 00000 0011 XXXX 0011. l Table 4 The 9-bit code column illustrates values stored in node descriptors 1402.
T,ooking at the first row in 'fable 4, the 9-bit code stored in the node descriptor 1402 is '101100100' and the corresponding 8-bit value '101100XX' is stored in the CAM
1 G06. The conversion logic 1604 converts the 9-bit code by searching the J-bit code from risllt to left for the first bit set to '1'. Looking at the bits in the 9-bit code from right to left, the first two bits arc set to 'U' and the third bit is set to '1'. The CUllVcr51011 lobic 160'1 Converts the '100' to two don't cares ('XX') because there arc two '0's to the right of the best '1'. 'Che .first '1' is ignored and the remaining bits are 1 S directly copied into the next bits of the 8-bit value.
Looking at the second row in Table 4, the 9-bit code stored in the node descriptor 1402 is '100100000', The; conversion logic 1604 converts the 9-bit code by searching ihc 9-bit code from right to left for the first 'I'. The firth digit stores a '1'. 'fhe 9-bit code is converted to an 8-bit value with the five Least Significant Bits ("I.SBs") set to "don't cares"("X"). I3y storing the node descriptors 1402 using 9-bit n1n bit length cncodinb, the nucnbcr of bits required per node descriptor 1402 is IT1111111117.cd, ihcreby increasing the number of node descriptors 1402 which can be stored in the lookup table 20U.
After converting the 9-bit node descriptors 1402 into 8-bit values, the 2S conversion logic l GO'1 loads the 8-bit values into the CAM 1 GOG. 'fhe 8-bit values arc loaded into the CAM 1 G06 in the same order as the node descriptors 1402 arc stored is the selected sparse subtree descriptor 1400; that is, from shortest to longest nvatch. The CAM 1606 provides storage for storing the maximum number of node descriptors 1402 per sparse subtree descriptor 1400. Thus, the CAM 1606 is 8 bits 3U wide by sixteen entries dceP to provide fifteen entries to store the maximum number of node descriptors 1402 for a mode 5 sparse subtrcc descriptor 1400 and a default mahhcr address. The CAM 160G has ternary capability and mufti-match resolver built in. A small size CAM 1606 can be implemented in gates, as opposed to providing a true contents addressable memory; that is, the CAM 1606 can be implemented in hardware circuitry which emulates a CAM.
The number of node descriptors 1402 stored in a sparse subtree descriptor S 1400 determines the subtree entry 404 in which the sparse subtree descriptor 1400 is stored. Sparse subtree descriptors 1400 storing node descriptors 1402 within the range for a particular mode are stored in the same subtree entry 404. A
default mapper address is computed for a default route for each subtree. A default 8-bit value is permanently stored in the first location in the CAM 1606 to compute the default mapper address.
After the 8-bit values for the selected sparse subtree 1400 have been loaded into the CAM 1606, the CAM 1606 is searched with the next portion of the key 210b. The entry in the CAM 1606 matching the greatest number of bits in the next portion of the key 210b is selected. The match address resulting from the search of the CAM is forwarded as the block offset 714. The block offset 714 is used to determine the mapper address 416 for the mapper entry corresponding to the route stored in the subtree mapper 418 (Fig. 5).
Figs. 19A-D illustrate the selection of a block offset 714 for a node in a sparsely populated subtree 1700. Fig. 17A is a graphical representation of the routes in the sparsely populated subtree 1700. Nodes in the subtree 1700 correspond to one of three routes, r0, r1 and r2, r0 is the default route for the subtree 1700.
Two routes r1, r2 are encoded in node descriptors 1402' and 14022 in a sparse subtree descriptor 1400. A value for the default route r0 is permanently stored in the first entry 1702 in the CAM 1606. Referring to Table 2, a sparse subtree descriptor 1400 with two node descriptors 1402 is stored in subtree entry 404 with the mode field 1404 set to ' 1'.
Looking at subtree 1700, r2 corresponds to all nodes matching lOxxxxxx and r1 corresponds to all nodes matching OlOxxxxx. In order to minimize the number of bits required by each node descriptor 1402', 14022 to describe each route in the sparse subtree descriptor 1400, the node descriptor 1402',14022 is coded using run bit length encoding. The method for coding uses one bit more than the number of bits used to fully encode the node. A '1' is inserted in the location of the first 'X' ('don't care') and the remaining Xs are coded as 0's. Thus, route lOxxxxxx is translated to 10100000 and Ol Oxxxxx is translated to 010100000.
Fig. 19B illustrates the storage of the node descriptors 1402' and 14022 in the sparse subtree descriptor 1400. The node descriptors 1402' and 14022 are stored in a subtree entry 404 with the mode field 1404 set to '1' because there are two node descriptors 1402' and 14022 stored in the sparse subtree descriptor 1400. The longest match for the subtree is r1 because r1 requires a match of the first three bits and r2 requires a match of the first two bits. The node descriptors 1402' and are stored in shortest to longest match order in the sparse subtree descriptor 1400, with the node descriptor 1402' for r2 stored first and the node descriptor 14022 for r1 stored next.
Fig. 19C illustrates the conversion of the node descriptor 14022 to an 8-bit masked value 1706. Looking at the node descriptor bits 1708' -17089 from left to right, the first ' 1' is stored in bit 17086 , this marks the end of the mask bits for the 8-bit masked value 1706. To convert the node descriptor 14022 to an 8-bit masked value 1706, the following bit conversions are performed. The '0' stored in node descriptor bit 1708' is converted to 'X' and stored in 8-bit masked value bit 1710'.
The '0' stored in node descriptor bit 17082 is converted to 'X' and stored in 8-bit masked value bit 17102. The '0' stored in node descriptor bit 17083 is converted to 'X' and stored in 8-bit masked value bit 17103 . The '0' stored in node descriptor bit 17084 is converted to 'X' and stored in 8-bit masked value bit 17104. The '0' stored in node descriptor bit 17085 is converted to 'X' and stored in 8-bit masked value bit 17105. The '1' stored in node descriptor bit 17086 is ignored. The '0' stored in node descriptor bit 1708'is stored in 8-bit masked value bit 17106. The '1' stored in node descriptor bit 17088 is stored in 8-bit masked value bit 1710'. The '0' stored in node descriptor bit 17089 is stored in 8-bit masked value bit 1710$.
Fig. 19D illustrates the storage of node descriptors 1402' and 14022 in the CAM 1606 and the corresponding mapper entries 504' - 5043 stored in the subtree mapper 418 (Fig. 5) for the selected sparse subtree descriptor 1400. The 9-bit node descriptors 1402' and 14022 stored in the selected subtree descriptor 1400 are converted in the conversion logic 1604 (Fig. 18) and loaded into CAM 1606. The first entry 1702 in the CAM 1606 is the default entry for r0 shown in subtree 1700 in Fig. 19A. The second entry 1704 is converted from the first node descriptor 1402' stored in the selected sparse subtree descriptor 1400. The second entry 1704 is the shortest match which is converted for r2. The second node descriptor 14022 stored in the selected subtree descriptor 1400 is converted from 010100000 to OIOXXXXX
and stored in the third entry 1706 in the CAM 1606.
A search of the CAM 1606 results in the block offset 714 (Fig. 18). The block offset 714 is used to determine the mapper address 416 for the mapper entry 504' - 5043 stored in the subtree mapper 418 (Fig. 5). The CAM 1606 is searched with the second portion of the key 21 Ob for the entry 1702,1704,1706 storing the longest match. The block offset 714 provided by the CAM 1606 is combined with a subtree base address dependent on one of the block base addresses 600', 600' stored in the pointers field 408 in the selected subtree entry 404.
Fig. 20 is a block diagram illustrating the sparse mode base select logic 1800 in the pointer logic 702 shown in Fig. 8. The pointer logic 702 selects the base address 716 used to compute the mapper address 416 for the mapper entry 504 (Fig.
6B) in the subtree mapper 418 (Fig. 5). The pointer logic 702 includes dense mode base select logic 710 and sparse mode base select logic 1800, one of which is selected, dependent on the state of type 1406 stored in the subtree entry descriptor 304 (Fig. 4), forwarded from the previous mapper level. As has already been described, the state of type 1406 indicates whether the subtree entry 404 is configured in dense mode.
Sparse mode base select logic 1800 computes the base address 716 for the sparse subtree descriptor 1400 if the subtree entry 404 stores a plurality of sparse subtree descriptors 1400. The sparse mode base select logic 1800 computes the base address 716 using the mode value 1608 stored in the mode field 1404 and subtree pointers 414 stored in block base address field 600', 6002 in the subtree entry 404 and the sparse subtree descriptor select 1408 stored in the subtree entry descriptor 304 (Fig. 4)forwarded from the previous mapper level. The base address 716 is computed as follows:

base address (for the sparse subtree descriptor) _ block base address + base offset where base offset = ((1 + nodes/subtree) * sparse subtree descriptor select)) ~ For example, to find the base address 716 for the start of subtree number 2 in a subtree entry 404 configured in sparse mode 4, the base offset is computed first.
The sparse subtree descriptor select 1408 for subtree number 2 is '1' and the number of nodes/subtree is 7 (See Table 2). The base offset is 8 ((1 + 7) * 1). Each block base address 600', 6002 is the base address for a block of 16 mapper addresses allocated for the subtree entry 404. The base offset for subtree number 2 is 8 which is less than 16 therefore the block base address for subtree 2 is block base address 600' and the base address 716 for the sparse subtree descriptor is block base address 600' + 8. Table 5 below illustrates the subtree base address for each of the four subtrees in a subtree entry 404 configured in mode 4.
1 S Subtree base addressSubtree block base address 1 1 + 0 block base address 2 2 + 8 block base address 3 2 + 0 block base address 4 2 + 8 Table 5 Fig. 21 illustrates a dense subtree descriptor and a sparse subtree descriptor stored in the subtree memory 400. Fig. 21 is described in conjunction with Fig. 15.
A dense subtree descriptor for subtree B (Fig. 21) is stored in the data field 406' in subtree entry 404' . A sparse subtree descriptor 14001 for subtree A (Fig. 21) is stored in the data field 4062 in subtree entry 4042 . The dense subtree descriptor stores a node bit for each node in the bottom level of subtree B as has already been described in conjunction with Fig. 6B. The sparse mode descriptor 14001 includes node descriptors 1402' and 14022 corresponding to routes r4 and r5 as has been described in conjunction with Fig. 19B. The subtree index 312 selects the subtree entry 404', 4042.

The subtree index 312 stored in a subtree entry descriptor 304 (Fig. 4) in a mapper entry 504 (Fig. 6B) in mapper 106a for s0 (Fig. 15) selects subtree entry 404'. The subtree index 312 stored in a subtree entry descriptor 304 (Fig. 4) in a mapper entry 504 (Fig. 6B) in mapper 106a for sl(Fig. 15) selects subtree entry 4042. Thus, the subtree memory 400 can store subtree entries 404',4042 for sparse subtrees and dense subtrees.
Fig. 22 is a flow chart illustrating a method for providing a mapper address 416 (Fig. 5) for a mapper entry 504 (Fig. 6B) in a subtree mapper 418 (Fig. 5) storing a route for a node in a sparsely populated subtree and a densely populated subtree. Any subtree entry 404 may store a plurality of sparse subtree descriptors or a single dense subtree descriptor. Any combination of sparse subtree descriptors and dense subtree descriptors is possible dependent on how the routes are distributed in the binary tree. The flexibility to mix and match sparse mode and dense subtree descriptors in subtree entries 404 in the subtree memory 400 allows better utilization of the subtree memory 400.
At step 1900, the configuration of the selected subtree entry 404 is determined from the state of type 1406 (Fig. 16C) stored in the subtree entry descriptor 304 (Fig. 4) selected in the previous mapper level. If the subtree entry 404 type is configured in sparse mode, processing continues with step 1902. If not, processing continues with step 1914.
At step 1902, the subtree entry 404 is configured in sparse mode. A subtree entry 404 configured in sparse mode stores a plurality of sparse subtree descriptors 1400. The number of sparse subtree descriptors 1400 stored in the subtree entry 404 is dependent on the state of the mode field 1404. The sparse mode logic 1502 in the offset logic 700 selects the sparse subtree descriptor 1400 from the subtree entry 404 dependent on the sparse subtree descriptor select 1408 stored in the subtree entry descriptor 304 (Fig. 4) forwarded from the previous mapper level and the contents of the mode field 1404 as was described earlier in conjunction with Fig. 14.
Processing continues with step 1904.
At step 1904, the 9-bit coded values stored in the node descriptors 1402 in the selected sparse subtree descriptor 1400 are converted into 8-bit values and stored in order of shortest to longest match in the CAM 1606. Processing continues with step 1906.
At step 1906, the CAM 1606 is searched with the next portion of the key 210b for the CAM entry storing the longest match. Processing continues with step 1908.
At step 1908, the address of the location in the CAM 1606 storing the longest match for the next portion of the key 210b is forwarded as the block offset 714. The block offset 714 is used to compute the mapper address 416 (Fig. 5) of the mapper entry 504 (Fig. 6B) in the subtree mapper 418 (Fig. 5). Processing continues with step 1910.
At step 1910, the base address 716 (Fig. 20) for the selected sparse subtree descriptor 1400 is computed dependent on the sparse subtree descriptor select stored in the subtree entry descriptor 304 (Fig. 4) forwarded from the previous mapper level and the contents of the mode field 1404 stored in the selected subtree entry 404. Processing continues with step 1912 At step 1912, the mapper address 416 is computed by adding the block offset 714 and the base address 716 in the adder logic 704 (Fig. 8). The mapper entry (Fig. 6B) identified by the mapper address 416 in the subtree mapper 418 (Fig.
5) either stores a route entry 302 (Fig. 4) or a subtree entry descriptor 304 (Fig. 4). If the mapper entry 504 (Fig. 6B) stores a route entry 302 (Fig. 4) the search is complete. If the mapper entry 504 (Fig. 6B) stores a subtree entry descriptor (Fig. 4), the search for the value corresponding to the key 210 continues in the next mapper level.
At step 1914, the subtree entry 404 is configured in dense mode and stores a single dense subtree descriptor in the data field 406. The block offset 714 is computed by counting the number of '1's stored in dense subtree descriptor stored in the data field 406 in the subtree entry 404 as has been described earlier in conjunction with Fig. 6B. Processing continues with step 1916.
At step 1916, the subtree entry 404 stores sixteen block base addresses 600 in the pointers field 408 in the subtree entry 404. One of the block base pointers 600 is selected by the dense mode base select logic 710 in the pointers logic 702 described earlier in conjunction with Fig. 8. Processing continues with step 1912.

Incremental Update Fig. 23 illustrates a binary tree representation of a new route to be added to the lookup table 200. The binary tree illustrates routes stored in the lookup table 200 for mapper level_1 2000, mapper level 2 2002 and mapper level_ 3 2004.
Mapper level 2 2002 stores routes for subtrees A and B. Mapper level 3 2004 stores routes for subtrees A,, A2, B, and BZ _ s5 represents a subtree entry descriptor 304 (Fig. 4) stored in a subtree mapper 418 (Fig. 5). The subtree entry descriptor 304 (Fig. 4)for s5 stores a pointer to Subtree BZ allowing the search for a longest match route for a key 210 to continue in mapper level 3 2004.
Subtree BZ 2006 is a sparse subtree because it has only two routes, r6 and hl.
Thus, node descriptors 1402 (Fig. 16A) for nodes r6 and hl are stored in a sparse subtree descriptor 1400, as has already been described in conjunction with Fig. 14A.
The sparse subtree descriptor 1400 for Subtree BZ 2006 is stored in a subtree entry 404 with mode field 1404 set to 1 in the subtree memory 400 because there are two node descriptors 1402 stored in the sparse subtree descriptor 1400.
A new route h2 shown in subtree BZ' 2008 is to be added to the lookup table 200. The new route hl can not be added directly to subtree Bz 2006 in the lookup table because the addition of a route to subtree BZ 2006 increases the number of node descriptors 1402 stored in the sparse subtree descriptor 1400 from 2 to 3. The addition of a node descriptor 1402 to a sparse subtree descriptor 1400 requires the allocation of a new sparse subtree descriptor 1400 in a subtree entry 404 with mode field 1404 set to '2'. Thus, the addition of new route hl requires the replacement of subtree Bz 2006 by subtree Bz' 2008.
Fig. 24 illustrates update routes stored in a processor memory 2400. A copy of the binary tree stored in the lookup table 200 is also stored in processor memory 2400 separate from the lookup table 200. The routes stored for Subtree BZ 2006 are copied to subtree Bz' 2008 in the processor memory 2400 and the new route h2 is added to subtree BZ' 2008.
A route update routine 2402 generates a sequence of route update instructions 2404 to add subtree Bz' 2008 to the lookup table 200 and forwards the route updates 2404 to the table update routine 2406. The table update routine generates table updates 2410 for the route updates 2402 and forwards the update cycles 2412 to update the lookup table 200 with the route updates 2404. The update cycles 2412 write the route updates to the appropriate memory locations in the subtree memory 400 (Fig. 5) and the subtree mapper 418 (Fig. 5).
Returning to Fig. 23, the update cycles 2412 include instructions to allocate a portion of the subtree mapper 418 (Fig. 5) to store the routes for the new subtree BZ' 2008 in mapper entries 504 (Fig. 6B). Subtree BZ' 2008 includes route entries stored in mapper entries 504 (Fig. 6B) for routes hl and r6 and new route h2. After the route entries for subtree BZ' 2008 are stored in mapper entries 504 (Fig. 6B) in the subtree mapper 418 (Fig. 5), the node descriptors 1402 for the routes are created and stored in a sparse subtree descriptor 1400. The sparse subtree descriptor 1400 is stored in a subtree entry 404. The mode 1404 of the subtree entry 404 is related to the number of node descriptors 1402 stored in the sparse subtree descriptor 1400.
After the sparse subtree descriptor 1400 for subtree Bz' 2008 is stored in a subtree entry 404 in subtree memory 400 in the lookup table 200, the subtree entry descriptor 304 (Fig. 4) represented by s5 is modified to point to subtree Bz' instead of subtree BZ 2006. While subtree BZ' 2008 is being added to the lookup table, routes r6 and hl stored in subtree Bz 2006 can be accessed through s5.
After subtree BZ' 2008 is stored in the lookup table, and s5 is modified to point to subtree BZ' 2008 and routes r6, hl and the new route h2 can be accessed. Thus, subtree BZ
2006 can continue to be searched for route indices corresponding to route r6 and hl while the new route h2 is being added to the lookup table 200.
Fig. 25 illustrates the new route h2 shown in Fig. 23 stored in mapper entry 504c4 in subtree mapper 418b in the lookup table 200. Fig. 25 is described in conjunction with the binary tree representation shown in Fig. 24.
Subtree B in mapper level 2 2002 has three routes; that is r3, s4 and s5.
Subtree B is a sparse subtree because it has less than sixteen routes. Node descriptors 1402a' - 1402a3 for the Subtree B r3, s4 and s5 are stored in a sparse subtree descriptor 1400a' in subtree entry 404a in subtree memory 400a. A
mapper entry 504a2 - 504a4 is stored in subtree mapper 418a for each route in Subtree B. A
default route for Subtree B is stored in mapper entry 504a' in the subtree mapper 418a. Each mapper entry 504a2 - 504a4 stores a route entry 302 (Fig. 4) or a subtree entry descriptor 304 (Fig. 4) for the node. A subtree entry descriptor 304 (Fig. 4) is stored for route s4 in 504a' and s5 in 504a4. The subtree entry descriptor 304 (Fig.
4) stored in mapper entry 504a' for s5 provides the subtree index 312b for subtree memory 400b to start the next level search; that is for mapper level 3 2004.
Subtree Bz is also a sparse subtree because it has two routes; that is, hl and r6. Node descriptors 1402b' - 1402b2 are stored in sparse subtree descriptor 1400b' in subtree entry 404b in subtree memory 400b. Each route in Subtree BZ is stored in a mapper entry 504b2 - 504b3 and the default route for Subtree BZ is stored in mapper entry 504b'.
To search for route hl in Subtree Bz 2006, the address of the subtree entry 404a storing the sparse subtree descriptor 1400a which stores the node descriptor 1402 for route s5 is forwarded on subtree index 312a to subtree memory 400a.
The data field 406 and pointers field 408 stored in the selected subtree entry 404a are forwarded on subtree data 412a and subtree pointers 414a to the mapper address logic 402a. The mapper address logic 402a generates the mapper address 416a for the mapper entry 504a4 storing the subtree entry for s5. The mapper address 416a is dependent on the subtree data 412a, subtree pointers 414a and a next portion of the key 210b. The subtree entry for s5 is forwarded on subtree index 312b to subtree memory 400b.
Subtree memory 400b stores node descriptors 1402b2, 1402b' for subtree BZ
2006. The sparse subtree descriptor 1400b' for BZ is stored in subtree entry 404b.
The data field 406 and pointers field 408 stored in subtree entry 404b are forwarded on subtree data 412b and subtree pointers 414b to mapper address logic 402b.
The mapper address logic 402b generates the mapper address 416b for the mapper entry 504b3 storing the route entry for hl. The mapper address 416b is dependent on the subtree data 412b, subtree pointers 414b and a next portion of the key 210c.
To add route h2 to subtree Bz 2006, a block of previously unused mapper entries 602c in subtree mapper 418b are allocated to store mapper entries 504cz-504c4 storing routes r6, hl and h2 for Subtree BZ' 2008. Mapper entry 504c' stores the default entry for Subtree Bz' 2008; that is, the same value stored in mapper entry 504b'. Mapper entry 504c2 stores the route entry for route r6; that is, the same value stored in mapper entry 504b2. Mapper entry 504c3 stores the route entry for route hl; that is the same value stored in mapper entry 504b3. Mapper entry 504c4 stores the route entry for the new route h2. While the block of mapper entries 504c'~4 are being written, the route entries stored in mapper entries 504b'-504b3 can be accessed through the subtree entry stored for route s5 in 504a4 in subtree mapper 418a.
Having stored the mapper entries 504c'~4 for subtree Bz' 2008 in subtree mapper 418b, a sparse subtree descriptor 1400c' is added to the subtree memory 400b. The number of node descriptors 1402c'-3 is less than sixteen, therefore, the node descriptors 1402c'-3 are stored in a sparse subtree descriptor 1400c'.
The location of the subtree descriptor 1400' in the subtree memory 400b is dependent on the number of node descriptors 1402c'-3 associated with the sparse subtree descriptor 1400c'. By adding a new route to subtree BZ 2006, the number of node descriptors 1402c'-1402c3 to be stored for the sparse subtree descriptor 1400c' has increased from two to three. The sparse subtree descriptor 1400c' is stored in a subtree entry 404c with three node descriptors per sparse subtree descriptor and the mode field 1404 set to '2'. The sparse subtree descriptor 1400c' is stored in a current mode 3 subtree entry 404c if there is space available or a new mode 3 subtree entry is allocated. Node descriptors for routes in Bz' 2008 are stored in node descriptors 1402c'-3 in the sparse subtree descriptor 1400c' in the mode 3 subtree entry 404c.
After the sparse subtree descriptor 1400c' and node descriptors 1402c'-3 have been stored in subtree memory 400b, subtree BZ' 2008 can be accessed. To provide access to BZ' 2008, subtree entry 504a' is modified to index sparse subtree descriptor 1400c' in subtree entry 404c instead of sparse subtree descriptor 1400b' in subtree entry 404b. The route entry for route h2 stored in mapper entry 504c4 and routes r6 and hl stored in respective mapper entries 504cz and 504c' can be accessed.
Mapper entries 504b'-504b3 can no longer be accessed and are deallocated and placed on a free list (not shown) for future allocation. Also, sparse subtree descriptor 1400b' can no longer be accessed. Thus, sparse subtree descriptor 1400b' is deallocated and placed on a free list (not shown) for future allocation.
The addition of a route to a sparse subtree has been described. A route can also be added to a dense subtree by storing a new dense subtree descriptor in a newly allocated subtree entry 404 and the corresponding mapper entries in the subtree mapper 418, and modifying the subtree entry stored in mapper entry 504a4 to index the newly allocated subtree entry 404.

Fig. 26 is a flowchart illustrating the steps for performing an incremental update to add a route to the lookup table 200 shown in Fig. 25.
At step 2200, the number of routes per subtree is computed to determine if the route update results in a sparse or dense subtree. If the subtree is dense after the route update, processing continues with step 2218. If the subtree is sparse after the route update, processing continues with step 2202.
At step 2202, the subtree is sparse. The sparse subtree mode is determined.
Processing continues with step 2204.
At step 2204, a list of partially filled subtree entries 404 stored in the subtree mapper 418 (Fig. S) is searched to determine if the new sparse subtree descriptor 1400c' can be stored in a previously allocated subtree entry 404. For example, four sparse subtree descriptors 1400c'-1400c4 can be stored in a mode 4 subtree entry 404. If only three are stored, the subtree entry 404 is partially filled and stored on the list of partially filled subtree entries 404. If there is a partially filled subtree entry 404 available, processing continues with step 2208. If not, processing continues with step 2206.
At step 2206, a new subtree entry 404c is allocated for storing the sparse subtree descriptor 1400c' and mapper entries 504c'-504c4 are allocated in the subtree mapper for storing the mapper entries 504 (Fig. 6B) for the node descriptors 1402c'-1402c3 stored in the sparse subtree descriptor 1400c' in the newly allocated subtree entry 404c. A pointer to an allocated block of mapper entries 504c'-504c4 in subtree mapper 418 (Fig. 5) is stored in the pointers field 408 in the new subtree entry 404c.
Processing continues with step 2208.
At step 2208, the location of the first mapper entry 504c' in the subtree mapper for the sparse subtree descriptor 1400c' is determined from pointers stored in the pointers field 408 in the subtree entry 404c and the mode stored in the mode field 1404 in the subtree entry 404c. Processing continues with step 2210.
At step 2210, the route entries for the sparse subtree are stored in the mapper entries 504c'-504c4 in the subtree mapper 418b. Processing continues with step 2212.

At step 2212, the sparse subtree descriptor 1400c' storing the node descriptors 1402c'-1402c' is stored in the subtree entry 404c. Processing continues with step 2214.
At step 2214, the subtree entry descriptor 304 (Fig. 4) stored in mapper entry S 504a4 is modified to index the new sparse subtree descriptor 1400c' stored in subtree entry 404c. The route entry for h2 stored in mapper entry 504c4 can now be accessed. Processing continues with step 2216.
At step 2216, mapper entries 504b' -504b' and sparse subtree descriptor 1400b can no longer be accessed. Mapper entries 504b1-504b3 are placed on a free list of mapper entries 504 (Fig. 6B) for subtree mapper 418b and can be allocated for storing other routes. The first available location in subtree entry 404b is updated in the list of partially filled subtree entries. Processing is complete.
At step 2218, a new subtree entry 404 is allocated from a list of free subtree entries 404 stored in processor memory 2400 (Fig. 24). The new subtree entry is allocated for storing a new dense subtree descriptor. Blocks of mapper entries 504 (Fig. 6B) in the subtree mapper 418b are allocated for storing the routes. The pointers to the blocks of allocated mapper entries 504 (Fig. 6B) are stored in the pointers field 408 (Fig. 7) in the subtree entry 404 (Fig.S). Processing continues with step 2220.
At step 2220, the new dense subtree descriptor is written in the data field in the new subtree entry 404 as has been described earlier in conjunction with Figs.
6A-B. Processing continues with step 2222.
At step 2222, the route entries for the dense subtree are stored in the mapper entries 504 (Fig. 6B) in subtree mapper 418 (Fig. 5) identified by the pointers stored in the pointers field 408 in the subtree entry 404. Processing continues with step 2224.
At step 2224, the subtree entry descriptor 304 (Fig. 4) stored in mapper entry 504a4 is modified to index the new dense subtree descriptor stored in the new subtree entry 404c. The route entry for h2 stored in mapper entry 504c4 can now be accessed. Processing continues with step 2226.
At step 2226, the mapper entries 504 (Fig. 6B) indexed by the pointers stored in the pointers field 408 in the old subtree entry 404 are returned to the free list of mapper entries stored in processor memory 2400 (Fig. 24). The old subtree entry 404b is added to a free list of subtree entries stored in processor memory 2400 (Fig.
24).
The process has been described for the addition of a route to the lookup table. A similar process is performed to delete a route from the lookup table.
For example, to delete h2 504c4 from Subtree Bz' requires storing a new sparse subtree descriptor with two node descriptors for routes r6 and hl, storing the sparse subtree descriptor in a mode 2 subtree entry, updating the corresponding subtree mapper and modifying the subtree entry descriptor 304 (Fig. 4) stored in mapper entry 504a4 to index the updated subtree descriptor stored in the new subtree entry 404.
While this invention has been particularly shown and described with references to preferred embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the scope of the invention encompassed by the appended claims.

Claims (21)

What is claimed is:
1.A lookup table comprising:
a first multi-level lookup unit which receives a key and provides a first result through a sequential multi-level search for a longest prefix match for the key in the first multi-level lookup unit; and a second multi-level lookup unit which receives the key and provides a second result through a sequential multi-level search for a longest prefix match for the key in the second multi-level lookup unit, both the first multi-level lookup unit and the second multi level lookup unit receiving the key in parallel and only one of the units providing a final result for the key stored in only one of the multi-level lookup units dependent on the first result and the second result.
2. The lookup table as claimed in Claim 1 wherein final results for the key are distributed amongst the multi-level lookup units dependent on the total number of final results to be stored in the multi-level lookup table.
3. The lookup table as claimed in Claim 1 wherein the key is an IP address and the final result is a route index.
4. The look-up table as claimed is Claim 1 wherein a first memory location in the first multi-level lookup unit indexed by the key stores the final result and a second memory location in the second multi-level lookup unit indexed by the key stores an indication that the final result is stored the first multi-level lookup unit.
5. The lookup table as claimed in Claim 1 wherein each of the first multi-level lookup unit and the second multi-level lookup unit includes a plurality of mapper levels and each of the mapper levels includes memory locations for storing results.
6. The lookup table as claimed in Claim 4 wherein the final result corresponding to the key stored is a location in any one of the plurality of mapper levels in the first multi-level lookup unit is forwarded to the second multi-level lookup unit and provided by the second lookup unit.
7. The lookup table as claimed in Claim 1 wherein the final result corresponding to the key is provided by the multi-level lookup unit in which the final result is stored.
8. A method for providing a final result for a key comprising the steps of:
receiving the key in parallel by a first multi-level lookup unit and a second lookup unit;
performing a sequential multi-level search for a longest prefix match for the key by the first multi-level lookup unit, resulting in a first result;
performing a sequential multi-level search by for the longest prefix match for the key the second multi-level lookup unit, resulting in a second result, and providing the final result for the key by one of the units dependent on the first result and the second result.
9. The method as claimed is Claim 8 further comprising the steps of:
distributing the final results amongst the units dependent on the total number of final results to be stored is the multi-level lookup tables.
10. The method as claimed in Claim 8 wherein the key is an IP address and the final result is a route index.
11. The method as claimed in Claim 8 wherein a first memory location in the first multi-level lookup unit indexed by a key stores the final result and a second memory location in the second multi-level lookup unit indexed by the key stores an indication that the final result is stored in the first multi-level lookup unit.
12. The method as claimed in Claim 8 wherein each of the first multi-level lookup unit and the second multi-level lookup unit includes a plurality of mapper levels and each of the mapper levels include memory locations for storing results.
13. The method as claimed in Claim 12 wherein the result corresponding to the key stored in a location in one of the plurality of mapper levels in the first multi-level lookup unit is forwarded to the second lookup and provided by the second lookup unit.
14. The method as claimed in Claim 8 wherein final the result corresponding to the key is provided by the multi-level lookup unit in which the final result is stored.
15. A lookup table comprising:
a first multi-level lookup unit which receives a key and provides a first result through a sequential multi-level search for a longest prefix match for the key is the first multi-level lookup unit;
a second multi-level lookup unit which receives the key and provides a second result through a sequential multi-level search for the longest prefix match for the key in the second multi-level lookup unit; and means for providing a final result stored in only one of the multi-level lookup units for the key received in parallel by both the first mult-level lookup unit end the second multi-level lookup unit dependent on the first result and the second result
16. The look-up table as claimed in Claim 15 wherein the final results for the key are distributed amongst multi-level the lookup units dependent on the total number of final results to be stored in the lookup table.
17. The lookup table as claimed is Claim 15 wherein the key is an IP address and the final result is a route index.
18. The lookup table as claimed in Claim 15 wherein a first memory location is the first multi-level lookup emit indexed by the key stores the result and a second multi-level memory location in the second lookup unit indexed by the key stores an indication that the result is stored in the first multi-level lookup unit
19. The lookup table as claimed in Claim 15 wherein each of the first multi-level lookup unit and the second lookup unit includes a plurality of mapper levels and each of the mapper levels include memory locations for storing.
20. The lookup table as claimed in Claim 19 wherein the final result corresponding to the key stored in a location in one of the plurality of mapper levels is the first multi-level lookup unit is forwarded to the second multi-level lookup unit and provided by the second multi-level lookup unit.
21. The lookup table as claimed in Claim 15 wherein the final result corresponding to the key is provided by the multi-level lookup unit in which the final result is stored.
CA 2393764 1999-12-10 2000-12-08 Method and apparatus for longest match address lookup Abandoned CA2393764A1 (en)

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
US17023299P 1999-12-10 1999-12-10
US60/170,232 1999-12-10
US21296600P 2000-06-21 2000-06-21
US60/212,966 2000-06-21
PCT/CA2000/001441 WO2001043400A2 (en) 1999-12-10 2000-12-08 Method and apparatus for longest match address lookup

Publications (1)

Publication Number Publication Date
CA2393764A1 true CA2393764A1 (en) 2001-06-14

Family

ID=26865861

Family Applications (4)

Application Number Title Priority Date Filing Date
CA 2393760 Expired - Fee Related CA2393760C (en) 1999-12-10 2000-12-08 Method and apparatus for longest match address lookup
CA 2395151 Expired - Fee Related CA2395151C (en) 1999-12-10 2000-12-08 Method and apparatus for longest match address lookup
CA 2393764 Abandoned CA2393764A1 (en) 1999-12-10 2000-12-08 Method and apparatus for longest match address lookup
CA 2397608 Expired - Fee Related CA2397608C (en) 1999-12-10 2000-12-08 Method and apparatus for longest match address lookup

Family Applications Before (2)

Application Number Title Priority Date Filing Date
CA 2393760 Expired - Fee Related CA2393760C (en) 1999-12-10 2000-12-08 Method and apparatus for longest match address lookup
CA 2395151 Expired - Fee Related CA2395151C (en) 1999-12-10 2000-12-08 Method and apparatus for longest match address lookup

Family Applications After (1)

Application Number Title Priority Date Filing Date
CA 2397608 Expired - Fee Related CA2397608C (en) 1999-12-10 2000-12-08 Method and apparatus for longest match address lookup

Country Status (10)

Country Link
US (5) US7423981B2 (en)
EP (4) EP1250775A2 (en)
JP (4) JP4741134B2 (en)
KR (4) KR100748772B1 (en)
CN (5) CN1435030A (en)
AU (4) AU2133701A (en)
CA (4) CA2393760C (en)
DE (4) DE10085390T1 (en)
GB (4) GB2373084A (en)
WO (4) WO2001043370A2 (en)

Families Citing this family (126)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4156112B2 (en) * 1998-12-25 2008-09-24 富士通株式会社 High-speed search method and high-speed search device
KR100748772B1 (en) * 1999-12-10 2007-08-13 모사이드 테크놀로지스 코포레이션 Method and apparatus for longest match address lookup
US6996808B1 (en) * 2000-02-12 2006-02-07 Microsoft Corporation Function injector
US6675163B1 (en) * 2000-04-06 2004-01-06 International Business Machines Corporation Full match (FM) search algorithm implementation for a network processor
US7089240B2 (en) * 2000-04-06 2006-08-08 International Business Machines Corporation Longest prefix match lookup using hash function
US6947931B1 (en) * 2000-04-06 2005-09-20 International Business Machines Corporation Longest prefix match (LPM) algorithm implementation for a network processor
US6826561B2 (en) * 2000-05-22 2004-11-30 Broadcom Corporation Method and apparatus for performing a binary search on an expanded tree
US7047317B1 (en) * 2000-06-14 2006-05-16 Altera Corporation High performance network address processor system
US6555511B2 (en) * 2000-06-19 2003-04-29 Lance L. Renfrow Stable hydrotropic surfactants comprising alkylamino propionate
US7966421B2 (en) * 2000-06-21 2011-06-21 SAtech Group, A.B. Limited Liability Company Method and apparatus for logically expanding the length of a search key
US7111071B1 (en) * 2000-06-29 2006-09-19 Intel Corporation Longest prefix match for IP routers
JP2002026973A (en) * 2000-07-12 2002-01-25 Nec Corp Path retrieval system and its method, and router used for it
US8161081B2 (en) 2001-03-16 2012-04-17 Michael Philip Kaufman System and method for generating automatic user interface for arbitrarily complex or large databases
KR100488414B1 (en) * 2000-12-30 2005-05-11 한국전자통신연구원 Node Structuring Method for multiway search tree, and Searching Method by using it
US6826669B1 (en) * 2001-05-08 2004-11-30 Lewiz Communications Multi-protocol memory lookup system and method
US6917954B2 (en) * 2001-05-30 2005-07-12 Mosaid Technologies, Inc. Load balancing in IP address lookup
US7062493B1 (en) * 2001-07-03 2006-06-13 Trilogy Software, Inc. Efficient technique for matching hierarchies of arbitrary size and structure without regard to ordering of elements
JP3845845B2 (en) * 2001-07-24 2006-11-15 株式会社テルミナス・テクノロジー Associative memory system and network equipment and network system
US20030031179A1 (en) * 2001-08-08 2003-02-13 Jintae Oh Self-updateable longest prefix matching method and apparatus
US7143382B2 (en) * 2001-08-23 2006-11-28 Cadence Design Systems, Inc. Method and apparatus for storing routes
US6665202B2 (en) 2001-09-25 2003-12-16 Integrated Device Technology, Inc. Content addressable memory (CAM) devices that can identify highest priority matches in non-sectored CAM arrays and methods of operating same
US20030140052A1 (en) 2001-12-18 2003-07-24 Shawn Thomas Method and system for asset transition quality control
US6691171B1 (en) * 2002-02-01 2004-02-10 Micrel, Inc. Method and system for address lookup in data communication
US7274698B2 (en) * 2002-03-15 2007-09-25 Broadcom Corporation Multilevel parser for conditional flow detection in a network device
US7764687B1 (en) * 2002-03-28 2010-07-27 Meriton Networks Us Inc. Longest prefix match search technique
US20030210654A1 (en) * 2002-05-07 2003-11-13 Dongguk University Method of parallel packet switching
US7325071B2 (en) * 2002-06-04 2008-01-29 Lucent Technologies Inc. Forwarding traffic in a network using a single forwarding table that includes forwarding information related to a plurality of logical networks
US6925464B2 (en) * 2002-06-13 2005-08-02 Intel Corporation Method and system for performing inserts and lookups in memory
JP4048861B2 (en) * 2002-07-23 2008-02-20 日本電気株式会社 Address search device
US6934252B2 (en) * 2002-09-16 2005-08-23 North Carolina State University Methods and systems for fast binary network address lookups using parent node information stored in routing table entries
US7346009B2 (en) * 2002-09-30 2008-03-18 Mosaid Technologies, Inc. Dense mode coding scheme
US7782853B2 (en) * 2002-12-06 2010-08-24 Stmicroelectronics, Inc. Apparatus and method of using fully configurable memory, multi-stage pipeline logic and an embedded processor to implement multi-bit trie algorithmic network search engine
US8018935B2 (en) * 2002-12-12 2011-09-13 Intel Corporation Address search
US7890633B2 (en) * 2003-02-13 2011-02-15 Oracle America, Inc. System and method of extending virtual address resolution for mapping networks
US7443854B2 (en) * 2003-04-02 2008-10-28 Intel Corporation Methods and apparatus to route packets in a policy driven networked environment
US9014195B2 (en) * 2003-05-02 2015-04-21 Vitesse Semiconductor Corporation Packet forwarding method and system
US7260675B1 (en) 2003-07-15 2007-08-21 Integrated Device Technology, Inc. CAM-based search engines that support pipelined multi-database search operations using encoded multi-database identifiers
US7058757B1 (en) * 2003-07-15 2006-06-06 Integrated Device Technology, Inc. Content addressable memory (CAM) devices that support distributed CAM control and methods of operating same
US7840696B2 (en) * 2003-07-25 2010-11-23 Broadcom Corporation Apparatus and method for classifier identification
US7299227B2 (en) * 2003-09-09 2007-11-20 Stmicroelectronics, Inc. Method and system for providing cascaded trie-based network packet search engines
US7028106B2 (en) * 2003-12-05 2006-04-11 Hewlett-Packard Development Company, L.P. Remapping routing information entries in an expander
US8868745B1 (en) * 2003-12-22 2014-10-21 Avaya Inc. Method and system for providing configurable route table limits in a service provider for managing VPN resource usage
US7937413B2 (en) * 2004-05-04 2011-05-03 International Business Machines Corporation Self-adaptive prefix encoding for stable node identifiers
US7483426B2 (en) * 2004-05-13 2009-01-27 Micrel, Inc. Look-up table expansion method
BRPI0513236A (en) * 2004-07-12 2008-04-29 Fexco direct currency conversion
US7653619B1 (en) 2004-07-23 2010-01-26 Netlogic Microsystems, Inc. Integrated search engine devices having pipelined search and tree maintenance sub-engines therein that support variable tree height
US7725450B1 (en) * 2004-07-23 2010-05-25 Netlogic Microsystems, Inc. Integrated search engine devices having pipelined search and tree maintenance sub-engines therein that maintain search coherence during multi-cycle update operations
US7603346B1 (en) 2004-07-23 2009-10-13 Netlogic Microsystems, Inc. Integrated search engine devices having pipelined search and b-tree maintenance sub-engines therein
US7747599B1 (en) 2004-07-23 2010-06-29 Netlogic Microsystems, Inc. Integrated search engine devices that utilize hierarchical memories containing b-trees and span prefix masks to support longest prefix match search operations
US8886677B1 (en) 2004-07-23 2014-11-11 Netlogic Microsystems, Inc. Integrated search engine devices that support LPM search operations using span prefix masks that encode key prefix length
CN100366023C (en) * 2004-07-28 2008-01-30 国家数字交换系统工程技术研究中心 Device for searching route
KR100735577B1 (en) 2004-08-12 2007-07-04 삼성전자주식회사 Apparatus and method for adaptively searching security key in wireless network
US7590130B2 (en) * 2004-12-22 2009-09-15 Exar Corporation Communications system with first and second scan tables
US20060133383A1 (en) * 2004-12-22 2006-06-22 Russell Homer Communications system with scan table identification
US7590119B2 (en) * 2005-01-27 2009-09-15 Cisco Technology, Inc. Method and apparatus for context-based prefix updates in border gateway protocol
US20060187923A1 (en) * 2005-02-18 2006-08-24 Broadcom Corporation Dynamic filter processor key generation based on packet type
US7876747B2 (en) * 2005-03-04 2011-01-25 Cisco Technology, Inc. Method and apparatus for efficient load distribution on link aggregations
US7839854B2 (en) * 2005-03-08 2010-11-23 Thomas Alexander System and method for a fast, programmable packet processing system
US7529726B2 (en) * 2005-08-22 2009-05-05 International Business Machines Corporation XML sub-document versioning method in XML databases using record storages
US8543614B2 (en) * 2005-08-22 2013-09-24 International Business Machines Corporation Packing nodes into records to store XML XQuery data model and other hierarchically structured data
US8572125B2 (en) 2005-08-22 2013-10-29 International Business Machines Corporation Scalable storage schemes for native XML column data of relational tables
CN100401723C (en) * 2005-10-13 2008-07-09 华为技术有限公司 Fast indexing method
US7869411B2 (en) * 2005-11-21 2011-01-11 Broadcom Corporation Compact packet operation device and method
US7889676B1 (en) * 2006-04-13 2011-02-15 Infoblox Inc. Systems and methods for storing and retrieving data
CN100421114C (en) * 2006-04-21 2008-09-24 华为技术有限公司 Data matching inquiry method based on key words
US7817629B2 (en) * 2006-05-12 2010-10-19 Agere Systems Inc. Methods and apparatus for performing network operations on packets of data in response to content of particular user-specified protocol header fields
US7697518B1 (en) 2006-09-15 2010-04-13 Netlogic Microsystems, Inc. Integrated search engine devices and methods of updating same using node splitting and merging operations
US8194664B2 (en) * 2006-10-10 2012-06-05 Cisco Technology, Inc. Two-level load-balancing of network traffic over an MPLS network
US8189585B2 (en) * 2006-10-10 2012-05-29 Cisco Technology, Inc. Techniques for virtual private network fast convergence
US7937541B2 (en) * 2006-10-12 2011-05-03 International Business Machines Corporation Speed selective table scan operation
US8364803B2 (en) * 2006-10-20 2013-01-29 Oricane Ab Method, device, computer program product and system for representing a partition of N W-bit intervals associated to D-bit data in a data communications network
SE531557C2 (en) * 2006-10-20 2009-05-19 Oricane Ab Method, classifier, computer program product and system for representing a portion of n w-bit intervals associated with d-bit data in a data communication network
US8086641B1 (en) 2006-11-27 2011-12-27 Netlogic Microsystems, Inc. Integrated search engine devices that utilize SPM-linked bit maps to reduce handle memory duplication and methods of operating same
US7953721B1 (en) 2006-11-27 2011-05-31 Netlogic Microsystems, Inc. Integrated search engine devices that support database key dumping and methods of operating same
US7831626B1 (en) 2006-11-27 2010-11-09 Netlogic Microsystems, Inc. Integrated search engine devices having a plurality of multi-way trees of search keys therein that share a common root node
US7987205B1 (en) * 2006-11-27 2011-07-26 Netlogic Microsystems, Inc. Integrated search engine devices having pipelined node maintenance sub-engines therein that support database flush operations
CN100531140C (en) * 2006-12-20 2009-08-19 中国科学院计算技术研究所 A method and device for maximum prefix matching searching without retrospection
US11095583B2 (en) 2007-06-28 2021-08-17 Voxer Ip Llc Real-time messaging method and apparatus
US8645477B2 (en) 2009-01-30 2014-02-04 Voxer Ip Llc Progressive messaging apparatus and method capable of supporting near real-time communication
US9178916B2 (en) 2007-06-28 2015-11-03 Voxer Ip Llc Real-time messaging method and apparatus
US8825772B2 (en) * 2007-06-28 2014-09-02 Voxer Ip Llc System and method for operating a server for real-time communication of time-based media
US8688789B2 (en) 2009-01-30 2014-04-01 Voxer Ip Llc Progressive messaging apparatus and method capable of supporting near real-time communication
US8180029B2 (en) 2007-06-28 2012-05-15 Voxer Ip Llc Telecommunication and multimedia management method and apparatus
US20110019662A1 (en) 2007-06-28 2011-01-27 Rebelvox Llc Method for downloading and using a communication application through a web browser
US20100198922A1 (en) 2009-01-30 2010-08-05 Rebelvox Llc Methods for using the addressing, protocols and the infrastructure of email to support near real-time communication
US7886176B1 (en) 2007-09-24 2011-02-08 Integrated Device Technology, Inc. DDR memory system for measuring a clock signal by identifying a delay value corresponding to a changed logic state during clock signal transitions
US20090100219A1 (en) * 2007-10-15 2009-04-16 Anurag Bhargava Method and apparatus for efficient cam lookup for internet protocol addresses
CN101420415B (en) * 2007-10-23 2012-08-22 华为技术有限公司 Method and apparatus for forming routing table
US7716204B1 (en) 2007-12-21 2010-05-11 Netlogic Microsystems, Inc. Handle allocation managers and methods for integated circuit search engine devices
TWI413910B (en) * 2008-01-25 2013-11-01 Univ Nat Taiwan Data item interval indentifier lookup method and system
CN101505262B (en) 2008-02-05 2011-07-20 华为技术有限公司 Method and apparatus for establishing node Id
US8335780B2 (en) * 2008-03-11 2012-12-18 James Madison Kelley Scalable high speed relational processor for databases and networks
US7801877B1 (en) 2008-04-14 2010-09-21 Netlogic Microsystems, Inc. Handle memory access managers and methods for integrated circuit search engine devices
US7933885B1 (en) 2008-04-25 2011-04-26 Netlogic Microsystems, Inc. Longest matching prefix search engine with hierarchical decoders
US8036141B2 (en) * 2008-08-15 2011-10-11 At&T Intellectual Property I, L.P Apparatus and method for managing a network
DE102008059204B9 (en) * 2008-11-27 2011-05-05 Infineon Technologies Ag Method for searching for a slave node in a communication network, master node and slave node for a communication network
US8849927B2 (en) 2009-01-30 2014-09-30 Voxer Ip Llc Method for implementing real-time voice messaging on a server node
US20100245072A1 (en) * 2009-03-25 2010-09-30 Syclipse Technologies, Inc. System and method for providing remote monitoring services
US20100246669A1 (en) * 2009-03-25 2010-09-30 Syclipse Technologies, Inc. System and method for bandwidth optimization in data transmission using a surveillance device
US20100245582A1 (en) * 2009-03-25 2010-09-30 Syclipse Technologies, Inc. System and method of remote surveillance and applications therefor
US9049157B1 (en) * 2009-08-16 2015-06-02 Compass Electro-Optical Systems Ltd Method and device for improving scalability of longest prefix match
US8363662B2 (en) * 2010-03-19 2013-01-29 Cisco Technology, Inc. Alternate down paths for directed acyclic graph (DAG) routing
JP5440691B2 (en) * 2010-03-24 2014-03-12 日本電気株式会社 Packet transfer system, control device, transfer device, processing rule creation method and program
US8751687B2 (en) 2010-04-30 2014-06-10 Microsoft Corporation Efficient encoding of structured data
US20130054565A1 (en) * 2010-05-04 2013-02-28 Mingoa Limited Identification and verification in communication systems
US9077669B2 (en) * 2010-06-14 2015-07-07 Dynamic Invention Llc Efficient lookup methods for ternary content addressable memory and associated devices and systems
US20140185614A1 (en) * 2012-12-28 2014-07-03 General Instrument Corporation Multiple domain addressing in message routing
US9331942B2 (en) * 2013-03-12 2016-05-03 Xpliant, Inc. Apparatus and method for processing alternately configured longest prefix match tables
CN104052669B (en) * 2013-03-12 2018-12-07 凯为公司 For handling the device for the longest prefix match table being alternately arranged
CN104426770A (en) * 2013-09-09 2015-03-18 中兴通讯股份有限公司 Routing lookup method, routing lookup device and method for constructing B-Tree tree structure
JP5950285B2 (en) * 2013-12-19 2016-07-13 インターナショナル・ビジネス・マシーンズ・コーポレーションInternational Business Machines Corporation A method for searching a tree using an instruction that operates on data having a plurality of predetermined bit widths, a computer for searching a tree using the instruction, and a computer thereof program
US9485179B2 (en) 2014-11-13 2016-11-01 Cavium, Inc. Apparatus and method for scalable and flexible table search in a network switch
US9553829B2 (en) 2014-11-13 2017-01-24 Cavium, Inc. Apparatus and method for fast search table update in a network switch
US10091056B1 (en) * 2015-08-06 2018-10-02 Amazon Technologies, Inc. Distribution of modular router configuration
US10419282B1 (en) 2015-09-24 2019-09-17 Amazon Technologies, Inc. Self-configuring network devices
US9584155B1 (en) 2015-09-24 2017-02-28 Intel Corporation Look-ahead hash chain matching for data compression
US10313231B1 (en) * 2016-02-08 2019-06-04 Barefoot Networks, Inc. Resilient hashing for forwarding packets
US10063407B1 (en) 2016-02-08 2018-08-28 Barefoot Networks, Inc. Identifying and marking failed egress links in data plane
US10521411B2 (en) * 2016-08-10 2019-12-31 Moonshadow Mobile, Inc. Systems, methods, and data structures for high-speed searching or filtering of large datasets
CN106549872B (en) * 2016-10-31 2019-05-24 西安空间无线电技术研究所 Longest prefix match with accurately match the spaceborne fast routing lookups system combined
US10404619B1 (en) 2017-03-05 2019-09-03 Barefoot Networks, Inc. Link aggregation group failover for multicast
US10237206B1 (en) 2017-03-05 2019-03-19 Barefoot Networks, Inc. Equal cost multiple path group failover for multicast
CN107169054A (en) * 2017-04-26 2017-09-15 四川长虹电器股份有限公司 Ip indexing means based on prefix forest
US10224957B1 (en) 2017-11-27 2019-03-05 Intel Corporation Hash-based data matching enhanced with backward matching for data compression
JP7088868B2 (en) * 2019-03-22 2022-06-21 エヌ・ティ・ティ・コミュニケーションズ株式会社 Communication equipment, communication methods and programs
CN111367665B (en) * 2020-02-28 2020-12-18 清华大学 Parallel communication route establishing method and system

Family Cites Families (47)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4450525A (en) * 1981-12-07 1984-05-22 Ibm Corporation Control unit for a functional processor
US4661658A (en) * 1985-02-12 1987-04-28 International Business Machines Corporation Offline PIN validation with DES
AU620994B2 (en) * 1989-07-12 1992-02-27 Digital Equipment Corporation Compressed prefix matching database searching
US5202986A (en) * 1989-09-28 1993-04-13 Bull Hn Information Systems Inc. Prefix search tree partial key branching
US5261088A (en) * 1990-04-26 1993-11-09 International Business Machines Corporation Managing locality in space reuse in a shadow written B-tree via interior node free space list
JPH0522297A (en) * 1991-07-09 1993-01-29 Nec Corp Communication network system
US5359724A (en) * 1992-03-30 1994-10-25 Arbor Software Corporation Method and apparatus for storing and retrieving multi-dimensional data in computer memory
US5329618A (en) * 1992-04-02 1994-07-12 Fibronics Ltd. Look-up table for a bridge in a data communications system
EP0606461B1 (en) * 1992-07-24 1999-11-24 Microsoft Corporation Computer method and system for allocating and freeing memory
US5323389A (en) * 1992-08-14 1994-06-21 Fore Systems, Inc. ATM cell interface and method for dispatching an ATM cell
US5386413A (en) * 1993-03-19 1995-01-31 Bell Communications Research, Inc. Fast multilevel hierarchical routing table lookup using content addressable memory
CA2643234C (en) * 1993-10-29 2012-05-15 Microsoft Corporation Method and system for generating a computer program
US5384568A (en) * 1993-12-02 1995-01-24 Bell Communications Research, Inc. Data compression
US5854196A (en) * 1994-08-23 1998-12-29 The Procter & Gamble Company Detergent compositions
US5754854A (en) * 1994-11-14 1998-05-19 Microsoft Corporation Method and system for providing a group of parallel resources as a proxy for a single shared resource
EP0829158A1 (en) * 1995-05-18 1998-03-18 Nortel Networks Corporation Telephony based delivery system of messages containing selected greetings
US5699276A (en) 1995-12-15 1997-12-16 Roos; Charles E. Utility meter providing an interface between a digital network and home electronics
US5857196A (en) * 1996-07-19 1999-01-05 Bay Networks, Inc. Method for storing a tree of potential keys in a sparse table
JP3520709B2 (en) * 1997-03-13 2004-04-19 三菱電機株式会社 Network address search method
US6034958A (en) * 1997-07-11 2000-03-07 Telefonaktiebolaget Lm Ericsson VP/VC lookup function
US5946679A (en) * 1997-07-31 1999-08-31 Torrent Networking Technologies, Corp. System and method for locating a route in a route table using hashing and compressed radix tree searching
US6512766B2 (en) * 1997-08-22 2003-01-28 Cisco Systems, Inc. Enhanced internet packet routing lookup
JPH1166096A (en) * 1997-08-25 1999-03-09 Nippon Telegr & Teleph Corp <Ntt> Data storage method, database stored by the same and retrieval method for the database
US6553002B1 (en) * 1997-08-29 2003-04-22 Ascend Communications, Inc. Apparatus and method for routing data packets through a communications network
EA200000328A1 (en) * 1997-09-15 2000-10-30 Эффнет Груп Аб METHOD AND SYSTEM FOR QUICK ROUTING SEARCHES
US6266706B1 (en) 1997-09-15 2001-07-24 Effnet Group Ab Fast routing lookup system using complete prefix tree, bit vector, and pointers in a routing table for determining where to route IP datagrams
US6141655A (en) * 1997-09-23 2000-10-31 At&T Corp Method and apparatus for optimizing and structuring data by designing a cube forest data structure for hierarchically split cube forest template
US6563823B1 (en) * 1997-10-30 2003-05-13 Marconi Communications, Inc. Multi-resolution tree for longest match address lookups
US6223172B1 (en) * 1997-10-31 2001-04-24 Nortel Networks Limited Address routing using address-sensitive mask decimation scheme
JP3186681B2 (en) * 1997-12-25 2001-07-11 日本電気株式会社 Route search circuit and communication control device
JP3570606B2 (en) * 1998-02-12 2004-09-29 日本電信電話株式会社 Data retrieval apparatus and method
JP3166700B2 (en) * 1998-03-12 2001-05-14 日本電気株式会社 Router and longest match search device
US6085188A (en) * 1998-03-30 2000-07-04 International Business Machines Corporation Method of hierarchical LDAP searching with relational tables
WO1999050963A2 (en) * 1998-03-31 1999-10-07 Samsung Electronics Co., Ltd. TURBO ENCODING/DECODING DEVICE AND METHOD FOR PROCESSING FRAME DATA ACCORDING TO QoS
US6067574A (en) * 1998-05-18 2000-05-23 Lucent Technologies Inc High speed routing using compressed tree process
US6247014B1 (en) * 1998-07-01 2001-06-12 Nortel Networks Limited Method and apparatus for performing hash lookups using valid bit tables with pointers
US6526055B1 (en) * 1998-10-20 2003-02-25 Sun Microsystems, Inc. Method and apparatus for longest prefix address lookup
IT1305140B1 (en) * 1998-10-27 2001-04-10 Cselt Centro Studi Lab Telecom MEMORY FOR SEARCHING FOR INFORMATION THROUGH ANALYSIS OF INPARTICULAR PREFIXES FOR THE CREATION OF ROAD TABLES IN KNOTS
TW468116B (en) * 1999-02-08 2001-12-11 Wen-Shian Chen High speed Internet protocol address lookups method for saving memory
US6460112B1 (en) * 1999-02-23 2002-10-01 Netlogic Microsystems, Llc Method and apparatus for determining a longest prefix match in a content addressable memory device
US6678274B1 (en) * 1999-07-30 2004-01-13 Riverstone Networks, Inc. Method and system for managing forwarding tables
EP1083768A1 (en) * 1999-09-08 2001-03-14 TELEFONAKTIEBOLAGET LM ERICSSON (publ) A method for facilitating data transmission
US6744775B1 (en) * 1999-09-27 2004-06-01 Nortel Networks Limited State information and routing table updates in large scale data networks
US6687247B1 (en) * 1999-10-27 2004-02-03 Cisco Technology, Inc. Architecture for high speed class of service enabled linecard
KR100748772B1 (en) * 1999-12-10 2007-08-13 모사이드 테크놀로지스 코포레이션 Method and apparatus for longest match address lookup
US6826561B2 (en) * 2000-05-22 2004-11-30 Broadcom Corporation Method and apparatus for performing a binary search on an expanded tree
US7106732B2 (en) * 2000-12-27 2006-09-12 Mosaid Technologies Incorporated Default route coding

Also Published As

Publication number Publication date
GB0213387D0 (en) 2002-07-24
JP4741134B2 (en) 2011-08-03
WO2001043370A2 (en) 2001-06-14
CN101510839A (en) 2009-08-19
JP2003516661A (en) 2003-05-13
GB0213390D0 (en) 2002-07-24
CA2397608A1 (en) 2001-06-14
WO2001043345A3 (en) 2001-11-29
GB2373083A (en) 2002-09-11
CA2393760C (en) 2010-07-13
KR20020081681A (en) 2002-10-30
CN1435032A (en) 2003-08-06
EP1250775A2 (en) 2002-10-23
WO2001043345A2 (en) 2001-06-14
US6836771B2 (en) 2004-12-28
AU2334101A (en) 2001-06-18
US20010042130A1 (en) 2001-11-15
AU2133801A (en) 2001-06-18
WO2001043370A3 (en) 2001-11-08
AU2133701A (en) 2001-06-18
CA2395151C (en) 2008-07-22
WO2001043400A2 (en) 2001-06-14
CA2393760A1 (en) 2001-06-14
KR20020081680A (en) 2002-10-30
DE10085388T1 (en) 2003-07-17
KR100748773B1 (en) 2007-08-13
US6691218B2 (en) 2004-02-10
GB2373082A (en) 2002-09-11
CN1435030A (en) 2003-08-06
CN1174587C (en) 2004-11-03
US7423981B2 (en) 2008-09-09
JP2003516666A (en) 2003-05-13
DE10085390T1 (en) 2003-04-30
KR100748771B1 (en) 2007-08-13
CA2395151A1 (en) 2001-06-14
US6539369B2 (en) 2003-03-25
GB2373082B (en) 2004-03-17
EP1250779A2 (en) 2002-10-23
CN1278525C (en) 2006-10-04
US20030126113A1 (en) 2003-07-03
DE10085389T1 (en) 2003-07-17
CN1434950A (en) 2003-08-06
WO2001043346A3 (en) 2001-11-29
EP1250778A2 (en) 2002-10-23
CN100432991C (en) 2008-11-12
GB0213389D0 (en) 2002-07-24
US20010044876A1 (en) 2001-11-22
CN101510839B (en) 2014-07-09
GB2374174A (en) 2002-10-09
KR20020082465A (en) 2002-10-31
GB2373083B (en) 2004-03-17
WO2001043400A3 (en) 2001-11-08
JP4565793B2 (en) 2010-10-20
GB0213391D0 (en) 2002-07-24
JP2003516670A (en) 2003-05-13
JP2003516660A (en) 2003-05-13
GB2373084A (en) 2002-09-11
US20020116526A1 (en) 2002-08-22
KR100748772B1 (en) 2007-08-13
CA2397608C (en) 2010-08-24
GB2374174B (en) 2004-03-03
AU2133901A (en) 2001-06-18
KR20020081679A (en) 2002-10-30
DE10085387T5 (en) 2004-07-01
CN1435031A (en) 2003-08-06
US20010043602A1 (en) 2001-11-22
WO2001043346A2 (en) 2001-06-14
EP1250662A2 (en) 2002-10-23

Similar Documents

Publication Publication Date Title
CA2393760C (en) Method and apparatus for longest match address lookup
US7633960B2 (en) Dense mode coding scheme
US7966421B2 (en) Method and apparatus for logically expanding the length of a search key
US7715385B2 (en) Default route coding
US6917954B2 (en) Load balancing in IP address lookup

Legal Events

Date Code Title Description
EEER Examination request
FZDE Discontinued