US20140233582A1 - Semaphore soft and hard hybrid architecture - Google Patents
Semaphore soft and hard hybrid architecture Download PDFInfo
- Publication number
- US20140233582A1 US20140233582A1 US13/973,744 US201313973744A US2014233582A1 US 20140233582 A1 US20140233582 A1 US 20140233582A1 US 201313973744 A US201313973744 A US 201313973744A US 2014233582 A1 US2014233582 A1 US 2014233582A1
- Authority
- US
- United States
- Prior art keywords
- packet
- resource
- processing
- remote
- processing stage
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L45/00—Routing or path finding of packets in data switching networks
- H04L45/44—Distributed routing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L49/00—Packet switching elements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L45/00—Routing or path finding of packets in data switching networks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L45/00—Routing or path finding of packets in data switching networks
- H04L45/12—Shortest path evaluation
- H04L45/121—Shortest path evaluation by minimising delays
Definitions
- the present disclosure relates to a network device that processes packets.
- a pipeline-type packet processing device typically includes processing stages arranged to process packets in what is understood as a pipeline configuration.
- packets are processed in stages with packets progressing from stage to stage through the pipeline.
- respective stages are programmable.
- a stage in the packet processing pipeline receives a packet that had been processed at one or more previous stages, along with a processing context for the packet.
- Various resources such as for example various table lookups, that are to be used by the respective stages in a packet processing pipeline typically are provided by units that are external to the packet processing pipeline and that are accessed by the stages when those services are needed to perform a processing operation.
- One or more embodiments of the disclosure generally relate to a network device that processes packets through a number of processing stages.
- Some packet processing operations update, i.e., read and write a value in a remote resource such as a memory location in a table. This situation is also referred to as “read-modify-write.”
- read-modify-write This situation is also referred to as “read-modify-write.”
- the remote resource when the processing of a given packet involves packet processing operations that will read from and subsequently write to a given remote resource, the remote resource is allocated or “locked” so that the processing of other packets cannot interfere with the remote resource until the processing carried out for the given packet reaches a point at which the remote resource is suitably released.
- Semaphores are used in example embodiments to lock corresponding remote resources. Notwithstanding the foregoing, not every example embodiment is required to possess all or even any of the features mentioned in this paragraph.
- a packet processing device having processing stages, including a first processing stage and a second processing stage arranged as a packet processing pipeline; the first processing stage and the second processing stage each have: a respective processor configured to process a packet of a packet stream, and a respective resource manager having a respective local resource lock corresponding to a remote resource; the respective processor is configured to request the respective resource manager to allocate the remote resource; the respective resource manager is further configured to respond to the request to allocate the remote resource by locking the remote resource with the respective local resource lock and allocating the remote resource; the respective processor is further configured to implement a packet processing operation associated with the allocated remote resource.
- a packet processing device includes: processing stages arranged as a packet processing pipeline; the processing stages each having processor cores and buffers; the processor cores and buffers of the processing stages defining a plurality of paths, for simultaneous packet processing, through the packet processing pipeline; an ingress front end configured to direct each packet of an incoming stream of packets into one of the plurality of paths; the paths including a hard path and a soft path, the hard path being configured to process received ones of the incoming stream of packets with fixed latency, the soft path being configured to process received ones of the incoming stream of packets with variable latency; and the processing stages each further including a respective resource manager configured to request allocation of a remote resource, for a given packet of the incoming stream of packets, in response to an instruction from one of the processor cores.
- the respective resource manager is further configured to allocate an available remote resource, thereby making the remote resource accessible only to the respective packet and to subsequent packet processing operations for that packet.
- the processing stages are further configured, in an example embodiment, to both request the release of a allocated resource, and to receive from another processing stage a request to release an allocated resource and subsequently to release the resource. A remote resource which has been released is available to be allocated.
- a packet processing method includes receiving, at a processor of a first processing stage, a first packet and a request for allocation of a remote resource; responding, by the processor, to the allocation request, by setting a semaphore corresponding to the remote resource to indicate a locked status; implementing a first packet processing operation, in association with the allocated remote resource, and in association with the first packet, to obtain a processed first packet; and outputting the processed first packet to a next processing stage of the pipeline packet processing device.
- FIG. 1A is a highly simplified illustrative drawing showing a concept of operation according to example embodiments.
- FIG. 1B is a highly simplified illustrative drawing showing another concept of operation according to example embodiments.
- FIG. 2 is a highly simplified schematic diagram of a device according to example embodiments.
- FIG. 3 is a highly simplified schematic diagram showing a particular aspect of a device according to example embodiments.
- FIG. 4 is a flow diagram of an example method according to example embodiments.
- FIG. 5 is a flow diagram of an example method according to example embodiments.
- FIG. 6 is a flow diagram of an example method according to example embodiments.
- FIG. 7 is a highly simplified schematic diagram of a device according to example embodiments.
- FIG. 8 is a highly simplified schematic diagram of a device according to example embodiments.
- FIG. 1A shows a Packet Processing Device ( 1000 ).
- Packet Processing Device ( 1000 ) includes a Plurality of Processing Stages ( 60 ), including a First Processing Stage ( 60 - 0 ), a Second Processing Stage ( 60 - 1 ), and other processing stages in accordance with particular example embodiments.
- the terms “first” and “second” do not connote that the First Processing Stage ( 60 - 0 ) is limited to being the first processing stage in a pipeline configuration, nor that the Second Processing Stage ( 60 - 1 ) is limited to being the second processing stage in the pipeline configuration.
- first and “second” simply indicate a relative order in the pipeline configuration such that the First Processing Stage ( 60 - 0 ) is relatively earlier along the pipeline configuration than is the Second Processing Stage ( 60 - 1 ).
- the First Processing Stage ( 60 - 0 ) and Second Processing Stage ( 60 - 1 ) are illustrated adjacent each other in the pipeline configuration, but this arrangement is merely an example embodiment, and other processing stages are interposed between the First Processing Stage ( 60 - 0 ) and the Second Processing Stage ( 60 - 1 ) in example embodiments.
- the First Processing Stage ( 60 - 0 ) is considered to be “upstream” with respect to the Second Processing Stage ( 60 - 1 ).
- the Second Processing Stage ( 60 - 1 ) is considered to be “downstream” with respect to the first processing stage, according to an example embodiment.
- the packet processing device ( 1000 ) is configured as a dataflow pipeline in which the respective stages ( 60 ) are programmable, and in which packets are received for processing at a downstream processing stage along with a processing context for processing the corresponding received packet.
- one or more individual stages of the Plurality of Processing Stages ( 60 ) are not programmable but have hard-coded functionality.
- the Plurality of Processing Stages ( 60 ) includes some programmable processing stages and some hard-coded processing stages.
- Each of the Plurality of Processing Stages ( 60 ) is substantially the same from a hardware perspective, according to an example embodiment.
- the Plurality of Processing Stages ( 60 ) includes different types of processing stages.
- the discussion below assumes that each of the Plurality of Processing Stages ( 60 ) has the same hardware structure. The components of each processing stage are described in more detail below. At the present, however, it is noted that each of the Plurality of Processing Stages ( 60 ) includes a respective Pool of Resource Locks ( 70 ).
- the respective Pool of Resource Locks ( 70 ) for the First Processing Stage ( 60 - 0 ) is shown as Pool of Resource Locks ( 70 - 0 ).
- the respective Pool of Resource Locks ( 70 ) for the Second Processing Stage ( 60 - 1 ) is shown as Pool of Resource Locks ( 70 - 1 ).
- Each of the Plurality of Processing Stages ( 60 ) has a respective Pool of Resource Locks ( 70 ), according to example embodiments.
- each respective Pool of Resource Locks ( 70 ) there exists a plurality of semaphores.
- the semaphores are represented as Local Resource Locks ( 50 , 51 , . . . 5 n ).
- the number of semaphores used depends on particular implementation details, and FIG. 1 shows only a few semaphores within a Pool of Resource Locks ( 70 ) for the sake of simplicity of illustration.
- Each respective Pool of Resource Locks ( 70 ) includes its own Local Resource Locks ( 50 , 51 , . . . 5 n ).
- the Pool of Resource Locks ( 70 - 0 ) for First Processing Stage ( 60 - 0 ) includes Respective Local Resource Lock ( 50 - 0 ), Respective Local Resource Lock ( 51 - 0 ), . . . and Respective Local Resource Lock ( 5 n - 0 ).
- the Pool of Resource Locks ( 70 - 1 ) for Second Processing Stage ( 60 - 1 ) includes Respective Local Resource Lock ( 50 - 1 ), Respective Local Resource Lock ( 51 - 1 ), . . . and Respective Local Resource Lock ( 5 n ⁇ 1).
- FIG. 1A Also shown in FIG. 1A are a Plurality of Remote Resources ( 85 ) including Remote Resource ( 85 - 0 ), Remote Resource ( 85 - 1 ), . . . and Remote Resource ( 85 - n ). Together, these represent remote resources that are available for READ and WRITE operations carried out by the Plurality of Processing Stages ( 60 ).
- the remote resources are memory locations that contain data such as, in a system for processing packets received on a network, forwarding tables, policy tables, MAC address learning, time-out management, and the like. This example is non-limiting, however, and the remote resources according to other exemplary embodiments contain data in structures other than tables.
- a remote resource contains a value which is accessed by either a READ or a WRITE Operation.
- a Processing Stage obtains a value stored in a memory location of the remote resource.
- the Processing Stage stores a value into the memory location of the remote resource. It is to be noted, however, that a remote resource may be locked by a process intended to subsequently perform alternate operations in addition to READ and WRITE operations, for example updating or synchronizing multiple tables and/or memory locations.
- a Processing Stage requires an Engine to access a Remote Resource.
- the Plurality of Processing Stages ( 60 ) are able to access the Plurality of Remote Resources ( 85 ) directly or without the use of an Engine.
- a given Remote Resource is accessed through a combination of engines.
- FIG. 1A there is illustrated a more detailed view of the Pool of Resource Locks ( 70 - 0 ) of the First Processing Stage ( 60 - 0 ), according to an example embodiment.
- This more detailed representation of the Pool of Resource Locks ( 70 - 0 ) is shown as a table containing two columns. Each row of the table represents a Local Resource Lock such as Respective Local Resource Lock ( 50 - 0 ), Respective Local Resource Lock ( 51 - 0 ), . . . Respective Local Resource Lock ( 5 n - 0 ).
- the left column of the table includes, in each row, an identifier of a corresponding one of the Plurality of Remote Resources ( 85 ).
- the right-column of the table indicates the status of the corresponding Remote Resource.
- the status is UNLOCKED or LOCKED.
- the table of FIG. 1A indicates that, in this example, the Remote Resource ( 85 - 0 ) is UNLOCKED, the Remote Resource ( 85 - 1 ) is LOCKED, and the Remote Resource ( 85 - n ) is UNLOCKED.
- the Processing Stages are connected to a Resource Release Bus ( 300 ), also referred to as a lock release message path.
- the Resource Release Bus ( 300 ) allows for communication, such as message exchange, among the Plurality of Processing Stages ( 60 ).
- the Resource Release Bus ( 300 ) allows for messages to be passed from upstream stages to downstream stages.
- the Resource Release Bus ( 300 ) is implemented in other ways according to example embodiments, and a bus that permits message exchange is a non-limiting example.
- the Resource Release Bus ( 300 ) does not carry the packets being processed through the pipeline configuration.
- a single Packet ( 20 - 0 ) is shown in FIG. 1A , for simplicity of illustration.
- the Packet ( 20 - 0 ) enters the First Processing Stage ( 60 - 0 ), and will pass to the Second Processing Stage ( 60 - 1 ) after any packet processing operations programmed into the First Processing Stage ( 60 - 0 ) are concluded.
- FIG. 1A also illustrates part of an overall concept of operation according to an example embodiment.
- 1 a - 1 indicates the entering of Packet ( 20 - 0 ) into the First Processing Stage ( 60 - 0 ).
- All locks of the Pool of Resource Locks ( 70 - 0 ) have an UNLOCKED status.
- the packet processing operations programmed into the First Processing Stage ( 60 - 0 ) include, for example, a READ of Remote Resource ( 85 - 1 ). Before carrying out the READ operation, however, the packet processing operations programmed into First Processing Stage ( 60 - 0 ) carry out a check to determine whether Remote Resource ( 85 - 1 ) is in a LOCKED or in an UNLOCKED status.
- the Pool of Resource Locks ( 70 - 0 ) is checked. More particularly, the particular Respective Local Resource Lock ( 51 - 0 ) corresponding to Remote Resource ( 85 - 1 ) is checked. At 1 a - 2 a , Finding the Respective Local Resource Lock ( 51 - 0 ) has an UNLOCKED status, the First Processing Stage ( 60 - 0 ) sets the Respective Local Resource Lock ( 51 - 0 ) to a LOCKED status.
- the First Processing Stage ( 60 - 0 ) carries out the READ operation with respect to Remote Resource ( 85 - 1 ) using an appropriate one of the Plurality of Engines ( 80 ).
- the value of Remote Resource ( 85 - 1 ) is thus read by the First Processing Stage ( 60 - 0 ) and this value is now available to use in accordance with the processing programmed into First Processing Stage ( 60 - 0 ) and others of the Plurality of Processing Stages ( 60 ).
- the Packet ( 20 - 0 ) is transferred to Second Processing Stage ( 60 - 1 ), together with other information to be described later.
- the setting of the status of Respective Local Resource Lock ( 51 - 0 ), corresponding to Remote Resource ( 85 - 1 ), from UNLOCKED to LOCKED implements a locking of the remote resource.
- the Respective Local Resource Lock ( 51 - 0 ) acts as a semaphore that indicates that Remote Resource ( 85 - 1 ) is an allocated remote resource.
- the mechanism for locking Remote Resource ( 85 - 1 ) and the mechanism for subsequent reading of the value of Remote Resource ( 85 - 1 ) are both located and activated locally within the First Processing Stage ( 60 - 0 ). If, for example, in contrast, semaphores were to be maintained in a shared pool of semaphores located outside the First Processing Stage ( 60 - 0 ), delays in making the determination as to the LOCKED or UNLOCKED status of Remote Resource ( 85 - 1 ) would likely result.
- one of the Plurality of Engines ( 80 ) would need to obtain LOCKED or UNLOCKED status information from such a shared pool of semaphores.
- the Lock & Read Operation ( 105 ) allocates the Remote Resource ( 85 - 1 ) by setting a value of Respective Local Resource Lock ( 51 - 0 ), thereby preventing First Processing Stage ( 60 - 0 ) from carrying out any operations with respect to Remote Resource ( 85 - 1 ) until the status of Remote Resource ( 85 - 1 ) in Respective Local Resource Lock ( 51 - 0 ) goes from LOCKED to UNLOCKED.
- the Lock & Read Operation ( 105 ) includes a READ operation with respect to the allocated Remote Resource ( 85 - 1 ).
- This READ operation is understood to constitute a packet processing operation associated with the allocated Remote Resource ( 85 - 1 ).
- Packet ( 20 - 0 ) completed processing at and exits from the First Processing Stage ( 60 - 0 ).
- This packet prior to undergoing the processing programmed into First Processing Stage ( 60 - 0 ), is understood to constitute a first packet.
- this packet is understood to constitute a processed first packet.
- the Packet ( 20 - 0 ) changes. Such changes, however, are not strictly required, and in some situations the Packet ( 20 - 0 ) stays the same. Therefore, it is noted that the processing programmed into First Processing Stage ( 60 - 0 ), performed with respect to the first packet, results, in any event, with the passing along of a processed first packet.
- Packet ( 20 - 1 ) represents a second packet.
- the Packet ( 20 - 1 ) is passed to Second Processing Stage ( 60 - 1 ).
- the Packet ( 20 - 1 ) is a packet which, in a prior processing stage, resulted in the locking of a remote resource such as Remote Resource ( 85 - 1 ).
- Packet ( 20 - 1 ) is the processed first packet which exits from First Processing Stage ( 60 - 0 ) at 1 a - 3 .
- One item of information in the Resource Lock Release Request is an identifier of the one of the Plurality of Processing Stages ( 60 ) that has the Respective Local Resource Lock ( 50 - 1 ) corresponding to the locked Remote Resource ( 85 - 1 ).
- the other item of information is an identifier of the particular local resource lock (here, Respective Local Resource Lock ( 50 - 1 )) to be released.
- the Resource Release Bus ( 300 ) Since the Resource Release Bus ( 300 ) needs to carry the Resource Lock Release Request only to upstream ones of the Plurality of Processing Stages ( 60 ), the Resource Release Bus ( 300 ) communicates in one direction, e.g. the upstream direction, in an example embodiment.
- the packet Packet ( 20 - 1 ) exits from the Second Processing Stage ( 60 - 1 ) as a processed second packet and enters into a subsequent one of the Plurality of Processing Stages ( 60 ).
- the First Processing Stage ( 60 - 0 ) receives the Resource Lock Release Request addressed to the First Processing Stage ( 60 - 0 ), and changes the status of Remote Resource ( 85 - 1 ) in Respective Local Resource Lock ( 51 - 0 ) of the Pool of Resource Locks ( 70 - 0 ) by replacing LOCKED with UNLOCKED (illustrated as “LOCKED ->UNLOCKED” in FIG. 1B ).
- the Remote Resource ( 85 - 1 ) is no longer locked, and is available to be allocated for use in a packet processing operation for another packet.
- Resource Lock Release Request is understood to constitute, more generally, a release request message.
- the status of one of the Plurality of Remote Resources ( 85 ) is represented in an alternative manner, such as a single binary digit with one value representing a LOCKED status, and the other value representing an UNLOCKED status.
- Other implementations of the Pool of Resource Locks ( 70 - 0 ) are within the ability of a person familiar with this field.
- the implementation of semaphores may be substituted with other locking mechanisms familiar to those skilled in the art including a single client lock such as a mutex or the like.
- local locks are used to restrict the accessibility of a remote resource to the processing for a single packet, along the respective processing stages.
- the First Processing Stage ( 60 - 0 ) reads from the Remote Resource ( 85 - 1 ) and passes the Packet ( 20 - 0 ) to the Second Processing Stage ( 60 - 1 ) as the Packet ( 20 - 1 ).
- the Second Processing Stage ( 60 - 1 ) then writes a value to the Remote Resource ( 85 - 1 ).
- the semaphore (Respective Local Resource Lock ( 50 - 1 )) allows exclusive access by the Second Processing Stage ( 60 - 1 ) for the purpose of writing.
- the Second Processing Stage ( 60 - 1 ) would not be guaranteed exclusive access to the Remote Resource ( 85 - 1 ). That is to say, the First Processing Stage ( 60 - 0 ) would have the opportunity, in response to a new packet requiring the same Remote Resource ( 85 - 1 ), to begin reading from the Remote Resource ( 85 - 1 ) while the Second Processing Stage ( 60 - 1 ) is beginning to write to the Remote Resource ( 85 - 1 ), resulting in possible problems due to such contention.
- the use of semaphores, in Packet Processing Device ( 1000 ), thus avoids such contention problems.
- FIG. 2 shows a more detailed view of the Packet Processing Device ( 1200 ).
- the Plurality of Processing Stages ( 60 ) includes First Processing Stage ( 60 - 0 ), Second Processing Stage ( 60 - 1 ), other processing stages, and Final Processing Stage ( 60 - f ).
- the Plurality of Engines ( 80 ) includes Engine ( 80 - 0 ), Engine ( 80 - 1 ), other engines, and Engine ( 80 - n ). These engines, according to an example embodiment, are input/output (I/O) processors that can interface with appropriate ones of the Plurality of Remote Resources ( 85 ) on behalf of the Plurality of Processing Stages ( 60 ).
- the Plurality of Processing Stages ( 60 ) communicate with the Plurality of Engines ( 80 ) through communication paths that are understood to constitute Engine Connections ( 100 - 0 , - 1 . . . f ).
- the Plurality of Engines ( 80 ) communicate with the Plurality of Remote Resources ( 85 ) through communication paths that are understood to constitute Engine to Resource Connections ( 101 - 0 , - 1 , . . . f ).
- the engine connections and the engine to resource connections are implemented, in an example embodiment, by an interconnect or other suitable connections.
- Each of the Plurality of Processing Stages ( 60 ) is substantially similar from a hardware point of view.
- the processing stage includes one or more Respective Processors ( 30 - 0 , 30 - 1 , . . . 3 n - 0 ).
- Each processor is configured to accept a packet such as Packet ( 20 - 0 ) from a Plurality of Parallel Packet Streams ( 20 ).
- the First Processing Stage ( 60 - 0 ) further includes a Respective Resource Manager ( 90 - 0 ) having Buffers ( 40 - 0 , 41 - 0 , . . . 4 n - 0 ).
- the Respective Resource Manager ( 90 - 0 ) of the First Processing Stage ( 60 - 0 ) further includes the Pool of Resource Locks ( 70 - 0 ) and the Respective Local Resource Locks ( 50 - 0 , 51 - 0 , . . . 5 n - 0 ).
- the First Processing Stage ( 60 - 0 ) is configured to receive the Packet ( 20 - 0 ) of a Plurality of Parallel Packet Streams ( 20 ). Although the packets of the Plurality of Parallel Packet Streams ( 20 ) are shown entering First Processing Stage ( 60 - 0 ) in a parallel fashion, the actual circuitry over which packets travel need not actually be implemented in such a manner. In FIG. 2 , there is shown a Packet Bus ( 200 ) along which each packet travels, according to an example embodiment.
- each of the Plurality of Processing Stages ( 60 ) is configured to implement a Lock & Read Operation ( 105 ), to implement a Write & Release Operation ( 106 ), to accept a Resource Lock Release Request received over the Resource Release Bus ( 300 ), and, subsequently, to release the indicated Resource Lock.
- the First Processing Stage ( 60 - 0 ) is configured, in accordance with the processing programmed the stage, and in response to the entrance of Packet ( 20 - 0 ), to allocate the Remote Resource ( 85 - 0 ) by locking Respective Local Resource Lock ( 51 - 0 ) of the Pool of Resource Locks ( 70 - 0 ), i.e., the Lock Operation.
- the First Processing Stage ( 60 - 0 ) is further configured to perform the Read Operation, through its Respective Resource Manager to Engine Connection ( 100 - 0 ), through Engine ( 80 - 0 ), and via Engine to Resource Connection ( 101 - 0 ) to Remote Resource ( 85 - 0 ).
- the First Processing Stage ( 60 - 0 ) is yet further configured to pass Packet ( 20 - 0 ) along the Packet Bus ( 200 ) as Packet ( 20 - 1 ).
- the Second Processing Stage ( 60 - 1 ) is configured, in response to receiving Packet ( 20 - 1 ), to perform the Write Operation, through Respective Resource Manager to Engine Connection ( 100 - 1 ), Engine ( 80 - 1 ), to Engine to Resource Connection ( 101 - 0 ), and to Remote Resource ( 85 - 0 ).
- the Second Processing Stage ( 60 - 1 ) is further configured to pass the Packet ( 20 - 1 ) along the Packet Bus ( 200 ) as Packet ( 20 - 2 ).
- the Second Processing Stage ( 60 - 1 ) is yet further configured to request the release of Remote Resource ( 85 - 0 ) by causing the unlocking of the Respective Local Resource Lock ( 51 - 0 ), i.e., the Release Operation.
- the Second Processing Stage ( 60 - 1 ) sends, along Resource Release Bus ( 300 ), a release request containing identifiers indicating the First Processing Stage ( 60 - 0 ) and its particular Respective Local Resource Lock ( 51 - 0 ).
- the First Processing Stage ( 60 - 0 ) is configured to receive the release request and to subsequently unlock the Respective Local Resource Lock ( 51 - 0 ), thereby deallocating the Remote Resource ( 85 - 0 ).
- the Remote Resource ( 85 - 0 ) is then available for subsequent allocation.
- FIG. 3 illustrates an example embodiment of a Packet Processing Device ( 1300 ).
- the Packet Processing Device ( 1300 ) includes a Receiver ( 110 ) for receiving packets, an Interface Arbiter ( 120 ) for directing packets along suitable paths through the pipeline configuration, a First Processing Stage ( 60 - 0 ) of a Plurality of Processing Stages ( 60 ), a Plurality of Engines ( 80 ), and a Remote Resource ( 85 - 0 ) of a Plurality of Remote Resources ( 85 ).
- the Receiver ( 110 ) receives a Packet ( 20 - 0 ) and subsequently passes the Packet ( 20 - 0 ) to the Interface Arbiter ( 120 ).
- the Interface Arbiter ( 120 ) directs the Packet ( 20 - 0 ) into the First Processing Stage ( 60 - 0 ).
- the Interface Arbiter ( 120 ) is also referred to as in ingress front end, in example embodiments.
- the First Processing Stage ( 60 - 0 ) includes a Respective Processor ( 30 - 0 ).
- the Respective Processor ( 30 - 0 ) includes a Packet Memory ( 160 ), an Execution Memory ( 150 ), an Instruction Memory ( 130 ), and a Processor Core ( 140 ).
- the Respective Processor ( 30 - 0 ) is configured to perform an operation consistent with the operations already described with respect to FIG. 1A and FIG. 1B , namely, a Lock & Read Operation ( 105 ) and a Write & Release Operation ( 106 ).
- the Packet Memory ( 160 ) stores a packet from the Plurality of Parallel Packet Streams ( 20 ).
- the Execution Memory ( 150 ) stores information related to the processing of the packet such as, for example, variable values and an indication as to whether one or more of the Plurality of Remote Resources ( 85 ) is allocated for the processing of the packet.
- the contents of both the Packet Memory ( 160 ) and the Execution Memory ( 150 ) are passed along, according to an example embodiment, as an Execution Context.
- the Processor Core ( 140 ) carries out operations in accordance with the contents of the Instruction Memory ( 130 ). The particular operations carried out depend on the type of packet being processed. In example embodiments the Processor Core ( 140 ) has been discussed in the context of programmable operations; it is noted that Processor Core ( 140 ) may be implemented multiple manners, such as a programmable processor core or as a hardware-designed processor core.
- multiple instances of Respective Processor ( 30 - 0 ) are provided in another example embodiment.
- the presence of multiple instances of Respective Processor ( 30 - 0 ) make it possible to carry out a number of operations within the First Processing Stage ( 60 - 0 ) before the packet is provided to the Second Processing Stage ( 60 - 1 ).
- the multiple instances of Respective Processor ( 30 - 0 ) carry out a packet processing operation associated with a single packet.
- the multiple instances of Respective Processor ( 30 - 0 ) carry out a packet processing operation associated with multiple packets.
- packet processing operations are performed on packets in different streams, depending on the particular processing context.
- the First Processing Stage ( 60 - 0 ) further includes a Respective Resource Manager ( 90 - 0 ) which includes a Buffer ( 40 ), a Pool of Resource Locks ( 70 - 0 ), and an I/O Access Unit ( 190 ).
- the Buffer ( 40 ) stores the Packet Memory ( 160 ) and the Execution Memory ( 150 ) for packets whose processing involves access to any of the Plurality of Remote Resources ( 85 ).
- the I/O Access Unit ( 190 ) further includes a Driver Table ( 180 ) utilized in the accessing of the Plurality of Engines ( 80 ) and, in some example embodiments, the Remote Resource ( 85 - 0 ).
- the First Processing Stage ( 60 - 0 ) communicates with the Plurality of Engines ( 80 ) by a Respective Resource Manager to Engine Connection ( 100 - 0 ), which then communicates with the Remote Resource ( 85 - 0 ) by an Engine to Resource Connection ( 105 - 0 ).
- An additional architecture is found in the disclosure of U.S. patent application Ser. No. 13/891,707 for “Hybrid Dataflow Processor,” filed May 10, 2013, owned by the same assignee and incorporated herein by reference for its useful descriptions and example architectures including descriptions of both hard and soft path packet processing operations.
- FIG. 4 is a flow diagram of an example method according to example embodiments.
- 5400 occurs when some Processing Stage accepts a Packet.
- the Processing Stage will read the Execution Context of the Packet and obtain Instructions based on the Execution Context.
- the Processing Stage will determine whether a Remote Resource is needed, in accordance with the Instructions.
- the Processing Stage will determine what other Packet Processing Operation to be performed.
- the thus-determined Packet Processing Operation is performed.
- the Packet is passed along the Packet Processing Pipeline as a Processed First Packet.
- the Processing Stage waits until the Resource Lock becomes available, and the Packet remains buffered in Buffer ( 40 ).
- the Processing Stage at S 405 locks the Local Resource Lock by setting the status to LOCKED, and then carries out a Packet Processing Operation in accordance with the Instructions.
- the information, that the Local Resource Lock has been set to the status of LOCKED travels as part of the execution context of the packet which is passed from the Processing Stage, at S 408 .
- the Processing Stage accesses the allocated Remote Resource, through an Engine, to Read a value from the Remote Resource.
- the value thus read is stored in the Packet's Execution Context.
- the Processed First Packet is passed along the pipeline configuration to a subsequent Processing Stage.
- FIG. 5 is a flow diagram of an example method according to example embodiments. The example method of FIG. 5 may be applied to multiple example embodiments as discussed and as understood by one of ordinary skill in the art.
- S 500 occurs when a Processing Stage accepts a Packet.
- the Processing Stage reads the Execution Context of the Packet and obtains Instructions based on the Execution Context.
- S 501 the Processing Stage determines whether a Resource is already Allocated in association with the received Packet.
- the Processing Stage carries out a WRITE operation with respect to the Allocated Resource.
- the Processing Stage then causes both S 505 and S 506 to occur.
- the Processing Stage pass along the Processed Second Packet, 5505 .
- the Processing Stage generates and sends, along the Resource Release Bus, a Resource Lock Release Request containing Identifiers of the Respective Resource Manager and of the particular Resource Lock to be released.
- FIG. 6 is a flow diagram of an example method according to example embodiments.
- Each Processing Stage is configured to monitor the Resource Release Bus ( 300 ) for any Resource Lock Release Request.
- S 601 represents continuous monitoring of the Resource Release Bus ( 300 ) until a Resource Lock Release Request (RLRR) is received.
- the Processing Stage determines at S 602 whether the Resource Lock Release Request is addressed to that Processing Stage.
- the use of the Local Resource Lock within the Processing Stage allows the treatment of multiple engines as a single resource.
- multiple engines are locked by setting the status of a single Local Resource Lock to LOCKED.
- FIG. 7 is a highly simplified schematic diagram of a device according to example embodiments, in which the Resource Locks are centrally managed by a Central Resource Lock Engine ( 700 - 0 ).
- a Packet Processing Device ( 1700 ) includes a Packet Bus ( 200 ), a Plurality of Packet Streams ( 20 ), a Plurality of Engines ( 80 ), a plurality of Remote Resources ( 85 - 0 , 85 - 1 , . . .
- the Central Resource Lock Engine ( 700 - 0 ) includes a Central Resource Lock Pool Local to Engine ( 780 ) which further includes an Engine Local Resource Lock ( 750 - 0 , 750 - 1 , . . . 750 - n ).
- the LOCK, READ, WRITE, and RELEASE Operations are each implemented by a separate Processing Stage.
- the First Processing Stage determines if a Central Resource Lock Pool Local to Engine ( 780 ) contains an Engine Local Resource Lock ( 750 - 0 , 750 - 1 , . . . 750 - n ) capable of locking some requested Resource. If available, the First Processing Stage ( 760 - 0 ) LOCKs the Resource.
- the First Processing Stage ( 760 - 0 ) passes Packet ( 20 - 0 ) along the Packet Bus ( 200 ) to the Second Processing Stage ( 760 - 1 ) as Packet ( 20 - 1 ) containing information that a Resource has been locked.
- the Second Processing Stage ( 760 - 1 ) is configured to READ from the locked Resource, through the Plurality of Engines ( 80 ), in response to the Packet ( 20 - 1 ).
- the Second Processing Stage ( 760 - 1 ) is further configured to pass the Packet ( 20 - 1 ) along the Packet Bus ( 200 ) to the Third Processing Stage ( 760 - 2 ) as the Packet ( 20 - 2 ), which contains information from the READ operation.
- the Third Processing Stage ( 760 - 2 ) is configured to WRITE to a Resource, through the Plurality of Engines ( 80 ), in response to the Packet ( 20 - 2 ).
- the Third Processing Stage ( 760 - 2 ) is further configured to pass the Packet ( 20 - 2 ) along the Packet Bus ( 200 ) to the Fourth Processing Stage ( 760 - 3 ) as the Packet ( 20 - 3 ), which contains information from the WRITE operations.
- the Fourth Processing Stage ( 760 - 3 ) is configured to generate a Resource Lock Release Request in response to the Packet ( 20 - 3 ).
- the Central Resource Lock Engine ( 700 - 0 ) is configured to receive the Resource Lock Release Request and subsequently release the Resource Lock from the Central Resource Lock Pool Local to Engine ( 780 ).
- the allocated resource has been modified and is now available to be reallocated.
- the Fourth Processing Stage ( 760 - 3 ) then passes the Packet ( 20 - 3 ) along the Packet Bus ( 200 ) as the Packet ( 20 - 4 ).
- the Central Resource Lock Engine ( 700 - 0 ) can avoid contention for Remote Resources without employing a Resource Release Bus ( 300 ).
- the Resource Locks are not local to the Plurality of Processing Stages ( 60 )
- the Lock & Read Operation ( 105 ) cannot be performed within a single Processing Stage.
- the Write & Release Operation ( 106 ) also cannot be performed within single Processing Stage.
- FIG. 8 is a highly simplified schematic diagram of a device according to example embodiments.
- the Resource Locks are local to the Engines and not local to the Processing Stages.
- Packet Processing Device ( 1800 ) includes a Packet Bus ( 200 ), a Packet ( 20 - 0 , - 1 , - 2 ) of a Plurality of Parallel Packet Streams ( 20 ), a First Processing Stage ( 860 - 0 ), a Second Processing Stage ( 860 - 1 ), Remote Resources ( 85 - 0 , - 1 , . . . - n ), and a Plurality of Engines ( 80 ) containing Resource Lock Engines ( 880 - 0 , 1 , . . . - n ).
- Each of the Resource Lock Engines ( 880 - 0 , - 1 , . . . - n ) contains a Respective Resource Lock Pool Local to each Engine ( 870 - 0 , - 1 , . . . - n ), and each Respective Resource Lock Pool Local to each Engine ( 870 - 0 , - 1 , . . . - n ) further contains Respective Engine Local Resource Locks ( 850 - 0 , 851 - 0 , . . . 85 n - 0 ; 850 - 1 , 851 - a , . . . 85 n ⁇ 1; . . . ; and 850 - f , a 851 - f , 85 n - f ) according to an example embodiment.
- the First Processing Stage ( 860 - 0 ) is configured to perform a Lock & Read Operation ( 105 ) in response to a Packet ( 20 - 0 ).
- the Lock & Read Operation ( 105 ) includes accessing a Resource Lock Engine ( 880 - 1 ) to access a Resource Lock Pool Local to Engine ( 870 - 1 ) to request availability of an Engine Local Resource Lock ( 850 - 1 ) corresponding to a Remote Resource ( 85 - 1 ), according to an example embodiment.
- the First Processing Stage ( 860 - 0 ) is further configured to LOCK the Remote Resource ( 85 - 1 ) if the Engine Local Resource Locks ( 850 - 1 ) is available.
- the First Processing Stage ( 860 - 0 ) is further configured to READ a value from the Remote Resource ( 85 - 1 ) and store information of that value into the Execution Context of the Packet ( 20 - 0 ).
- the First Processing Stage ( 860 - 0 ) then passes the Packet ( 20 - 0 ) along the Packet Bus ( 200 ) as a Packet ( 20 - 1 ).
- the Second Processing Stage ( 860 - 1 ) is configured to accept the Packet ( 20 - 1 ) and subsequently perform a Write & Release Operation ( 106 ) associated with the Remote Resource ( 85 - 1 ) through the Resource Lock Engine ( 880 - 1 ).
- the Second Processing Stage ( 860 - 1 ) is configured to WRITE to the Remote Resource ( 85 - 1 ) and subsequently send a Resource Lock Release Request to the Resource Lock Engine ( 880 - 1 ).
- the allocated resource has been modified and is now available to be reallocated.
- the Lock & Read Operation ( 105 ) can be performed within a single Processing Stage, as can the Write & Release Operation ( 106 ).
- the single Processing Stage may perform a Lock & Read Operation ( 105 ) or a Write & Release Operation ( 106 ) within a single stage.
- the Resource Release Bus ( 300 ) is not required.
- the Resource Locks are local to each of the Engines, the problem of contention is avoided only for operations seeking the same Remote Resource through the same Engine.
- a hybrid architecture implementing a soft path and a hard path will now be discussed in the context of FIGS. 2 and 3 , according to example embodiments.
- a path is to be understood as the collection of processing stage resources, and the associated packet processing operations, through which a packet passes during its time traversing the pipeline configuration.
- the path through which packets labeled (P — 0, P — 0′, P — 0′′, . . .
- P — 0f) pass includes Respective Processor ( 30 - 0 ) and Buffer ( 40 - 0 ) of Processing Stage ( 60 - 0 ); Respective Processor ( 30 - 1 ) and Buffer ( 40 - 1 ) of Processing Stage ( 60 - 1 ); and so on until the path is concluded with Respective Processor ( 30 - n ) and Buffer ( 40 - f ) of Processing Stage ( 60 - f ).
- the Respective Processor ( 30 - 0 , 1 , . . . n) is implemented in an example embodiment as a plurality of individual processors, each of which is part of the path through which packets labeled (P — 0, P — 0′, P — 0′′, . . . P — 0f) pass.
- P — 0, P — 0′, P — 0′′, . . . P — 0f another path exists for the packets labeled (P — 1, P — 1′, P — 1′′, . . .
- FIGS. 2 and 3 thus illustrate a packet processing device through which multiple paths exist. Multiple streams of packets are handled at the same time along the multiple paths, i.e., are processed in parallel.
- Hard paths provide fixed latency, and soft paths do not provide fixed latency.
- a hard path therefore completes the processing of every hard path packet within a certain, well-defined time period, if at all.
- a soft path does not achieve the processing of soft path packets in any particular time period, and the soft path packet processing is typically implemented on a best effort basis so as not to impede the processing of hard path packets.
- soft paths guarantee the processing of soft path packets will complete (i.e., guaranteed delivery), while hard paths sometimes drop hard path packets.
- a packet processing device includes hard paths and one or more soft paths.
- hard path packets under certain circumstances are dropped and not fully processed through the pipeline configuration. That is to say, if the processing for a hard path packet requires a Remote Resource ( 85 - 0 , - 1 , . . . - n ), but the corresponding Local Resource Lock ( 50 ) has a LOCKED state, the hard path packet is dropped. Dropping the hard path packet under such circumstances guarantees that the fixed processing latency is not subverted failure to obtain a resource allocation.
- one or more soft paths are configured to allow for a pause in soft path processing in response to a soft path packet being unable to allocate a Remote Resource. That is to say, if the processing for a soft path packet requires a Remote Resource ( 85 - 0 , - 1 , . . . - n ), but the corresponding Local Resource Lock ( 50 ) has a LOCKED state, the soft path packet remains in Buffer ( 40 ) for that particular Processing Stage until the Local Resource Lock ( 50 ) has an UNLOCKED state.
- the Local Resource Lock ( 50 ) acts as a semaphore indicating whether the processing for a soft path packet should be paused. That is to say, when a Local Resource Lock ( 50 ) corresponding to a Remote Resource ( 85 - 0 , - 1 , . . . - n ) required for the subsequent processing of a soft path packet has a LOCKED state, the processing for the soft path packet is thus paused.
- a beneficial effect of such a pause is that it mitigates the possibility that the processing for a soft path packet might adversely affect the processing of hard path packets.
- the use of Local Resource Lock ( 50 ) provides a mechanism that permits processing of soft path packets to be interleaved with the processing of hard path packets in such a manner that the soft path packet processing is not carried out when remote resources are already allocated to hard path processing, or to other soft path processing.
- an Interface Arbiter ( 120 ) directs data packets through one of the hard paths, and directs control and management packets along the one or more soft paths.
- a packet, for which a Remote Resource must be accessed as part of its processing is directed along a soft path by Interface Arbiter ( 120 ).
- Interface Arbiter 120
Landscapes
- Engineering & Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Data Exchanges In Wide-Area Networks (AREA)
Abstract
A packet processing device has a plurality of processing stages, including a first processing stage and a second processing stage arranged as a packet processing pipeline. The first processing stage and the second processing stage each have a respective processor configured to process a packet of a packet stream and a respective resource manager having a respective local resource lock corresponding to a remote resource. The respective processor requests the respective resource manager to allocate the remote resource. The respective resource manager responds to the request to allocate the remote resource by locking the remote resource with the respective local resource lock and allocating the remote resource. The respective processor implements a packet processing operation associated with the allocated remote resource.
Description
- This application claims priority to and the benefit of U.S. Provisional Patent Application No. 61/694,483 filed Aug. 29, 2012, and U.S. Provisional Patent Application No. 61/753,767 filed Jan. 17, 2013, the disclosures of both of which are incorporated by reference herein in their entirety. This application is related in content to U.S. patent application Ser. No. 13/891,707 for “Hybrid Dataflow Processor,” filed May 10, 2013, the entire disclosure of which is incorporated by reference herein in its entirety.
- The present disclosure relates to a network device that processes packets.
- The background description provided herein is for the purpose of generally presenting the context of the disclosure. Work of the presently named inventor, to the extent it is described in this background section, as well as aspects of the description that may not otherwise qualify as prior art at the time of filing, are neither expressly nor impliedly admitted as prior art against the present disclosure.
- A pipeline-type packet processing device typically includes processing stages arranged to process packets in what is understood as a pipeline configuration. In a pipeline configuration, packets are processed in stages with packets progressing from stage to stage through the pipeline. In some packet processing pipelines, such as a dataflow packet processing pipeline, respective stages are programmable. A stage in the packet processing pipeline receives a packet that had been processed at one or more previous stages, along with a processing context for the packet. Various resources, such as for example various table lookups, that are to be used by the respective stages in a packet processing pipeline typically are provided by units that are external to the packet processing pipeline and that are accessed by the stages when those services are needed to perform a processing operation.
- One or more embodiments of the disclosure generally relate to a network device that processes packets through a number of processing stages. Some packet processing operations update, i.e., read and write a value in a remote resource such as a memory location in a table. This situation is also referred to as “read-modify-write.” When the processing of a packet across multiple stages involves updating a remote resource, the possibility of contention for the remote resource, from packet processing operations implemented for prior or succeeding packets, is mitigated. According to example embodiments, when the processing of a given packet involves packet processing operations that will read from and subsequently write to a given remote resource, the remote resource is allocated or “locked” so that the processing of other packets cannot interfere with the remote resource until the processing carried out for the given packet reaches a point at which the remote resource is suitably released. Semaphores are used in example embodiments to lock corresponding remote resources. Notwithstanding the foregoing, not every example embodiment is required to possess all or even any of the features mentioned in this paragraph.
- According to an aspect of the present disclosure, there is provided a packet processing device having processing stages, including a first processing stage and a second processing stage arranged as a packet processing pipeline; the first processing stage and the second processing stage each have: a respective processor configured to process a packet of a packet stream, and a respective resource manager having a respective local resource lock corresponding to a remote resource; the respective processor is configured to request the respective resource manager to allocate the remote resource; the respective resource manager is further configured to respond to the request to allocate the remote resource by locking the remote resource with the respective local resource lock and allocating the remote resource; the respective processor is further configured to implement a packet processing operation associated with the allocated remote resource.
- According to another example embodiment, a packet processing device includes: processing stages arranged as a packet processing pipeline; the processing stages each having processor cores and buffers; the processor cores and buffers of the processing stages defining a plurality of paths, for simultaneous packet processing, through the packet processing pipeline; an ingress front end configured to direct each packet of an incoming stream of packets into one of the plurality of paths; the paths including a hard path and a soft path, the hard path being configured to process received ones of the incoming stream of packets with fixed latency, the soft path being configured to process received ones of the incoming stream of packets with variable latency; and the processing stages each further including a respective resource manager configured to request allocation of a remote resource, for a given packet of the incoming stream of packets, in response to an instruction from one of the processor cores. According to another example embodiment, the respective resource manager is further configured to allocate an available remote resource, thereby making the remote resource accessible only to the respective packet and to subsequent packet processing operations for that packet. Also, the processing stages are further configured, in an example embodiment, to both request the release of a allocated resource, and to receive from another processing stage a request to release an allocated resource and subsequently to release the resource. A remote resource which has been released is available to be allocated.
- In yet another example embodiment, a packet processing method includes receiving, at a processor of a first processing stage, a first packet and a request for allocation of a remote resource; responding, by the processor, to the allocation request, by setting a semaphore corresponding to the remote resource to indicate a locked status; implementing a first packet processing operation, in association with the allocated remote resource, and in association with the first packet, to obtain a processed first packet; and outputting the processed first packet to a next processing stage of the pipeline packet processing device.
-
FIG. 1A is a highly simplified illustrative drawing showing a concept of operation according to example embodiments. -
FIG. 1B is a highly simplified illustrative drawing showing another concept of operation according to example embodiments. -
FIG. 2 is a highly simplified schematic diagram of a device according to example embodiments. -
FIG. 3 is a highly simplified schematic diagram showing a particular aspect of a device according to example embodiments. -
FIG. 4 is a flow diagram of an example method according to example embodiments. -
FIG. 5 is a flow diagram of an example method according to example embodiments. -
FIG. 6 is a flow diagram of an example method according to example embodiments. -
FIG. 7 is a highly simplified schematic diagram of a device according to example embodiments. -
FIG. 8 is a highly simplified schematic diagram of a device according to example embodiments. - In the following discussion, descriptions of well-known functions and constructions may be omitted for increased clarity and conciseness.
-
FIG. 1A shows a Packet Processing Device (1000). Packet Processing Device (1000) includes a Plurality of Processing Stages (60), including a First Processing Stage (60-0), a Second Processing Stage (60-1), and other processing stages in accordance with particular example embodiments. Here, the terms “first” and “second” do not connote that the First Processing Stage (60-0) is limited to being the first processing stage in a pipeline configuration, nor that the Second Processing Stage (60-1) is limited to being the second processing stage in the pipeline configuration. Instead, “first” and “second” simply indicate a relative order in the pipeline configuration such that the First Processing Stage (60-0) is relatively earlier along the pipeline configuration than is the Second Processing Stage (60-1). In this discussion and in the drawing figures, the First Processing Stage (60-0) and Second Processing Stage (60-1) are illustrated adjacent each other in the pipeline configuration, but this arrangement is merely an example embodiment, and other processing stages are interposed between the First Processing Stage (60-0) and the Second Processing Stage (60-1) in example embodiments. Further, the First Processing Stage (60-0) is considered to be “upstream” with respect to the Second Processing Stage (60-1). Likewise, the Second Processing Stage (60-1) is considered to be “downstream” with respect to the first processing stage, according to an example embodiment. In an embodiment the packet processing device (1000) is configured as a dataflow pipeline in which the respective stages (60) are programmable, and in which packets are received for processing at a downstream processing stage along with a processing context for processing the corresponding received packet. In an alternative example embodiment, one or more individual stages of the Plurality of Processing Stages (60) are not programmable but have hard-coded functionality. In yet another example embodiment, the Plurality of Processing Stages (60) includes some programmable processing stages and some hard-coded processing stages. - Each of the Plurality of Processing Stages (60) is substantially the same from a hardware perspective, according to an example embodiment. In another example embodiment, the Plurality of Processing Stages (60) includes different types of processing stages. The discussion below assumes that each of the Plurality of Processing Stages (60) has the same hardware structure. The components of each processing stage are described in more detail below. At the present, however, it is noted that each of the Plurality of Processing Stages (60) includes a respective Pool of Resource Locks (70). The respective Pool of Resource Locks (70) for the First Processing Stage (60-0) is shown as Pool of Resource Locks (70-0). Likewise, the respective Pool of Resource Locks (70) for the Second Processing Stage (60-1) is shown as Pool of Resource Locks (70-1). Each of the Plurality of Processing Stages (60) has a respective Pool of Resource Locks (70), according to example embodiments.
- Within each respective Pool of Resource Locks (70) there exists a plurality of semaphores. In the present example embodiment, the semaphores are represented as Local Resource Locks (50, 51, . . . 5 n). The number of semaphores used depends on particular implementation details, and
FIG. 1 shows only a few semaphores within a Pool of Resource Locks (70) for the sake of simplicity of illustration. Each respective Pool of Resource Locks (70) includes its own Local Resource Locks (50, 51, . . . 5 n). The Pool of Resource Locks (70-0) for First Processing Stage (60-0) includes Respective Local Resource Lock (50-0), Respective Local Resource Lock (51-0), . . . and Respective Local Resource Lock (5 n-0). Likewise, the Pool of Resource Locks (70-1) for Second Processing Stage (60-1) includes Respective Local Resource Lock (50-1), Respective Local Resource Lock (51-1), . . . and Respective Local Resource Lock (5 n−1). - Also shown in
FIG. 1A are a Plurality of Remote Resources (85) including Remote Resource (85-0), Remote Resource (85-1), . . . and Remote Resource (85-n). Together, these represent remote resources that are available for READ and WRITE operations carried out by the Plurality of Processing Stages (60). In an example embodiment, the remote resources are memory locations that contain data such as, in a system for processing packets received on a network, forwarding tables, policy tables, MAC address learning, time-out management, and the like. This example is non-limiting, however, and the remote resources according to other exemplary embodiments contain data in structures other than tables. In an example embodiment, a remote resource contains a value which is accessed by either a READ or a WRITE Operation. In a READ Operation, a Processing Stage obtains a value stored in a memory location of the remote resource. In a WRITE Operation, the Processing Stage stores a value into the memory location of the remote resource. It is to be noted, however, that a remote resource may be locked by a process intended to subsequently perform alternate operations in addition to READ and WRITE operations, for example updating or synchronizing multiple tables and/or memory locations. - Between the Plurality of Processing Stages (60) and the Plurality of Remote Resources (85) is illustrated a Plurality of Engines (80). In an example embodiment, a Processing Stage requires an Engine to access a Remote Resource. In other example embodiments, however, the Plurality of Processing Stages (60) are able to access the Plurality of Remote Resources (85) directly or without the use of an Engine. According to another example embodiment, a given Remote Resource is accessed through a combination of engines.
- At the lower portion of
FIG. 1A , there is illustrated a more detailed view of the Pool of Resource Locks (70-0) of the First Processing Stage (60-0), according to an example embodiment. This more detailed representation of the Pool of Resource Locks (70-0) is shown as a table containing two columns. Each row of the table represents a Local Resource Lock such as Respective Local Resource Lock (50-0), Respective Local Resource Lock (51-0), . . . Respective Local Resource Lock (5 n-0). The left column of the table includes, in each row, an identifier of a corresponding one of the Plurality of Remote Resources (85). The right-column of the table indicates the status of the corresponding Remote Resource. In this non-limiting, simplified example the status is UNLOCKED or LOCKED. The table ofFIG. 1A indicates that, in this example, the Remote Resource (85-0) is UNLOCKED, the Remote Resource (85-1) is LOCKED, and the Remote Resource (85-n) is UNLOCKED. - In the example embodiment shown in
FIG. 1A , the Processing Stages are connected to a Resource Release Bus (300), also referred to as a lock release message path. The Resource Release Bus (300) allows for communication, such as message exchange, among the Plurality of Processing Stages (60). According to an example embodiment, the Resource Release Bus (300) allows for messages to be passed from upstream stages to downstream stages. The Resource Release Bus (300), however, is implemented in other ways according to example embodiments, and a bus that permits message exchange is a non-limiting example. In the present example embodiment, the Resource Release Bus (300) does not carry the packets being processed through the pipeline configuration. - A single Packet (20-0) is shown in
FIG. 1A , for simplicity of illustration. The Packet (20-0) enters the First Processing Stage (60-0), and will pass to the Second Processing Stage (60-1) after any packet processing operations programmed into the First Processing Stage (60-0) are concluded. -
FIG. 1A also illustrates part of an overall concept of operation according to an example embodiment. In particular, 1 a-1 indicates the entering of Packet (20-0) into the First Processing Stage (60-0). At this instant, all locks of the Pool of Resource Locks (70-0) have an UNLOCKED status. The packet processing operations programmed into the First Processing Stage (60-0) include, for example, a READ of Remote Resource (85-1). Before carrying out the READ operation, however, the packet processing operations programmed into First Processing Stage (60-0) carry out a check to determine whether Remote Resource (85-1) is in a LOCKED or in an UNLOCKED status. To make this determination the Pool of Resource Locks (70-0) is checked. More particularly, the particular Respective Local Resource Lock (51-0) corresponding to Remote Resource (85-1) is checked. At 1 a-2 a, Finding the Respective Local Resource Lock (51-0) has an UNLOCKED status, the First Processing Stage (60-0) sets the Respective Local Resource Lock (51-0) to a LOCKED status. At 1 a-2 b, having locked Respective Local Resource Lock (51-0), the First Processing Stage (60-0) carries out the READ operation with respect to Remote Resource (85-1) using an appropriate one of the Plurality of Engines (80). The value of Remote Resource (85-1) is thus read by the First Processing Stage (60-0) and this value is now available to use in accordance with the processing programmed into First Processing Stage (60-0) and others of the Plurality of Processing Stages (60). At 1 a-3, the Packet (20-0) is transferred to Second Processing Stage (60-1), together with other information to be described later. - In this example embodiment, the setting of the status of Respective Local Resource Lock (51-0), corresponding to Remote Resource (85-1), from UNLOCKED to LOCKED, implements a locking of the remote resource. To put it another way, the Respective Local Resource Lock (51-0) acts as a semaphore that indicates that Remote Resource (85-1) is an allocated remote resource.
- It is noted that, in an embodiment the mechanism for locking Remote Resource (85-1) and the mechanism for subsequent reading of the value of Remote Resource (85-1) are both located and activated locally within the First Processing Stage (60-0). If, for example, in contrast, semaphores were to be maintained in a shared pool of semaphores located outside the First Processing Stage (60-0), delays in making the determination as to the LOCKED or UNLOCKED status of Remote Resource (85-1) would likely result. Although maintaining a shared pool of semaphores (not shown) streamlines some aspects of the processor architecture, in an example embodiment, one of the Plurality of Engines (80) would need to obtain LOCKED or UNLOCKED status information from such a shared pool of semaphores.
- Since the locking of Remote Resource (85-1) and the subsequent reading of the value of Remote Resource (85-1) take place within the First Processing Stage (60-0), within a very short time of each other, it is noted that the lock and read operations constitute a Lock & Read Operation (105). It is further noted that the Lock & Read Operation (105) allocates the Remote Resource (85-1) by setting a value of Respective Local Resource Lock (51-0), thereby preventing First Processing Stage (60-0) from carrying out any operations with respect to Remote Resource (85-1) until the status of Remote Resource (85-1) in Respective Local Resource Lock (51-0) goes from LOCKED to UNLOCKED.
- It is noted that, in First Processing Stage (60-0), the Lock & Read Operation (105) includes a READ operation with respect to the allocated Remote Resource (85-1). This READ operation is understood to constitute a packet processing operation associated with the allocated Remote Resource (85-1).
- Having thus described an overall concept of operation for the Lock & Read Operation (105), the discussion will proceed to a further concept of operation relating to a Write & Release Operation (106).
- At 1 a-3 of
FIG. 1A , Packet (20-0) completed processing at and exits from the First Processing Stage (60-0). This packet, prior to undergoing the processing programmed into First Processing Stage (60-0), is understood to constitute a first packet. At 1 a-3, however, this packet is understood to constitute a processed first packet. In some circumstances, depending on the processing programmed into the given processing stage, the Packet (20-0) changes. Such changes, however, are not strictly required, and in some situations the Packet (20-0) stays the same. Therefore, it is noted that the processing programmed into First Processing Stage (60-0), performed with respect to the first packet, results, in any event, with the passing along of a processed first packet. - Turning to
FIG. 1B , Packet (20-1) represents a second packet. At 1 b-1, the Packet (20-1) is passed to Second Processing Stage (60-1). In this example, the Packet (20-1) is a packet which, in a prior processing stage, resulted in the locking of a remote resource such as Remote Resource (85-1). For the sake of this discussion, Packet (20-1) is the processed first packet which exits from First Processing Stage (60-0) at 1 a-3. - In
FIG. 1B , at 1 b-2 a, the processing programmed into Second Processing Stage (60-1) causes a WRITE operation to be carried out with respect to Remote Resource (85-1). Having concluded the WRITE operation, it is suitable to release the Remote Resource (85-1). At 1 b-2 b, to effect the release of Remote Resource (85-1), Second Processing Stage (60-1) sends a Resource Lock Release Request over Resource Release Bus (300). As shown inFIG. 1B , the Resource Lock Release Request includes two items of information. One item of information in the Resource Lock Release Request is an identifier of the one of the Plurality of Processing Stages (60) that has the Respective Local Resource Lock (50-1) corresponding to the locked Remote Resource (85-1). The other item of information is an identifier of the particular local resource lock (here, Respective Local Resource Lock (50-1)) to be released. - Since the Resource Release Bus (300) needs to carry the Resource Lock Release Request only to upstream ones of the Plurality of Processing Stages (60), the Resource Release Bus (300) communicates in one direction, e.g. the upstream direction, in an example embodiment.
- At 1 b-3 a, the packet Packet (20-1) exits from the Second Processing Stage (60-1) as a processed second packet and enters into a subsequent one of the Plurality of Processing Stages (60).
- At 1 b-3 b, the First Processing Stage (60-0) receives the Resource Lock Release Request addressed to the First Processing Stage (60-0), and changes the status of Remote Resource (85-1) in Respective Local Resource Lock (51-0) of the Pool of Resource Locks (70-0) by replacing LOCKED with UNLOCKED (illustrated as “LOCKED ->UNLOCKED” in
FIG. 1B ). After the status of Remote Resource (85-1) is updated in Respective Local Resource Lock (51-0), the Remote Resource (85-1) is no longer locked, and is available to be allocated for use in a packet processing operation for another packet. - Since the WRITE operation and the subsequent sending of the Resource Lock Release Request take place within the Second Processing Stage (60-1), within a very short time of each other, it is noted that these two actions constitute a Write & Release Operation (106). It is further understood that the Write & Release Operation (106) releases, or deallocates the Remote Resource (85-1) by causing the First Processing Stage (60-0) to set a value of its Respective Local Resource Lock (51-0), thereby enabling First Processing Stage (60-0) to carry out operations with respect to Remote Resource (85-1) if necessary.
- In an example embodiment the Resource Lock Release Request is understood to constitute, more generally, a release request message.
- In an example embodiment, the status of one of the Plurality of Remote Resources (85) is represented in an alternative manner, such as a single binary digit with one value representing a LOCKED status, and the other value representing an UNLOCKED status. Other implementations of the Pool of Resource Locks (70-0) are within the ability of a person familiar with this field. Further, the implementation of semaphores may be substituted with other locking mechanisms familiar to those skilled in the art including a single client lock such as a mutex or the like.
- According to an example embodiment, therefore, local locks are used to restrict the accessibility of a remote resource to the processing for a single packet, along the respective processing stages. According to an example embodiment, the First Processing Stage (60-0) reads from the Remote Resource (85-1) and passes the Packet (20-0) to the Second Processing Stage (60-1) as the Packet (20-1). The Second Processing Stage (60-1) then writes a value to the Remote Resource (85-1). The semaphore (Respective Local Resource Lock (50-1)) allows exclusive access by the Second Processing Stage (60-1) for the purpose of writing. However, without such a semaphore, the Second Processing Stage (60-1) would not be guaranteed exclusive access to the Remote Resource (85-1). That is to say, the First Processing Stage (60-0) would have the opportunity, in response to a new packet requiring the same Remote Resource (85-1), to begin reading from the Remote Resource (85-1) while the Second Processing Stage (60-1) is beginning to write to the Remote Resource (85-1), resulting in possible problems due to such contention. The use of semaphores, in Packet Processing Device (1000), thus avoids such contention problems.
-
FIG. 2 shows a more detailed view of the Packet Processing Device (1200). InFIG. 2 , the Plurality of Processing Stages (60) includes First Processing Stage (60-0), Second Processing Stage (60-1), other processing stages, and Final Processing Stage (60-f). - The Plurality of Engines (80) includes Engine (80-0), Engine (80-1), other engines, and Engine (80-n). These engines, according to an example embodiment, are input/output (I/O) processors that can interface with appropriate ones of the Plurality of Remote Resources (85) on behalf of the Plurality of Processing Stages (60). The Plurality of Processing Stages (60) communicate with the Plurality of Engines (80) through communication paths that are understood to constitute Engine Connections (100-0, -1 . . . f). The Plurality of Engines (80) communicate with the Plurality of Remote Resources (85) through communication paths that are understood to constitute Engine to Resource Connections (101-0, -1, . . . f). The engine connections and the engine to resource connections are implemented, in an example embodiment, by an interconnect or other suitable connections.
- Each of the Plurality of Processing Stages (60) is substantially similar from a hardware point of view. Taking the First Processing Stage (60-0) as an example, the processing stage includes one or more Respective Processors (30-0, 30-1, . . . 3 n-0). Each processor is configured to accept a packet such as Packet (20-0) from a Plurality of Parallel Packet Streams (20).
- The First Processing Stage (60-0) further includes a Respective Resource Manager (90-0) having Buffers (40-0, 41-0, . . . 4 n-0). The Respective Resource Manager (90-0) of the First Processing Stage (60-0) further includes the Pool of Resource Locks (70-0) and the Respective Local Resource Locks (50-0, 51-0, . . . 5 n-0).
- The First Processing Stage (60-0) is configured to receive the Packet (20-0) of a Plurality of Parallel Packet Streams (20). Although the packets of the Plurality of Parallel Packet Streams (20) are shown entering First Processing Stage (60-0) in a parallel fashion, the actual circuitry over which packets travel need not actually be implemented in such a manner. In
FIG. 2 , there is shown a Packet Bus (200) along which each packet travels, according to an example embodiment. - As previously mentioned in the context of
FIGS. 1 a, 1 b, and further below inFIG. 3 which will be discussed momentarily, each of the Plurality of Processing Stages (60) is configured to implement a Lock & Read Operation (105), to implement a Write & Release Operation (106), to accept a Resource Lock Release Request received over the Resource Release Bus (300), and, subsequently, to release the indicated Resource Lock. According to an example embodiment, the First Processing Stage (60-0) is configured, in accordance with the processing programmed the stage, and in response to the entrance of Packet (20-0), to allocate the Remote Resource (85-0) by locking Respective Local Resource Lock (51-0) of the Pool of Resource Locks (70-0), i.e., the Lock Operation. - The First Processing Stage (60-0) is further configured to perform the Read Operation, through its Respective Resource Manager to Engine Connection (100-0), through Engine (80-0), and via Engine to Resource Connection (101-0) to Remote Resource (85-0). The First Processing Stage (60-0) is yet further configured to pass Packet (20-0) along the Packet Bus (200) as Packet (20-1).
- The Second Processing Stage (60-1) is configured, in response to receiving Packet (20-1), to perform the Write Operation, through Respective Resource Manager to Engine Connection (100-1), Engine (80-1), to Engine to Resource Connection (101-0), and to Remote Resource (85-0). The Second Processing Stage (60-1) is further configured to pass the Packet (20-1) along the Packet Bus (200) as Packet (20-2). The Second Processing Stage (60-1) is yet further configured to request the release of Remote Resource (85-0) by causing the unlocking of the Respective Local Resource Lock (51-0), i.e., the Release Operation. To cause the unlocking, the Second Processing Stage (60-1) sends, along Resource Release Bus (300), a release request containing identifiers indicating the First Processing Stage (60-0) and its particular Respective Local Resource Lock (51-0). The First Processing Stage (60-0) is configured to receive the release request and to subsequently unlock the Respective Local Resource Lock (51-0), thereby deallocating the Remote Resource (85-0). The Remote Resource (85-0) is then available for subsequent allocation.
-
FIG. 3 illustrates an example embodiment of a Packet Processing Device (1300). The Packet Processing Device (1300) includes a Receiver (110) for receiving packets, an Interface Arbiter (120) for directing packets along suitable paths through the pipeline configuration, a First Processing Stage (60-0) of a Plurality of Processing Stages (60), a Plurality of Engines (80), and a Remote Resource (85-0) of a Plurality of Remote Resources (85). The Receiver (110) receives a Packet (20-0) and subsequently passes the Packet (20-0) to the Interface Arbiter (120). The Interface Arbiter (120) directs the Packet (20-0) into the First Processing Stage (60-0). The Interface Arbiter (120) is also referred to as in ingress front end, in example embodiments. - The First Processing Stage (60-0) includes a Respective Processor (30-0). The Respective Processor (30-0) includes a Packet Memory (160), an Execution Memory (150), an Instruction Memory (130), and a Processor Core (140). The Respective Processor (30-0) is configured to perform an operation consistent with the operations already described with respect to
FIG. 1A andFIG. 1B , namely, a Lock & Read Operation (105) and a Write & Release Operation (106). - The Packet Memory (160) stores a packet from the Plurality of Parallel Packet Streams (20). The Execution Memory (150) stores information related to the processing of the packet such as, for example, variable values and an indication as to whether one or more of the Plurality of Remote Resources (85) is allocated for the processing of the packet. When a packet exits from the First Processing Stage (60-0) to the Second Processing Stage (60-1), the contents of both the Packet Memory (160) and the Execution Memory (150) are passed along, according to an example embodiment, as an Execution Context.
- The Processor Core (140) carries out operations in accordance with the contents of the Instruction Memory (130). The particular operations carried out depend on the type of packet being processed. In example embodiments the Processor Core (140) has been discussed in the context of programmable operations; it is noted that Processor Core (140) may be implemented multiple manners, such as a programmable processor core or as a hardware-designed processor core.
- As shown in
FIG. 3 , multiple instances of Respective Processor (30-0) are provided in another example embodiment. The presence of multiple instances of Respective Processor (30-0) make it possible to carry out a number of operations within the First Processing Stage (60-0) before the packet is provided to the Second Processing Stage (60-1). According to an example embodiment, the multiple instances of Respective Processor (30-0) carry out a packet processing operation associated with a single packet. According to another example embodiment, the multiple instances of Respective Processor (30-0) carry out a packet processing operation associated with multiple packets. According to yet another example embodiment, packet processing operations are performed on packets in different streams, depending on the particular processing context. - The First Processing Stage (60-0) further includes a Respective Resource Manager (90-0) which includes a Buffer (40), a Pool of Resource Locks (70-0), and an I/O Access Unit (190). The Buffer (40) stores the Packet Memory (160) and the Execution Memory (150) for packets whose processing involves access to any of the Plurality of Remote Resources (85).
- The I/O Access Unit (190) further includes a Driver Table (180) utilized in the accessing of the Plurality of Engines (80) and, in some example embodiments, the Remote Resource (85-0).
- The First Processing Stage (60-0) communicates with the Plurality of Engines (80) by a Respective Resource Manager to Engine Connection (100-0), which then communicates with the Remote Resource (85-0) by an Engine to Resource Connection (105-0).
- A more detailed architecture that is suitable for implementing one of the Plurality of Processing Stages (60), according to an example embodiment, is described in Jakob Carlstrom, and Thomas Boden, “Synchronous Dataflow Architecture for Network Processors,” IEEE MICRO, (September-October 2004), the content of which is incorporated in its entirety herein for its useful description and example architecture. An additional architecture is found in the disclosure of U.S. patent application Ser. No. 13/891,707 for “Hybrid Dataflow Processor,” filed May 10, 2013, owned by the same assignee and incorporated herein by reference for its useful descriptions and example architectures including descriptions of both hard and soft path packet processing operations.
-
FIG. 4 . is a flow diagram of an example method according to example embodiments. 5400 occurs when some Processing Stage accepts a Packet. The Processing Stage will read the Execution Context of the Packet and obtain Instructions based on the Execution Context. At S401, the Processing Stage will determine whether a Remote Resource is needed, in accordance with the Instructions. - If at S401 a Remote Resource is not needed, the Processing Stage will determine what other Packet Processing Operation to be performed. At S402, the thus-determined Packet Processing Operation is performed. At S408, the Packet is passed along the Packet Processing Pipeline as a Processed First Packet.
- If at S401 the Instructions indicate that a Remote Resource is needed, at S403 the Processing Stage will check to see if a Resource Lock is available for the needed Remote Resource.
- If at S403 the Resource Lock is unavailable, i.e., in a LOCKED status as indicated by the corresponding Respective Local Resource Lock, at S404 the Processing Stage waits until the Resource Lock becomes available, and the Packet remains buffered in Buffer (40).
- If at S403 the Resource Lock is available, i.e., in an UNLOCKED status, the Processing Stage at S405 locks the Local Resource Lock by setting the status to LOCKED, and then carries out a Packet Processing Operation in accordance with the Instructions. According to an example embodiment, the information, that the Local Resource Lock has been set to the status of LOCKED, travels as part of the execution context of the packet which is passed from the Processing Stage, at S408.
- More particularly, at S406, the Processing Stage accesses the allocated Remote Resource, through an Engine, to Read a value from the Remote Resource. At S407, the value thus read is stored in the Packet's Execution Context. At S408, the Processed First Packet is passed along the pipeline configuration to a subsequent Processing Stage.
-
FIG. 5 . is a flow diagram of an example method according to example embodiments. The example method ofFIG. 5 may be applied to multiple example embodiments as discussed and as understood by one of ordinary skill in the art. S500 occurs when a Processing Stage accepts a Packet. The Processing Stage reads the Execution Context of the Packet and obtains Instructions based on the Execution Context. At S501, the Processing Stage determines whether a Resource is already Allocated in association with the received Packet. - At S501, if a Resource is not allocated, at S503 some Packet Processing Operation will be carried out in accordance with the Instructions. The Packet will then be passed along the Packet Processing Pipeline as a Processed Second Packet.
- At S501, if a Resource is allocated, at S502 the Processing Stage will then determine whether the Instructions indicate a WRITE Operation is needed.
- At S502, if a WRITE Operation is not needed, at S503 some other Packet Processing Operation is carried out in accordance with the Instructions.
- At S502, if a WRITE Operation is needed, at S504 the Processing Stage carries out a WRITE operation with respect to the Allocated Resource. The Processing Stage then causes both S505 and S506 to occur.
- At S505, the Processing Stage pass along the Processed Second Packet, 5505. At S506, the Processing Stage generates and sends, along the Resource Release Bus, a Resource Lock Release Request containing Identifiers of the Respective Resource Manager and of the particular Resource Lock to be released.
-
FIG. 6 . is a flow diagram of an example method according to example embodiments. Each Processing Stage is configured to monitor the Resource Release Bus (300) for any Resource Lock Release Request. S601 represents continuous monitoring of the Resource Release Bus (300) until a Resource Lock Release Request (RLRR) is received. - At S601, when a Processing Stage does receive a Resource Lock Release Request, the Processing Stage determines at S602 whether the Resource Lock Release Request is addressed to that Processing Stage.
- At S602, when the Processing Stage determines that the Resource Lock Release Request does not indicate the address of that Processing Stage, the Resource Lock Release Request is ignored.
- At S602, when a Processing Stage determines that the Resource Lock Release Request does indicate the address of that Processing Stage, at S603 the Resource Lock Release Request is implemented so as to release whichever Local Resource Lock is identified in the Resource Lock Release Request.
- It is to be noted that, in an example embodiment, the use of the Local Resource Lock within the Processing Stage allows the treatment of multiple engines as a single resource. In such an example embodiment, multiple engines are locked by setting the status of a single Local Resource Lock to LOCKED.
-
FIG. 7 is a highly simplified schematic diagram of a device according to example embodiments, in which the Resource Locks are centrally managed by a Central Resource Lock Engine (700-0). InFIG. 7 , a Packet Processing Device (1700) includes a Packet Bus (200), a Plurality of Packet Streams (20), a Plurality of Engines (80), a plurality of Remote Resources (85-0, 85-1, . . . 85-n), a First Processing Stage (760-0), a Second Processing Stage (760-1), a Third Processing Stage (760-2), a Fourth Processing Stage (760-3), a Central Resource Lock Engine (700-0). The Central Resource Lock Engine (700-0) includes a Central Resource Lock Pool Local to Engine (780) which further includes an Engine Local Resource Lock (750-0, 750-1, . . . 750-n). - According to this example embodiment, the LOCK, READ, WRITE, and RELEASE Operations are each implemented by a separate Processing Stage. In response to a request for a resource designated by a Packet (20-0), the First Processing Stage (760-0) determines if a Central Resource Lock Pool Local to Engine (780) contains an Engine Local Resource Lock (750-0, 750-1, . . . 750-n) capable of locking some requested Resource. If available, the First Processing Stage (760-0) LOCKs the Resource. The First Processing Stage (760-0) passes Packet (20-0) along the Packet Bus (200) to the Second Processing Stage (760-1) as Packet (20-1) containing information that a Resource has been locked.
- The Second Processing Stage (760-1) is configured to READ from the locked Resource, through the Plurality of Engines (80), in response to the Packet (20-1). The Second Processing Stage (760-1) is further configured to pass the Packet (20-1) along the Packet Bus (200) to the Third Processing Stage (760-2) as the Packet (20-2), which contains information from the READ operation.
- The Third Processing Stage (760-2) is configured to WRITE to a Resource, through the Plurality of Engines (80), in response to the Packet (20-2). The Third Processing Stage (760-2) is further configured to pass the Packet (20-2) along the Packet Bus (200) to the Fourth Processing Stage (760-3) as the Packet (20-3), which contains information from the WRITE operations.
- The Fourth Processing Stage (760-3) is configured to generate a Resource Lock Release Request in response to the Packet (20-3). The Central Resource Lock Engine (700-0) is configured to receive the Resource Lock Release Request and subsequently release the Resource Lock from the Central Resource Lock Pool Local to Engine (780). The allocated resource has been modified and is now available to be reallocated. The Fourth Processing Stage (760-3) then passes the Packet (20-3) along the Packet Bus (200) as the Packet (20-4).
- In this example embodiment, the Central Resource Lock Engine (700-0) can avoid contention for Remote Resources without employing a Resource Release Bus (300). On the other hand, since the Resource Locks are not local to the Plurality of Processing Stages (60), the Lock & Read Operation (105) cannot be performed within a single Processing Stage. Likewise, the Write & Release Operation (106) also cannot be performed within single Processing Stage.
-
FIG. 8 is a highly simplified schematic diagram of a device according to example embodiments. In this example embodiment, the Resource Locks are local to the Engines and not local to the Processing Stages. InFIG. 8 , Packet Processing Device (1800) includes a Packet Bus (200), a Packet (20-0, -1, -2) of a Plurality of Parallel Packet Streams (20), a First Processing Stage (860-0), a Second Processing Stage (860-1), Remote Resources (85-0, -1, . . . -n), and a Plurality of Engines (80) containing Resource Lock Engines (880-0, 1, . . . -n). - Each of the Resource Lock Engines (880-0, -1, . . . -n) contains a Respective Resource Lock Pool Local to each Engine (870-0, -1, . . . -n), and each Respective Resource Lock Pool Local to each Engine (870-0, -1, . . . -n) further contains Respective Engine Local Resource Locks (850-0, 851-0, . . . 85 n-0; 850-1, 851-a, . . . 85 n−1; . . . ; and 850-f, a851-f, 85 n-f) according to an example embodiment.
- The First Processing Stage (860-0) is configured to perform a Lock & Read Operation (105) in response to a Packet (20-0). The Lock & Read Operation (105) includes accessing a Resource Lock Engine (880-1) to access a Resource Lock Pool Local to Engine (870-1) to request availability of an Engine Local Resource Lock (850-1) corresponding to a Remote Resource (85-1), according to an example embodiment.
- The First Processing Stage (860-0) is further configured to LOCK the Remote Resource (85-1) if the Engine Local Resource Locks (850-1) is available. The First Processing Stage (860-0) is further configured to READ a value from the Remote Resource (85-1) and store information of that value into the Execution Context of the Packet (20-0). The First Processing Stage (860-0) then passes the Packet (20-0) along the Packet Bus (200) as a Packet (20-1).
- The Second Processing Stage (860-1) is configured to accept the Packet (20-1) and subsequently perform a Write & Release Operation (106) associated with the Remote Resource (85-1) through the Resource Lock Engine (880-1). The Second Processing Stage (860-1) is configured to WRITE to the Remote Resource (85-1) and subsequently send a Resource Lock Release Request to the Resource Lock Engine (880-1). The allocated resource has been modified and is now available to be reallocated.
- In the example embodiment of
FIG. 8 , the Lock & Read Operation (105) can be performed within a single Processing Stage, as can the Write & Release Operation (106). The single Processing Stage, according to an example embodiment, may perform a Lock & Read Operation (105) or a Write & Release Operation (106) within a single stage. The Resource Release Bus (300) is not required. On the other hand, since the Resource Locks are local to each of the Engines, the problem of contention is avoided only for operations seeking the same Remote Resource through the same Engine. - A hybrid architecture implementing a soft path and a hard path will now be discussed in the context of
FIGS. 2 and 3 , according to example embodiments. - First, the concept of a “path” will be discussed. In
FIG. 2 , a path is to be understood as the collection of processing stage resources, and the associated packet processing operations, through which a packet passes during its time traversing the pipeline configuration. For example, inFIG. 2 , the path through which packets labeled (P —0,P —0′,P —0″, . . . P—0f) pass includes Respective Processor (30-0) and Buffer (40-0) of Processing Stage (60-0); Respective Processor (30-1) and Buffer (40-1) of Processing Stage (60-1); and so on until the path is concluded with Respective Processor (30-n) and Buffer (40-f) of Processing Stage (60-f). - As shown in
FIG. 3 , the Respective Processor (30-0, 1, . . . n) is implemented in an example embodiment as a plurality of individual processors, each of which is part of the path through which packets labeled (P —0,P —0′,P —0″, . . . P—0f) pass. Similarly, another path exists for the packets labeled (P —1,P —1′,P —1″, . . . P—1f) through Respective Processor (31-0) and Buffer (41-0) of Processing Stage (60-0) and on through Respective Processor (31-n) and Buffer (41-f) of Processing Stage (60-f). Still other paths are apparent inFIG. 2 .FIGS. 2 and 3 , according to an example embodiment, thus illustrate a packet processing device through which multiple paths exist. Multiple streams of packets are handled at the same time along the multiple paths, i.e., are processed in parallel. - Having explained in more detail the concept of a path, and having explained that an example embodiment provides for multiple paths that process respective packet streams in parallel, some differences between soft and hard paths will now be explained. One difference relates to latency, and the other relates to delivery. Hard paths provide fixed latency, and soft paths do not provide fixed latency. A hard path therefore completes the processing of every hard path packet within a certain, well-defined time period, if at all. A soft path does not achieve the processing of soft path packets in any particular time period, and the soft path packet processing is typically implemented on a best effort basis so as not to impede the processing of hard path packets. On the other hand, soft paths guarantee the processing of soft path packets will complete (i.e., guaranteed delivery), while hard paths sometimes drop hard path packets.
- In a hybrid architecture, a packet processing device includes hard paths and one or more soft paths.
- In the hybrid architecture, according to an example embodiment, hard path packets under certain circumstances are dropped and not fully processed through the pipeline configuration. That is to say, if the processing for a hard path packet requires a Remote Resource (85-0, -1, . . . -n), but the corresponding Local Resource Lock (50) has a LOCKED state, the hard path packet is dropped. Dropping the hard path packet under such circumstances guarantees that the fixed processing latency is not subverted failure to obtain a resource allocation.
- In the hybrid architecture, according to an example embodiment, one or more soft paths are configured to allow for a pause in soft path processing in response to a soft path packet being unable to allocate a Remote Resource. That is to say, if the processing for a soft path packet requires a Remote Resource (85-0, -1, . . . -n), but the corresponding Local Resource Lock (50) has a LOCKED state, the soft path packet remains in Buffer (40) for that particular Processing Stage until the Local Resource Lock (50) has an UNLOCKED state.
- According to an example embodiment, in a packet processing device having the hybrid architecture, the Local Resource Lock (50) acts as a semaphore indicating whether the processing for a soft path packet should be paused. That is to say, when a Local Resource Lock (50) corresponding to a Remote Resource (85-0, -1, . . . -n) required for the subsequent processing of a soft path packet has a LOCKED state, the processing for the soft path packet is thus paused. A beneficial effect of such a pause is that it mitigates the possibility that the processing for a soft path packet might adversely affect the processing of hard path packets.
- To put this another way, in an example embodiment, the use of Local Resource Lock (50) provides a mechanism that permits processing of soft path packets to be interleaved with the processing of hard path packets in such a manner that the soft path packet processing is not carried out when remote resources are already allocated to hard path processing, or to other soft path processing.
- According to example embodiments, an Interface Arbiter (120) directs data packets through one of the hard paths, and directs control and management packets along the one or more soft paths.
- According to other example embodiments, a packet, for which a Remote Resource must be accessed as part of its processing, is directed along a soft path by Interface Arbiter (120). By directing such packets requiring access to a Remote Resource along the soft path, the interruption of the processing of data packets along the hard path is avoided.
- Although the inventive concept has been described above with respect to the various embodiments, it is noted that there can be a variety of permutations and modifications of the described features by those who are familiar with this field, without departing from the technical ideas and scope of the features, which shall be defined by the appended claims.
- Further, while this specification contains many features, the features should not be construed as limitations on the scope of the disclosure or the appended claims. Certain features described in the context of separate embodiments can also be implemented in combination. Conversely, various features described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable sub-combination.
- Although the drawings describe operations in a specific order and/or show specific arrangements of components, one should not interpret that such specific order and/or arrangements are limited, or that all the operations performed and the components disclosed are needed to obtain a desired result. Accordingly, other implementations are within the scope of the following claims.
Claims (20)
1. A packet processing device comprising:
a plurality of processing stages, including a first processing stage and a second processing stage arranged as a packet processing pipeline;
the first processing stage and the second processing stage each comprising:
a respective processor configured to process a packet of a packet stream, and
a respective resource manager having a respective local resource lock corresponding to a remote resource;
the respective processor being further configured to request the respective resource manager to allocate the remote resource;
the respective resource manager being further configured to respond to the request to allocate the remote resource by locking the remote resource with the respective local resource lock and allocating the remote resource;
the respective processor being further configured to implement a packet processing operation associated with the allocated remote resource.
2. The packet processing device according to claim 1 , wherein the first processing stage and the second processing stage are further configured to perform the packet processing operation, in association with the allocated remote resource, before passing the packet to a next one of the plurality of processing stages along the processing pipeline.
3. The packet processing device according to claim 1 , further comprising:
a resource release bus connected to the plurality of processing stages; and
the respective resource manager being further configured to:
send to upstream ones of the plurality of processing stages a release request message; and
release the remote resource with the respective local resource lock in response to receiving the release request message.
4. The packet processing device according to claim 2 , wherein the respective resource manager is further configured to:
send a release request message comprising identifiers for an identified resource manager and for an identified resource lock.
5. The packet processing device according to claim 2 , wherein the respective resource manager is further configured to:
receive a release request message comprising identifiers for an identified resource manager and an identified resource lock.
6. The packet processing device according to claim 1 , wherein the respective resource manager is further configured to:
detect when the remote resource cannot be locked; and
pause implementation of the packet processing operation until the remote resource is locked.
7. The packet processing device according to claim 1 , further comprising the plurality of processing stages being connected in series within a same integrated circuit.
8. A packet processing device, comprising:
a plurality of processing stages arranged as a packet processing pipeline;
the processing stages each having processor cores and buffers;
the processor cores and buffers of the processing stages defining a plurality of paths, for simultaneous packet processing, through the packet processing pipeline;
an ingress front end configured to direct each packet of an incoming stream of packets into one of the plurality of paths;
the plurality of paths including a hard path and a soft path, the hard path being configured to process received ones of the incoming stream of packets with fixed latency, the soft path being configured to process received ones of the incoming stream of packets with variable latency; and
the processing stages each further including a respective resource manager configured to request allocation of a remote resource, for a given packet of the incoming stream of packets in the soft path, in response to an instruction from one of the processor cores, whereby the remote resource is allocated for processing of only the given packet.
9. The packet processing device as set forth in claim 8 , further comprising the respective resource manager requesting allocation of the remote resource by referring to a semaphore locally available in the one of the plurality of processing stages in which the respective resource manager is included.
10. The packet processing device as set forth in claim 8 , further comprising:
a semaphore engine storing a plurality of semaphores, each corresponding to one of a plurality of remote resources; and
the respective resource manager requesting allocation of one of the plurality of remote resources by referring to a semaphore controlled by the semaphore engine.
11. The packet processing device as set forth in claim 8 , further comprising:
a plurality of engines configured to interface each said respective resource manager with one of a plurality of remote resources; and
the respective resource manager requesting allocation of one of the plurality of remote resources by referring to a semaphore controlled by one of the plurality of engines.
12. A packet processing method, for a processing stage of a pipeline packet processing device, comprising:
receiving, at a processor of a first processing stage, a first packet and a request for allocation of a remote resource;
responding, by the processor, to the allocation request, by setting a semaphore corresponding to the remote resource to indicate a locked status;
implementing a first packet processing operation, in association with the allocated remote resource, and in association with the first packet, to obtain a processed first packet; and
outputting the processed first packet to a next processing stage of the pipeline packet processing device.
13. The method of claim 12 , further comprising:
receiving a second packet and an indication of an allocation of a remote resource;
implementing a second packet processing operation, associated with the allocated remote resource, on the second packet, to obtain a second processed packet;
requesting the release of the remote resource; and
outputting the second processed packet to a next processing stage of the pipeline packet processing device.
14. The method of claim 12 , further comprising outputting a remote resource release indication to an upstream processing stage of the pipeline packet processing device in response to a completion of the second packet processing operation associated with the allocated remote resource.
15. The method of claim 12 , further comprising obtaining the indication of the allocation of the remote resource from an execution context received together with the processed first packet.
16. A packet processing device comprising:
a plurality of processing stages, arranged as a packet processing pipeline, configured to process packets in a first direction along the packet processing pipeline;
a lock release message path connecting the plurality of processing stages, configured to communicate along a second direction opposite to said first direction;
each of the plurality of processing stages including:
a respective local resource lock, and
a respective processor core; and
the respective processor core being configured to send, along the lock release message path, a resource lock release request message identifying one of the plurality of processing stages and identifying a resource to be released.
17. The packet processing device as set forth in claim 16 , further comprising a local lock buffer configured to buffer a packet when the respective local resource lock indicates that a corresponding resource is not unlocked.
18. The packet processing device as set forth in claim 16 , wherein each of the plurality of processing stages further comprises a plurality of processor cores, the processor cores and buffers of the plurality of processing stages defining a plurality of paths, for simultaneous packet processing, through the packet processing pipeline.
19. The packet processing device as set forth in claim 18 , further comprising an ingress front end configured to direct each of the packets into one of the plurality of paths.
20. The packet processing device as set forth in claim 19 , further comprising the plurality of paths being configured as a hard path and a soft path, the hard path being configured to process received ones of the packets with fixed latency, and the soft path being configured to process received ones of the packets with variable latency, the ingress front end directing a packet requiring a remote resource to only the soft path.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/973,744 US20140233582A1 (en) | 2012-08-29 | 2013-08-22 | Semaphore soft and hard hybrid architecture |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201261694483P | 2012-08-29 | 2012-08-29 | |
US201361753767P | 2013-01-17 | 2013-01-17 | |
US13/973,744 US20140233582A1 (en) | 2012-08-29 | 2013-08-22 | Semaphore soft and hard hybrid architecture |
Publications (1)
Publication Number | Publication Date |
---|---|
US20140233582A1 true US20140233582A1 (en) | 2014-08-21 |
Family
ID=50187555
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/973,744 Abandoned US20140233582A1 (en) | 2012-08-29 | 2013-08-22 | Semaphore soft and hard hybrid architecture |
US13/973,751 Active 2034-02-27 US9525621B2 (en) | 2012-08-29 | 2013-08-22 | Semaphore soft and hard hybrid architecture |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/973,751 Active 2034-02-27 US9525621B2 (en) | 2012-08-29 | 2013-08-22 | Semaphore soft and hard hybrid architecture |
Country Status (2)
Country | Link |
---|---|
US (2) | US20140233582A1 (en) |
CN (1) | CN103677755A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10110233B2 (en) * | 2016-06-23 | 2018-10-23 | Altera Corporation | Methods for specifying processor architectures for programmable integrated circuits |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140233582A1 (en) * | 2012-08-29 | 2014-08-21 | Marvell World Trade Ltd. | Semaphore soft and hard hybrid architecture |
CN103984544A (en) * | 2014-04-24 | 2014-08-13 | 浪潮电子信息产业股份有限公司 | Method for avoiding simultaneous execution of tasks |
US9575804B2 (en) * | 2015-03-27 | 2017-02-21 | Commvault Systems, Inc. | Job management and resource allocation |
US11972292B1 (en) * | 2020-03-30 | 2024-04-30 | XConn Technologies Holdings, Inc. | Interconnect switch using multi-bank scheduling |
US11934318B2 (en) | 2022-06-07 | 2024-03-19 | XConn Technologies Holdings, Inc. | Multi-host memory sharing |
US11947483B2 (en) | 2022-06-17 | 2024-04-02 | XConn Technologies Holdings, Inc. | Data flow management |
Citations (64)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5123106A (en) * | 1987-09-24 | 1992-06-16 | Nec Corporation | Multiprocessor system with shared memory includes primary processor which selectively accesses primary local memory and common memories without using arbiter |
US5353414A (en) * | 1989-06-30 | 1994-10-04 | Nec Corporation | Bus lock control apparatus capable of controlling without stopping bus arbiters |
US5361334A (en) * | 1988-01-15 | 1994-11-01 | Quantel Limited | Data processing and communication |
US5499356A (en) * | 1989-12-29 | 1996-03-12 | Cray Research, Inc. | Method and apparatus for a multiprocessor resource lockout instruction |
US5699500A (en) * | 1995-06-01 | 1997-12-16 | Ncr Corporation | Reliable datagram service provider for fast messaging in a clustered environment |
US6127871A (en) * | 1997-11-18 | 2000-10-03 | Hewlett-Packard Company | Variable digital delay line |
US6170025B1 (en) * | 1997-08-29 | 2001-01-02 | Intel Corporation | Distributed computer system supporting remote interrupts and lock mechanism |
US6199094B1 (en) * | 1998-06-05 | 2001-03-06 | International Business Machines Corp. | Protecting shared resources using mutex striping |
US20010037421A1 (en) * | 1999-12-29 | 2001-11-01 | Intel Corporation | Enhanced highly pipelined bus architecture |
US20020010793A1 (en) * | 1997-08-22 | 2002-01-24 | Michael Noll | Method and apparatus for performing frame processing for a network |
US20020083149A1 (en) * | 2000-12-22 | 2002-06-27 | International Business Machines Corporation | Method for deadlock avoidance in a cluster environment |
US6460174B1 (en) * | 1999-10-01 | 2002-10-01 | Stmicroelectronics, Ltd. | Methods and models for use in designing an integrated circuit |
US20020174305A1 (en) * | 2000-12-28 | 2002-11-21 | Vartti Kelvin S. | Method and apparatus for controlling memory storage locks based on cache line ownership |
US20020181440A1 (en) * | 2001-06-01 | 2002-12-05 | Norman Richard S. | Cell-based switch fabric with cell-to-line-card control for regulating injection of packets |
US6529983B1 (en) * | 1999-11-03 | 2003-03-04 | Cisco Technology, Inc. | Group and virtual locking mechanism for inter processor synchronization |
US6532478B1 (en) * | 1999-07-14 | 2003-03-11 | Fujitsu Limited | File loader in information processing system of multiprocessor configuration |
US6542926B2 (en) * | 1998-06-10 | 2003-04-01 | Compaq Information Technologies Group, L.P. | Software partitioned multi-processor system with flexible resource sharing levels |
US6549961B1 (en) * | 1999-10-27 | 2003-04-15 | Infineon Technologies North America Corporation | Semaphore access in a multiprocessor system |
US20030101307A1 (en) * | 2001-03-15 | 2003-05-29 | Riccardo Gemelli | System of distributed microprocessor interfaces toward macro-cell based designs implemented as ASIC or FPGA bread boarding and relative common bus protocol |
US20030103499A1 (en) * | 2001-11-21 | 2003-06-05 | Tom Davis | Configurable packet processor |
US20040088488A1 (en) * | 2002-11-01 | 2004-05-06 | Infineon Technologies North America Corp. | Multi-threaded embedded processor using deterministic instruction memory to guarantee execution of pre-selected threads during blocking events |
US20050038964A1 (en) * | 2003-08-14 | 2005-02-17 | Hooper Donald F. | Folding for a multi-threaded network processor |
US20050076169A1 (en) * | 2000-12-22 | 2005-04-07 | Modelski Richard P. | External memory engine selectable pipeline architecture |
US6898687B2 (en) * | 2002-12-13 | 2005-05-24 | Sun Microsystems, Inc. | System and method for synchronizing access to shared resources |
US20050207387A1 (en) * | 2004-02-09 | 2005-09-22 | Semtech Corporation | Method and apparatus for aligning time references when separated by an unreliable data packet network |
US20050235134A1 (en) * | 2002-08-07 | 2005-10-20 | Mmagix Technology Limited | Apparatus, method and system for a synchronicity independent, resource delegating, power and instruction optimizing processor |
US6973521B1 (en) * | 2000-05-16 | 2005-12-06 | Cisco Technology, Inc. | Lock controller supporting blocking and non-blocking requests |
US20060140119A1 (en) * | 2004-12-29 | 2006-06-29 | Alcatel | Predictive congestion management in a data communications switch using traffic and system statistics |
US20060171386A1 (en) * | 2004-09-01 | 2006-08-03 | Interactic Holdings, Llc | Means and apparatus for a scaleable congestion free switching system with intelligent control III |
US7099275B2 (en) * | 2001-09-21 | 2006-08-29 | Slt Logic Llc | Programmable multi-service queue scheduler |
US20060206635A1 (en) * | 2005-03-11 | 2006-09-14 | Pmc-Sierra, Inc. | DMA engine for protocol processing |
US7131120B2 (en) * | 2002-05-16 | 2006-10-31 | Sun Microsystems, Inc. | Inter Java virtual machine (JVM) resource locking mechanism |
US7219342B2 (en) * | 2000-08-07 | 2007-05-15 | Altera Corporation | Software-to-hardware compiler |
US7233182B1 (en) * | 2004-12-10 | 2007-06-19 | Marvell International Ltd. | Circuitry for eliminating false lock in delay-locked loops |
US7266646B2 (en) * | 2003-05-05 | 2007-09-04 | Marvell International Ltd. | Least mean square dynamic cache-locking |
US20080219175A1 (en) * | 2007-03-02 | 2008-09-11 | Paul Stephan Bedrosian | Connectionless configurations for stress testing timing and synchronization in data packet networks |
US20080244150A1 (en) * | 2007-02-02 | 2008-10-02 | Viswa Sharma | Processor chip arcitecture having integrated high-speed packet switched serial interface |
US20080282008A1 (en) * | 2002-09-30 | 2008-11-13 | Marvell International Ltd. | System and Apparatus for Early Fixed Latency Subtractive Decoding |
US20090080339A1 (en) * | 2007-09-20 | 2009-03-26 | Nicholas Geoffrey Duffield | Multicast-based inference of temporal delay characteristics in packet data networks |
US20090161685A1 (en) * | 2007-02-10 | 2009-06-25 | Huawei Technologies Co., Ltd. | Method, system and node for backpressure in multistage switching network |
US7743191B1 (en) * | 2007-12-20 | 2010-06-22 | Pmc-Sierra, Inc. | On-chip shared memory based device architecture |
US7757070B1 (en) * | 2007-04-10 | 2010-07-13 | Marvell International Ltd. | Methods, apparatuses, and system for facilitating control of multiple instruction threads |
US20100208614A1 (en) * | 2007-10-19 | 2010-08-19 | Harmatos Janos | Method and arrangement for scheduling data packets in a communication network system |
US7996642B1 (en) * | 2007-04-25 | 2011-08-09 | Marvell International Ltd. | Digital locked loop on channel tagged memory requests for memory optimization |
US20110289180A1 (en) * | 2009-04-27 | 2011-11-24 | Lsi Corporation | Data caching in a network communications processor architecture |
US20110302345A1 (en) * | 2010-06-03 | 2011-12-08 | Philippe Boucard | Network on chip (noc) with qos features |
US20120011351A1 (en) * | 2010-07-08 | 2012-01-12 | Texas Instruments Incorporated | Security Processing Engines, Circuits and Systems and Adaptive Processes and Other Processes |
US8127262B1 (en) * | 2008-12-18 | 2012-02-28 | Xilinx, Inc. | Communicating state data between stages of pipelined packet processor |
US20120066432A1 (en) * | 2009-05-22 | 2012-03-15 | Hitachi, Ltd. | Semiconductor Device |
US20120079495A1 (en) * | 2010-09-29 | 2012-03-29 | International Business Machines Corporation | Managing access to a shared resource in a data processing system |
US20120117625A1 (en) * | 2010-11-09 | 2012-05-10 | Microsoft Corporation | Security system for computing resources pre-releases |
US20120151153A1 (en) * | 2009-07-03 | 2012-06-14 | Axel JANTSCH | Programmable Controller |
US20120216012A1 (en) * | 2008-10-15 | 2012-08-23 | Hyperion Core, Inc. | Sequential processor comprising an alu array |
US8332831B1 (en) * | 2000-08-07 | 2012-12-11 | Altera Corporation | Software-to-hardware compiler with symbol set inference analysis |
US20130003556A1 (en) * | 2011-06-28 | 2013-01-03 | Xelerated Ab | Scheduling packets in a packet-processing pipeline |
US20130028079A1 (en) * | 2011-07-26 | 2013-01-31 | Telefonaktiebolaget L M Ericsson (Publ) | Systems and methods for resource booking for admission control and scheduling |
US20130100948A1 (en) * | 2011-10-21 | 2013-04-25 | Garrick Thomas Irvine | Methods and Apparatuses for Maintaining Synchronization Between a Radio Equipment Controller and an Item of Radio Equipment |
US20130170435A1 (en) * | 2012-01-04 | 2013-07-04 | Esmael Hejazi Dinan | Network Site for Wireless Communications |
US20130301408A1 (en) * | 2012-05-10 | 2013-11-14 | Marvell World Trade Ltd. | Hybrid dataflow processor |
US20130315259A1 (en) * | 2012-05-22 | 2013-11-28 | Marvell World Trade Ltd. | Method and Apparatus for Internal/External Memory Packet and Byte Counting |
US20130316728A1 (en) * | 2012-03-21 | 2013-11-28 | Digimarc Corporation | Positioning systems for wireless networks |
US20140064271A1 (en) * | 2012-08-29 | 2014-03-06 | Marvell World Trade Ltd. | Semaphore soft and hard hybrid architecture |
US8688922B1 (en) * | 2010-03-11 | 2014-04-01 | Marvell International Ltd | Hardware-supported memory management |
US20140247835A1 (en) * | 2002-04-04 | 2014-09-04 | Marvell International Ltd. | System and method for modifying, in a processing pipeline, a length of a data packet in a data block without modifying a length of the data block |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS60107170A (en) * | 1983-11-15 | 1985-06-12 | Nec Corp | Multiprocessor control system |
US7466653B1 (en) * | 2004-06-30 | 2008-12-16 | Marvell International Ltd. | Quality of service for a stackable network switch |
US7653804B2 (en) * | 2006-01-26 | 2010-01-26 | Xilinx, Inc. | Resource sharing in multiple parallel pipelines |
WO2010088680A1 (en) * | 2009-02-02 | 2010-08-05 | Research In Motion Limited | Indication of uplink semi-persistent scheduling explicit release using a downlink physical downlink control channel |
JP5463076B2 (en) * | 2009-05-28 | 2014-04-09 | パナソニック株式会社 | Multithreaded processor |
US20130080672A1 (en) * | 2011-09-27 | 2013-03-28 | Kaminario Technologies Ltd. | System, method and computer program product for access control |
US8880687B1 (en) * | 2012-02-06 | 2014-11-04 | Netapp, Inc. | Detecting and managing idle virtual storage servers |
US9141430B2 (en) * | 2012-04-30 | 2015-09-22 | Hewlett-Packard Development Company, L.P. | Scheduling mapreduce job sets |
-
2013
- 2013-08-22 US US13/973,744 patent/US20140233582A1/en not_active Abandoned
- 2013-08-22 US US13/973,751 patent/US9525621B2/en active Active
- 2013-08-29 CN CN201310389251.9A patent/CN103677755A/en active Pending
Patent Citations (64)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5123106A (en) * | 1987-09-24 | 1992-06-16 | Nec Corporation | Multiprocessor system with shared memory includes primary processor which selectively accesses primary local memory and common memories without using arbiter |
US5361334A (en) * | 1988-01-15 | 1994-11-01 | Quantel Limited | Data processing and communication |
US5353414A (en) * | 1989-06-30 | 1994-10-04 | Nec Corporation | Bus lock control apparatus capable of controlling without stopping bus arbiters |
US5499356A (en) * | 1989-12-29 | 1996-03-12 | Cray Research, Inc. | Method and apparatus for a multiprocessor resource lockout instruction |
US5699500A (en) * | 1995-06-01 | 1997-12-16 | Ncr Corporation | Reliable datagram service provider for fast messaging in a clustered environment |
US20020010793A1 (en) * | 1997-08-22 | 2002-01-24 | Michael Noll | Method and apparatus for performing frame processing for a network |
US6170025B1 (en) * | 1997-08-29 | 2001-01-02 | Intel Corporation | Distributed computer system supporting remote interrupts and lock mechanism |
US6127871A (en) * | 1997-11-18 | 2000-10-03 | Hewlett-Packard Company | Variable digital delay line |
US6199094B1 (en) * | 1998-06-05 | 2001-03-06 | International Business Machines Corp. | Protecting shared resources using mutex striping |
US6542926B2 (en) * | 1998-06-10 | 2003-04-01 | Compaq Information Technologies Group, L.P. | Software partitioned multi-processor system with flexible resource sharing levels |
US6532478B1 (en) * | 1999-07-14 | 2003-03-11 | Fujitsu Limited | File loader in information processing system of multiprocessor configuration |
US6460174B1 (en) * | 1999-10-01 | 2002-10-01 | Stmicroelectronics, Ltd. | Methods and models for use in designing an integrated circuit |
US6549961B1 (en) * | 1999-10-27 | 2003-04-15 | Infineon Technologies North America Corporation | Semaphore access in a multiprocessor system |
US6529983B1 (en) * | 1999-11-03 | 2003-03-04 | Cisco Technology, Inc. | Group and virtual locking mechanism for inter processor synchronization |
US20010037421A1 (en) * | 1999-12-29 | 2001-11-01 | Intel Corporation | Enhanced highly pipelined bus architecture |
US6973521B1 (en) * | 2000-05-16 | 2005-12-06 | Cisco Technology, Inc. | Lock controller supporting blocking and non-blocking requests |
US8332831B1 (en) * | 2000-08-07 | 2012-12-11 | Altera Corporation | Software-to-hardware compiler with symbol set inference analysis |
US7219342B2 (en) * | 2000-08-07 | 2007-05-15 | Altera Corporation | Software-to-hardware compiler |
US20020083149A1 (en) * | 2000-12-22 | 2002-06-27 | International Business Machines Corporation | Method for deadlock avoidance in a cluster environment |
US20050076169A1 (en) * | 2000-12-22 | 2005-04-07 | Modelski Richard P. | External memory engine selectable pipeline architecture |
US20020174305A1 (en) * | 2000-12-28 | 2002-11-21 | Vartti Kelvin S. | Method and apparatus for controlling memory storage locks based on cache line ownership |
US20030101307A1 (en) * | 2001-03-15 | 2003-05-29 | Riccardo Gemelli | System of distributed microprocessor interfaces toward macro-cell based designs implemented as ASIC or FPGA bread boarding and relative common bus protocol |
US20020181440A1 (en) * | 2001-06-01 | 2002-12-05 | Norman Richard S. | Cell-based switch fabric with cell-to-line-card control for regulating injection of packets |
US7099275B2 (en) * | 2001-09-21 | 2006-08-29 | Slt Logic Llc | Programmable multi-service queue scheduler |
US20030103499A1 (en) * | 2001-11-21 | 2003-06-05 | Tom Davis | Configurable packet processor |
US20140247835A1 (en) * | 2002-04-04 | 2014-09-04 | Marvell International Ltd. | System and method for modifying, in a processing pipeline, a length of a data packet in a data block without modifying a length of the data block |
US7131120B2 (en) * | 2002-05-16 | 2006-10-31 | Sun Microsystems, Inc. | Inter Java virtual machine (JVM) resource locking mechanism |
US20050235134A1 (en) * | 2002-08-07 | 2005-10-20 | Mmagix Technology Limited | Apparatus, method and system for a synchronicity independent, resource delegating, power and instruction optimizing processor |
US20080282008A1 (en) * | 2002-09-30 | 2008-11-13 | Marvell International Ltd. | System and Apparatus for Early Fixed Latency Subtractive Decoding |
US20040088488A1 (en) * | 2002-11-01 | 2004-05-06 | Infineon Technologies North America Corp. | Multi-threaded embedded processor using deterministic instruction memory to guarantee execution of pre-selected threads during blocking events |
US6898687B2 (en) * | 2002-12-13 | 2005-05-24 | Sun Microsystems, Inc. | System and method for synchronizing access to shared resources |
US7266646B2 (en) * | 2003-05-05 | 2007-09-04 | Marvell International Ltd. | Least mean square dynamic cache-locking |
US20050038964A1 (en) * | 2003-08-14 | 2005-02-17 | Hooper Donald F. | Folding for a multi-threaded network processor |
US20050207387A1 (en) * | 2004-02-09 | 2005-09-22 | Semtech Corporation | Method and apparatus for aligning time references when separated by an unreliable data packet network |
US20060171386A1 (en) * | 2004-09-01 | 2006-08-03 | Interactic Holdings, Llc | Means and apparatus for a scaleable congestion free switching system with intelligent control III |
US7233182B1 (en) * | 2004-12-10 | 2007-06-19 | Marvell International Ltd. | Circuitry for eliminating false lock in delay-locked loops |
US20060140119A1 (en) * | 2004-12-29 | 2006-06-29 | Alcatel | Predictive congestion management in a data communications switch using traffic and system statistics |
US20060206635A1 (en) * | 2005-03-11 | 2006-09-14 | Pmc-Sierra, Inc. | DMA engine for protocol processing |
US20080244150A1 (en) * | 2007-02-02 | 2008-10-02 | Viswa Sharma | Processor chip arcitecture having integrated high-speed packet switched serial interface |
US20090161685A1 (en) * | 2007-02-10 | 2009-06-25 | Huawei Technologies Co., Ltd. | Method, system and node for backpressure in multistage switching network |
US20080219175A1 (en) * | 2007-03-02 | 2008-09-11 | Paul Stephan Bedrosian | Connectionless configurations for stress testing timing and synchronization in data packet networks |
US7757070B1 (en) * | 2007-04-10 | 2010-07-13 | Marvell International Ltd. | Methods, apparatuses, and system for facilitating control of multiple instruction threads |
US7996642B1 (en) * | 2007-04-25 | 2011-08-09 | Marvell International Ltd. | Digital locked loop on channel tagged memory requests for memory optimization |
US20090080339A1 (en) * | 2007-09-20 | 2009-03-26 | Nicholas Geoffrey Duffield | Multicast-based inference of temporal delay characteristics in packet data networks |
US20100208614A1 (en) * | 2007-10-19 | 2010-08-19 | Harmatos Janos | Method and arrangement for scheduling data packets in a communication network system |
US7743191B1 (en) * | 2007-12-20 | 2010-06-22 | Pmc-Sierra, Inc. | On-chip shared memory based device architecture |
US20120216012A1 (en) * | 2008-10-15 | 2012-08-23 | Hyperion Core, Inc. | Sequential processor comprising an alu array |
US8127262B1 (en) * | 2008-12-18 | 2012-02-28 | Xilinx, Inc. | Communicating state data between stages of pipelined packet processor |
US20110289180A1 (en) * | 2009-04-27 | 2011-11-24 | Lsi Corporation | Data caching in a network communications processor architecture |
US20120066432A1 (en) * | 2009-05-22 | 2012-03-15 | Hitachi, Ltd. | Semiconductor Device |
US20120151153A1 (en) * | 2009-07-03 | 2012-06-14 | Axel JANTSCH | Programmable Controller |
US8688922B1 (en) * | 2010-03-11 | 2014-04-01 | Marvell International Ltd | Hardware-supported memory management |
US20110302345A1 (en) * | 2010-06-03 | 2011-12-08 | Philippe Boucard | Network on chip (noc) with qos features |
US20120011351A1 (en) * | 2010-07-08 | 2012-01-12 | Texas Instruments Incorporated | Security Processing Engines, Circuits and Systems and Adaptive Processes and Other Processes |
US20120079495A1 (en) * | 2010-09-29 | 2012-03-29 | International Business Machines Corporation | Managing access to a shared resource in a data processing system |
US20120117625A1 (en) * | 2010-11-09 | 2012-05-10 | Microsoft Corporation | Security system for computing resources pre-releases |
US20130003556A1 (en) * | 2011-06-28 | 2013-01-03 | Xelerated Ab | Scheduling packets in a packet-processing pipeline |
US20130028079A1 (en) * | 2011-07-26 | 2013-01-31 | Telefonaktiebolaget L M Ericsson (Publ) | Systems and methods for resource booking for admission control and scheduling |
US20130100948A1 (en) * | 2011-10-21 | 2013-04-25 | Garrick Thomas Irvine | Methods and Apparatuses for Maintaining Synchronization Between a Radio Equipment Controller and an Item of Radio Equipment |
US20130170435A1 (en) * | 2012-01-04 | 2013-07-04 | Esmael Hejazi Dinan | Network Site for Wireless Communications |
US20130316728A1 (en) * | 2012-03-21 | 2013-11-28 | Digimarc Corporation | Positioning systems for wireless networks |
US20130301408A1 (en) * | 2012-05-10 | 2013-11-14 | Marvell World Trade Ltd. | Hybrid dataflow processor |
US20130315259A1 (en) * | 2012-05-22 | 2013-11-28 | Marvell World Trade Ltd. | Method and Apparatus for Internal/External Memory Packet and Byte Counting |
US20140064271A1 (en) * | 2012-08-29 | 2014-03-06 | Marvell World Trade Ltd. | Semaphore soft and hard hybrid architecture |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10110233B2 (en) * | 2016-06-23 | 2018-10-23 | Altera Corporation | Methods for specifying processor architectures for programmable integrated circuits |
Also Published As
Publication number | Publication date |
---|---|
US9525621B2 (en) | 2016-12-20 |
US20140064271A1 (en) | 2014-03-06 |
CN103677755A (en) | 2014-03-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9525621B2 (en) | Semaphore soft and hard hybrid architecture | |
US9467399B2 (en) | Processing concurrency in a network device | |
US9639409B2 (en) | Device and method for communicating between cores | |
KR100834432B1 (en) | Method and apparatus for managing resource contention in a multisystem cluster | |
CN106815251A (en) | Distributed data base system, data bank access method and device | |
CN113065957B (en) | Futures trading system | |
US20040139142A1 (en) | Method and apparatus for managing resource contention | |
US9146794B2 (en) | Enhanced arbitration protocol for nodes in a cluster | |
CN104461862B (en) | Resource recovery method and apparatus after data handling system and thread collapse | |
CN108984327A (en) | Message forwarding method, multi-core CPU and the network equipment | |
US8140728B1 (en) | Data packet arbitration system | |
JP4833220B2 (en) | Method, system, and program for dividing client resources by local authority | |
US20020009098A1 (en) | Communication control method and device | |
KR101227885B1 (en) | Channel multiplexing method and apparatus in shared memory | |
US9996468B1 (en) | Scalable dynamic memory management in a network device | |
US10831561B2 (en) | Method for changing allocation of data using synchronization token | |
CN106453223A (en) | Seamless world game server system | |
US8782161B2 (en) | Method and system for offloading computation flexibly to a communication adapter | |
US7483422B2 (en) | Data processing system, method and interconnect fabric for selective link information allocation in a data processing system | |
US9015717B2 (en) | Method for processing tasks in parallel and selecting a network for communication | |
JP2013041361A (en) | Resource arbitration system and resource arbitration method | |
US8688880B2 (en) | Centralized serialization of requests in a multiprocessor system | |
JP2007241922A (en) | Arbitration method for use of shared resource, and arbitration device therefor | |
Nelson et al. | Understanding RDMA behavior in NUMA systems | |
US12099463B2 (en) | Hierarchical ring-based interconnection network for symmetric multiprocessors |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |