WO1999013405A1 - A fully-pipelined fixed-latency communications system with a real-time dynamic bandwidth allocation - Google Patents

A fully-pipelined fixed-latency communications system with a real-time dynamic bandwidth allocation Download PDF

Info

Publication number
WO1999013405A1
WO1999013405A1 PCT/US1998/015006 US9815006W WO9913405A1 WO 1999013405 A1 WO1999013405 A1 WO 1999013405A1 US 9815006 W US9815006 W US 9815006W WO 9913405 A1 WO9913405 A1 WO 9913405A1
Authority
WO
WIPO (PCT)
Prior art keywords
initiator
subsystem
bus
target
cycle
Prior art date
Application number
PCT/US1998/015006
Other languages
French (fr)
Inventor
Drew Eric Wingard
Geert Paul Rosseel
Original Assignee
Sonics, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sonics, Inc. filed Critical Sonics, Inc.
Priority to EP98936927A priority Critical patent/EP1027657B1/en
Priority to KR1020007002388A priority patent/KR100572372B1/en
Priority to DE69832410T priority patent/DE69832410T2/en
Priority to AU85764/98A priority patent/AU8576498A/en
Priority to AT98936927T priority patent/ATE310277T1/en
Priority to JP2000511115A priority patent/JP4259751B2/en
Publication of WO1999013405A1 publication Critical patent/WO1999013405A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F13/00Interconnection of, or transfer of information or other signals between, memories, input/output devices or central processing units
    • G06F13/14Handling requests for interconnection or transfer
    • G06F13/36Handling requests for interconnection or transfer for access to common bus or bus system
    • G06F13/368Handling requests for interconnection or transfer for access to common bus or bus system with decentralised access control
    • G06F13/37Handling requests for interconnection or transfer for access to common bus or bus system with decentralised access control using a physical-position-dependent priority, e.g. daisy chain, round robin or token passing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F13/00Interconnection of, or transfer of information or other signals between, memories, input/output devices or central processing units

Definitions

  • the present invention relates to a bus system to couple computing devices.
  • a model may be written for the proven subsystem that can provide accurate results when analyzing the requirements and performance of a new system design; the model for a new, unproven subsystem is likely to be neither as accurate as the proven subsystem, nor built in time to influence the design.
  • proven subsystems can serve as building blocks that simplify the overall design process by allowing the system designer to focus at a higher level of abstraction, while providing improved predictability in the resulting system implementation.
  • re-use of hardware subsystems protects the investment in software to control those subsystems, and allows the system software implementation to proceed as soon as the hardware building blocks have been chosen.
  • subsystem re-use protects the investment in verification and testing. Since the desired systems are highly integrated, the required subsystems end up deeply embedded within an integrated circuit. In deeply-embedded designs, verifying the design functionality becomes very challenging and testing an individual system to prove that it is correctly built can lead to expensive delays or costly system rework. Thus, maintenance of the integrity of subsystem verification and test is likely the single biggest gain from design re-use.
  • An essential aspect of such approaches is the communications interface the various subsystems present to one another.
  • One approach is to define customized point-to-point interfaces between a subsystem and each peer to which it must communicate. This interface style is particularly popular in applications where data flows from a first subsystem into the subsystem being designed, which processes the data and outputs the data to a second subsystem.
  • This customized approach offers protocol simplicity, guaranteed performance, and isolation from dependencies on unrelated subsystems.
  • customized interfaces by their very nature are inflexible. If a new application needs an existing subsystem to be interfaced to one that does not share its interface, design re-work is required.
  • a second approach is to define a system using standardized interfaces. It is common to see standardized interfaces paired with various forms of shared interconnect to form a communications subsystem. Many standardized interfaces are based on pre-established computer bus protocols. One significant advantage to this approach is flexibility in communications patterns: the ability for every agent to communicate with every other agent in the system. Computer buses allow flexibility in system design, since as many different agents may be connected together as required by the system, as long as the bus has sufficient performance.
  • a final advantage is the natural mapping of address/data transfers, which are fundamental to processor-memory communications, into systems where complex algorithms that are implemented in software need to communicate with hardware subsystems.
  • shared interconnect comes a requirement to allocate the communications resources among the various initiator devices and target subsystems. In the case of computer buses, resource allocation is typically referred to as arbitration.
  • a principal disadvantage of standardized bus interfaces is the uncertainty in transfer delay (normally called latency) that results from arbitration. Latency uncertainty causes trouble for subsystems that must satisfy real-time constraints, since data that does not arrive in time to meet a deadline can result in improper system behavior.
  • a second disadvantage is a lack of total available transfer opportunities (normally called bandwidth) that results from the computer system heritage. Historically, as well as presently, a bus typically is designed to support the peak transfer rate of the processor connected to it, and transfer opportunities not claimed by the processor are made available to other agents. Systems with combined bandwidths in excess of the peak processor bandwidth therefore must resort to multilevel bus schemes, or a mix of buses and dedicated connections, to separate the bandwidth.
  • a final disadvantage is the inability to effectively support communication that does not map easily into address/data transfers (for instance, handshaking signals between two agents) or communications originated by an agent that is not an initiator (for instance, an interrupt signal driven by a target subsystem to signal an initiator device that data is available).
  • subsystem communications requirements vary greatly.
  • Some subsystems such as input/output devices like keyboards, infrared remote controllers, and LED displays, have very low bandwidth requirements and are very tolerant of variable latency.
  • Other subsystems such as a RISC CPU or a digital signal processor, desire high available bandwidth and low latency to memory, but can tolerate bandwidth limits and variable latency at the expense of slower program execution.
  • Still other subsystems such as a Tl telephony link or an SVGA display, have moderate to high bandwidth requirements, but cannot tolerate uncertainties in bandwidth nor latency.
  • Such real-time subsystems need guaranteed bandwidth and latency, and there is no advantage to providing any extra performance.
  • This invention streamlines the design of very complex systems by providing highly-predictable communications capabilities that allow higher degrees of design re-use, simpler performance analysis, flexible communications capabilities, and shorter design times.
  • the single chip computer bus system includes at least one initiator device and at least one target subsystem.
  • the bus cycles are divided up into recurring frames.
  • At least one initiator device is pre-allocated certain request frames to transmit a request and a certain reply frame a predetermined number of clock cycles after the request frame.
  • the initiator device issues the request during the pre-allocated frame.
  • the requests contain an address identifying the target subsystem and a command indicating the action to be taken by the target.
  • Each target subsystem coupled to the bus monitors the request packets issued and responds to those request packets containing the address that corresponds to the target subsystem address. The target responds by performing the command noted in the packet and transmitting a response packet the predetermined number of clock cycles after the request packet was issued.
  • cycles are pre-allocated to initiators with high bandwidth and low latency requirements.
  • the remaining initiators arbitrate for available frames, such as on a round-robin basis.
  • Figure 1 illustrates one embodiment of a single chip communication system that operates in accordance with the teachings of the present invention.
  • Figure 2a illustrates a simplified block diagram of one embodiment of an initiator interface module.
  • Figures 2b, 2c, 2d, 2e, 2f, 2g, 2h describe signals into and out of the interface module of Figure 2a.
  • Figure 3a is a simplified flow diagram of one embodiment of the process for setting up arbitration for access to the bus.
  • Figure 3b describes one embodiment of the process for issuing requests and replying to requests.
  • FIGS 4a and 4b illustrate one embodiment of the arbitration states for the second level of arbitration.
  • FIG. 5 is an illustrative timing diagram showing bus transactions in accordance with the teachings of the present invention.
  • Figure 6 illustrates a retry process in accordance with the teachings of the present invention.
  • the present invention provides an innovative communications protocol and bus structure on a single semiconductor device to support different subsystems having different timing constraints.
  • the system and method of the present invention provides an efficient mechanism for providing a high bandwidth and low latency communications path between two or more subsystems as well as provides an efficient mechanism for providing a minimum guaranteed bandwidth to any subsystem that needs real-time performance.
  • the system and method of the present invention provides an efficient mechanism for decoupling the requirements of the subsystems from the system level requirements. This enables coupled subsystems to be re-used in systems with widely varying system latency and bandwidth requirements without modification of the subsystem.
  • the system and method of the present invention provides an efficient mechanism to implement address/data-style memory mapped communication, dedicated point-to-point communication, and one-to-many broadcast-style communication.
  • the proposed communications method governs the communication between different subsystems that are part of a larger system and that are communicating to each other through a shared communication bus.
  • the system can consist of one or more integrated circuits or chips, and a subsystem can be a logic block on an integrated circuit containing many logic blocks or an integrated circuit implementing a single logic function.
  • FIG. 1 shows one embodiment of a system that operates in accordance with the teachings of the present invention.
  • the system is coupled to three target subsystems 10, 15 and 20 and two initiator subsystems 25, 30 communicating to each other over the communications bus 35.
  • a target subsystem 10, 15, 20 is defined herein as a subsystem that can only receive and respond to requests.
  • Typical target subsystems are memory subsystems and input/output (I O) devices.
  • An initiator subsystem 25, 30 is a subsystem that can receive and respond to requests, and can also issue requests.
  • Typical initiator subsystems are central processing unit (CPU) cores, digital signal processor (DSP) cores direct memory access (DMA) engines, co-processors, etc.
  • CPU central processing unit
  • DSP digital signal processor
  • DMA direct memory access
  • an interface unit 40, 45, 50, 55, 60 is connected between each subsystem 10, 15, 20, 25, 30 and the bus 35.
  • These interface modules 40, 45, 50, 55, 60 contain logic which issues and/or replies to packets in accordance with the bus protocol.
  • the interface synchronizes the subsystem and bus clocks for efficient data transfer.
  • there is no need for a dedicated controller unit to manage the transactions over the communications bus as it is preferred that the functionality is distributed among the interface units 40, 45, 50, 55, 60.
  • the interface units can be merged into the logic of the subsystem or can be implemented as a separate block (40, 45, 50, 55, 60) coupled to bus.
  • the communications bus 35 in Figure 1 consists of the following wires: command/address lines 65, assert lines 70, clock lines 75, data/response lines 80 and flag lines 85.
  • the system also includes token ring lines, e.g., lines 90, 95 for implementing a second level of arbitration as will be described below.
  • the address/command lines 65 are used to transmit the address that identifies the target subsystem.
  • the address lines further identify a resource within the target.
  • the specific allocation of lines between the address of a specific target subsystem and a resource within a subsystem can be varied and different for each target. Furthermore, it is preferred that the allocation is programmable.
  • the assert line 70 is preferably a single line used by an initiator subsystem to indicate that it is going to use the packet preallocated to it to transmit a request. If the assert line is not active, then the initiator holding the token will gain access to that packet.
  • the clock line 75 carries the clock from which the subsystems synchronize operation across the bus.
  • the data/response lines 80 are used to transmit data and convey addressed target responses to requests.
  • the flag lines 85 are general purpose lines available for specialized communication, thus adding further to the flexibility and robustness of the bus system. This particularly is advantageous to support communication that does not map easily into address/data transfers.
  • these general purpose lines can be used to support handshaking signals between two subsystems or a communication originated by a subsystem that is not an initiator, such as an interrupt signal driven by a target subsystem to signal an initiator subsystem that data is available.
  • FIGS 2a and 2b illustrate a simplified block diagram of the connections to one embodiment of an initiator interface module.
  • the module contains logic and memory necessary to perform the functions described herein, including issuing requests (e.g., logic 250), interfacing to the client (clock/synchronizer 245), maintaining a configuration (configuration registers 240) and updating the configuration in accordance with commands received (e.g., logic 250).
  • requests e.g., logic 250
  • configuration registers 240 configuration registers 240
  • the configuration registers 240 e.g., logic 250
  • the initiator In order to receive commands to update the configuration of operation of the module, the initiator also functions as a target and therefore includes address/decode logic 242.
  • initiators are described as having embedded targets as their primary function is that of an initiator; however it maintains the functionality of the target as described herein.
  • Cycle SO is referred to herein as the arbitration cycle.
  • Cycle S 1 is referred to herein as the command/address cycle.
  • Cycle SM which occurs a predetermined number of cycles after SI, is the data/response cycle.
  • Cycle SN is the error cycle.
  • the clock signal line sbclock 205 receives the bus clock. If the client operates synchronously to the bus or at some derived frequency, the clock circuitry 242 performs any clock division necessary to drive the interface to the client.
  • the signals sbTokenln (round robin token grant in) and sbTokenOut (round robin token grant out) 207, 209 are used in the second level of arbitration which is implemented in a token ring to indicate an incoming token and outgoing token, respectively.
  • sbAssert 211 is asserted by owners of preallocated packet to state the intention to use the packet.
  • the signal sbCmd 213 identifies the type of transaction as illustrated in Figure 2c.
  • the signal sbAddr 215 conveys the address of the target. All commands except the idle and broadcast command use this address mechanism.
  • the sb Width 217 signal indicates the data width, sbData 219, conveys data, sbResp 221 conveys response info such as that depicted in Figure 2d.
  • the sbFlagNum 223, as will be described later, identifies particular flags of sbFlag 225 for the receiver of the signal to monitor.
  • the signals sbFlag 225 are configurable flag lines that can be used a variety of ways depending upon implementation.
  • the interface module which interfaces with a target client further contains target subsystem signal lines 230 to interface with a target client.
  • the signals are discussed in Figure 2e. These include the transaction type, exemplary types illustrated in Figure 2f.
  • the interface module which interfaces with an initiator client further contains initiator subsystem signal lines 235 as illustrated in Figure 2g, including initiator client transaction types, such as those illustrated in Figure 2h.
  • Communications are achieved by combining the following: framed operation, time-division multiplex access, synchronous operation, pipelined operation, fixed-latency operation, and split-transactions.
  • TDMA time division multiplex access
  • bus cycles are divided into frames.
  • Device transfers between predetermined source and destination devices are pre-assigned frames.
  • a synchronous system is one in which devices perform transfers based upon the same clock.
  • a pipelined bus is a bus in which the address and command fields associated with a transaction are transmitted over the bus on a clock cycle before the data and response are transmitted allowing overlapping of the data/response of a first transaction with the address/command portion of a later transaction.
  • a new transaction can be initiated on every cycle of the clock.
  • additional transactions can be initiated before completion of a prior transaction.
  • certain performance benefits can be achieved by fixing the latency between initiation of a transaction and completion of a transaction.
  • the latency between a command and an associated data transfer is a known guaranteed value. This characteristic is extremely important for a bus that communicates with a processor or digital signal processing cores, which themselves operate in a fixed latency fully pipelined manner.
  • the frames are typically allocated for prespecified point-to-point communications, such as point- to-point voice communications. Any changes to the point-to-point communications happens slowly and typically due to the changes in either end of the point-to-point communications link.
  • the present invention overcomes these shortfalls by allocating only the right to initiate a transfer/operation command. This enables the initiator subsystem to easily communicate with a multiplicity of target subsystems on a real-time basis.
  • the initiator subsystem issues requests that include address information identifying the address of the target subsystem; thus, the initiator can communicate during a frame with one or more target subsystems based upon the command and address used.
  • the communication system of the present invention provides a very high performance, low latency communication path between subsystems over a shared interconnect.
  • Differences between the bus frequency and the subsystem frequency can present significant challenges in the subsystem design, depending on the patterns of communication required by each subsystem.
  • a single subsystem will likely be unable to keep up with a faster communications bus over long sequences of traffic to or from the subsystem; the problem is that the subsystem should not need to deal with data more rapidly than its own peak bandwidth requirements.
  • a burst is a sequence of data transfers occurring on consecutive bus cycles, typically involving implied address incrementation.
  • Traditional computer busses emphasize such burst traffic as a way to improve performance by reducing the time inefficiencies resulting from bus arbitration and address transfers.
  • the system cost implications of emphasizing burst transfers over a communications bus that operates at a higher frequency than the sending and receiving subsystems are that of expensive storage locations at the sender and receiver.
  • the sender cannot initiate its transfer until it has buffered up a burst length's worth of data, and the receiver must provide a buffer into which to transfer the data.
  • the very process of waiting for the last data to be created at the sender before transmitting the first data adds latency to the overall transfer.
  • a better solution is one that allows data to be transferred at the natural operating rate of the sender or receiver.
  • a time- interleaved set of individual transfers would replace a series of sequential burst transfers. This approach reduces storage and improves efficiency at both ends of the transfer.
  • an appropriately configured time-interleaving scheme can deliver seamless integration of traditional computing traffic with demanding multimedia and communications traffic characterized by hard real-time deadlines. Delivering all of these features requires enhancements to the fundamental arbitration scheme of the communication bus.
  • a two level arbitration is used. This generally is described with reference to Figures 3a and 3b.
  • the bandwidth allocation is performed.
  • the allocation can be performed at initialization of the system.
  • reallocation can be performed dynamically during operation by updating the configuration registers of the subsystems using available write bus transactions.
  • the packets are preallocated to the different initiator subsystems.
  • varying numbers of packets may be allocated to different initiator subsystems. For example, one initiator subsystem may have high bandwidth requirements and is preallocated a large percentage of the available packets.
  • Another initiator subsystem may be preallocated a small percentage of available packets, while still another initiator subsystem may not be preallocated any packets. Furthermore, it may be decided to not preallocate some packets such that there are always packets available for round robin arbitration.
  • the token ring network is established to implement a second level of arbitration, for example a round robin arbitration.
  • the members to the token ring may be initiator subsystems that were not preallocated packets; however, initiator subsystems that have been preallocated packets may also be members of the token ring network.
  • the bus system can begin operation. Referring to Figure 3b, at step 306, it is determined whether the initiator preallocated a particular upcoming packet is going to use the packet. For example, this can be determined by monitoring a particular bus line, such as the assert line of the present embodiment. If asserted, the initiator preallocated the packet has access to perform a transaction, step 308 and a request is issued, step 310. If the initiator does not require access, then the packet is available for a second level of arbitration, which is performed via the token ring network.
  • the second level of arbitration (step 312, 314) is performed concurrently with the first level of arbitration (i.e., steps 306, 308) such that the token bearing initiator issues a request immediately after the preallocated initiator declines access.
  • the arbitration processes may also be implemented in a sequential manner.
  • step 312 if the current holder of the token requires the packet, the token bearing initiator maintains the token and is granted access, at step 314, if the preallocated initiator declines access. Otherwise, the token is passed to the next initiator subsystem in the ring to permit the next initiator to gain access. This process continues until the token is located at an initiator that requires access.
  • the receiver of the token is again the initial holder of the token indicating therefore the token was passed around the ring without an initiator subsystem requiring access, the initial holder will maintain the token, and if granted access at step 314, will issue a dummy request to enable the arbitration to be completed.
  • the request issued by an initiator subsystem contains an address of the intended target.
  • the request further contains an address of a particular resource in the target.
  • the target subsystems monitor each request issued to determine whether the address corresponds to the address of the target, step 322, and if a correspondence (e.g., a match) is determined, the target subsystem responds to the request a predetermined number of clock cycles later, step 324. For example, if the request is a write operation, the target subsystem receives the data and issues a completion response the predetermined number of clock cycles after the issuance of the request. If the request is a read operation, the target subsystem provides the read data the predetermined number of clock cycles after the issuance of the request. In one embodiment, a 1 : 1 match between the issued address and target address is contemplated. However, various levels of correspondence, such as comparisons performed in conjunction with masks to match one or more devices are also contemplated.
  • the first level of arbitration is implemented in the following way.
  • the total available number of bus cycles are divided up into recurring frames.
  • the frame is divided up in a number of packets, which corresponds to a number of cycles.
  • a 256 cycle frame is divided up in 128 2-cycle packets.
  • a frame could be composed of 128 2-cycle packets, and in such a design the frame would repeat every 256 bus cycles.
  • the first level arbitration scheme revolves around the allocation of these packets to the different subsystems. Since only initiator subsystems can issue commands, the packets should only be allocated to the initiator subsystems.
  • every initiator subsystem contains a synchronous counter, which runs in lock step with all other counters.
  • the output of this counter is used to index a programmable logic circuit whose output determines if the packet associated with the output of the current counter value is allocated to this particular interface module.
  • the system implements a fully distributed TDMA time wheel, preferably as long as no two initiator interface modules are allocated the same packet.
  • the programmable logic circuits in the interface modules are pre-programmed at system initialization with information regarding allocation and/or may be re- programmed during run-time of the system by commands issued over the communications bus to the different initiator subsystems. This allows for dynamic re-allocation of the bandwidth in the system.
  • the allocation of the packet occurs preferably during a bus cycle before the occurrence of the packet; for example, the bus cycle immediately prior to the occurrence of the packet.
  • the most versatile implementation of the programmable logic circuit is as a lookup table in the form of a small random access memory (RAM).
  • the depth of such a RAM is equal to the number of packets in the frame.
  • the TDMA time wheel is implemented simply by indexing into the RAM with the frame counter.
  • the RAM where the tables are stored are user-visible and accessible for reading and writing over the communication bus by software using the same read/write commands as used in standard data transactions. This allows the bandwidth allocation of the system to be changed at transaction speeds comparable to normal read/write transactions over the bus, and to change the allocation to unevenly distribute bandwidth as needed to support certain applications.
  • subsystems may not use one or more of the preallocated cycles. This may occur as although some system traffic is predictable, even the predictable traffic may not be uniform. For instance, frame buffer traffic transferred to a display is very predictable, but the traffic flow is interrupted while the display performs horizontal and vertical retrace operations. Furthermore, some initiator subsystems may have very low performance requirements, and pre-allocating any bandwidth to such a subsystem may lead to a lot of unusable cycles. Therefore, it is desirable to enhance the time-division multiplexed access protocol with the second level of arbitration.
  • packets that are not pre-allocated or packets that were pre-allocated but not used are contended for by the other initiator subsystems on the bus in a fair round-robin scheme using a token passing mechanism.
  • This second level of arbitration ensures that no packet remains unused if any subsystem in the system can use the bus.
  • the mechanism of the second level of arbitration retains all the advantages of the previously described communications method, but increases the overall efficiency of use of the available system bandwidth.
  • the second level of arbitration is implemented as a fair round-robin scheme using a token passing mechanism. All initiators that participate in the second level of arbitration are connected in a ring to pass the token around. Typically, those initiators that can tolerate a unpredictable latency participate in the second level of arbitration.
  • subsystems that participate in the first level of arbitration can also participate in the second level of arbitration.
  • a token is allocated to one initiator subsystem at initialization. The token signifies a conditional right to issue bus commands during the bus cycles associated with the current packet. This right is conditional because it is dependent on the pre-allocated owner of the packet not asserting its right to use the packet or on the packet being unallocated.
  • the TDMA scheme actually allocates a right of first refusal: the pre-allocated owner must assert its right to use its packet, or else the packet is made available to the initiator that ends up with the token.
  • the token On every arbitration cycle (i.e. every packet), the token may remain at the initiator that held it on the previous cycle, or it may pass around the ring to the next initiator that is requesting the token, unlike existing token ring systems, in one embodiment of the present invention the token may be passed completely around the ring back to the initiator that started it.
  • the token's action depends upon system conditions and fairness constraints as set forth in the system's rules. For example, it is unfair for one initiator to use the token for more than one packet, since other initiator subsystems may be waiting. It is equally unfair for an initiator subsystem to be forced to give up the token before having a chance to use it.
  • the token-passing rules are structured such that an initiator subsystem that requests and receives the token may keep it until getting an opportunity to initiate precisely one packet's worth of commands across the communication bus. That opportunity may occur in the arbitration cycle in which the token arrives, or it may not occur for a large number of cycles, depending on the usage decisions made by the devices to which the packets are pre-allocated to owners. Finally, once the initiator with the token has a chance to use it, the initiator starts the token around the ring by driving its TokenOut signal to logic one. If no other initiator requests the token, it will pass completely around the ring and return to its sender.
  • the initiator subsystem that starts the token In order to prevent the token from circulating around the ring indefinitely, it is preferred that the initiator subsystem that starts the token must temporarily break the ring. Thus, the initiator subsystem that starts the token on any given arbitration cycle can end up with it again by the end of that arbitration cycle. It should be noted that it is desirable that the communication bus signals always be actively driven so that the wires do not float to an undetermined state where improper bus operation and increased power dissipation may occur. As such, the initiator that ends up with the token should initiate a bus command whenever no pre-allocated packet owner asserts its right of first refusal.
  • the initiator with the token should initiate a useless command (a harmless read or write, or a NOP/Idle command) if it has no useful commands to issue. If the initiator that owns the token at the beginning of the arbitration cycle has not had an opportunity to use the token and has commands to issue, it can keep the token. If not, the initiator pulls a signal wire high (e.g., Token Out) indicating to other initiator subsystems that the token is available for the contention through the second level of arbitration. The initiator subsystem that has the token (i.e. the initiator subsystem that used the token last) passes the token on to the next initiator subsystem in the ring, which then, in turn, can appropriate the token if it can use the cycle.
  • a signal wire high e.g., Token Out
  • the token is passed around until the token is allocated, in which case that specific initiator subsystem keeps the token as a starting point for the next token ring arbitration. If no initiator subsystem uses the token, it comes back to the initiator that initially had the token. All packet arbitration, both the first level and second level, occurs in the same cycle. This second level of arbitration does not provide guaranteed bandwidth, but does provide for guaranteed access. Using the second level of arbitration, no initiator system can command the bus for more than 1 packet if another initiator also wants access to the bus.
  • Figure 4a illustrates the states of one embodiment of the second level of arbitration and the table of Figure 4b sets forth a brief explanation of the states.
  • the propagation delay around the ring is minimized by two implementation aspects.
  • the interface module generating the token does so purely as a function of the current state.
  • the interface modules along the way decide to pass the token or not purely as a function of the current state.
  • the input to the states are ReqToken, Tokenin, Assert, Reset_n, First and the outputs are TokenOut, Grant Token and the next state.
  • ReqToken indicates that the interface is requesting the token
  • Tokenin indicates that the token is available to this initiator
  • Assert is used to indicate that the preallocated initiator will use its packet
  • reset_n is used to force the system into a reset state
  • TokenOut is used to pass the token to the next initiator in the ring
  • GrantToken is used to tell the initiator that it should prepare to issue a request
  • Next State indicates the next state of the particular interface module of the initiator subsystem.
  • state NOTOKEN the interface module of the subsystem neither desires nor has the token. Thus, if the token arrives on Tokenin it is immediately passed on via TokenOut. The state machine leaves NOTOKEN once it samples the interface module requesting the token. In state WANT TOKEN, the interface module does not have, but desires, the token. It therefore breaks the ring, such that when the token arrives on Tokenin this module will grab it. Once the token arrives, the interface module will use it (by noticing GrantToken) if no preallocated slot owner drove Assert. If Assert is active, then the module will keep the token until it gets an opportunity to us it, so the state machine transitions to HAVETOKEN.
  • state HAVETOKEN the module has the token and is waiting for an opportunity to use it. As soon as no module raises Assert, the slot is available and this module will use it (by noticing GrantToken). When either this occurs or the interface module stops asserting ReqToken (presumably because its request was satisfied via a pre-allocated slot), the state machine transitions to GENTOKEN. Since this module has the token, there is no need to pay attention to the Tokenin line (e.g., it must be zero). In state GENTOKEN, the module will start the token around the ring at the beginning of the arbitration cycle. The module must break the ring in case no module is in WANTTOKEN (i.e., is actively requesting the token).
  • the interface module will receive GrantToken and be required to initiate a command, independently of whether the device requested the token. While the state machine is in reset (Reset_n is driven low), the starting state is set to NOTOKEN for all interface modules except the one labeled First.
  • the First device is unique in that it starts off with the token at reset, thus ensuring that there is precisely one token in the ring at any time. Furthermore, it should be noted that in the present embodiment, the Don't Care entries in the ReqToken column (other than during reset) guarantee that precisely one device always initiates a command.
  • the interface module that ends up with the token at the end of arbitration must initiate a command if no pre-allocated slot owner drives the Assert line high.
  • FIG. 5 shows a timing diagram of exemplary transactions across such a fully pipelined, fixed- latency bus that operates in accordance with the teachings of the present invention. Every transaction consists of an arbitration part, a command/address part, and a data/response part. Since the bus is a fixed-latency bus, there is a fixed delay (in number of clock cycles) between the two transactions, i.e., request and response.
  • the latency between arbitration and command/address is one cycle
  • the latency between command/address and data is assumed to be 2 two cycles
  • a packet (the arbitration unit) is one bus cycle.
  • initiator subsystem I-D asserts its right to use its pre-allocated packet, so in cycle time slot 2, I-D issues a command and address on the bus to a target.
  • the command is labeled Command 1.
  • All target interface modules contain address decoding logic to decide if a transaction is addressed to their client subsystem.
  • the addressed target interface module responds to the request command.
  • the read data is sent from the target to the initiator.
  • the data to be written is sent from the initiator to the target in cycle 4.
  • a response signal that is driven by the target in cycle 4 from the target over the data response lines indicates if the transaction was successful or not.
  • the "Valid” response indicates that the transaction was successful.
  • the "Busy” response indicates that the target module could respond to the command.
  • the "Retry” response is only allowed for read commands, and indicates that the target is attempting to satisfy the command but could not do so within the fixed bus latency. The initiator must retry its request later. In case of a "No Response", the address provided with the command did not match any of the targets, and therefore no target responded.
  • a second arbitration cycle begins in clock cycle 2 and another transaction is issued in cycle 3.
  • initiator subsystem I-E which ended up with the token in cycle 2 issues a command and address in cycle 3.
  • the data and response signals associated with this the command are issued transfer in cycle 5, a bus cycle latency later.
  • the interface modules preferably are programmable to provide for dynamic reallocation of the packets.
  • the interface modules for each subsystem contain a set of configuration registers (240, Figure 2) to store the configuration data of the modules and the bus.
  • the configuration registers store data identifying packets preallocated to the corresponding subsystem.
  • the configuration registers have addresses in the system address space, and therefore can be read from and written to using available bus transactions.
  • Data written to a buffered register is stored in a holding register location and only becomes visible after that data has been transferred from the holding register to the configuration register.
  • the transfer is initiated through a broadcast command over the bus.
  • a broadcast command is a command that affects all interface modules simultaneously.
  • buffered registers typically are used when it is desirable to update several interface modules simultaneously before the new data can be allowed to be seen by the rest of the system.
  • buffered registers One example of the use of buffered registers is the assignment of the address map in the system.
  • Each target interface or module has a locally stored mask and match data value that specifies to which address(es) the target will respond.
  • the address on the bus is compared to the match field and the result is masked. If all bits in the masked result are 0, this specific interface module is addressed.
  • the mask/match data is stored in a configuration register and can be written to by regular bus commands. Proper operation of the system requires that target address ranges must be uniquely assigned in the system. If address changes are to be implemented, for example, the swapping of addresses, the new assignments should take place simultaneously in all affected interface modules. Otherwise, overlapping address-spaces may exist during the transition from one address map to the other.
  • this invention is not limited to the use of buffered and unbuffered configuration registers.
  • all or some of the data can be stored in read only memory (ROM).
  • ROM typically requires less die area, and prevents undesired writes to configuration locations that resulted in improper system configuration.
  • a typical system contains interface modules having a combination of buffered configuration registers, unbuffered configuration registers, and ROM.
  • Information included in the configuration registers consists of data regarding packets preallocated to initiator subsystem, latency, frequency of communication between interface module and the coupled address device of the target for purposes of request address matching.
  • configuration data can be written during real-time operation of the system using available normal read and write commands as the registers are part of the system address space. This allows the system designer to change many significant features of the communication system via simple software or even pre-computed command sequences stored in a ROM.
  • the bus latency is programmable at initialization of the system or during the run-time of the system.
  • the latency is programmable over the communications bus. Because of the fully pipelined nature of the communication bus, such an operation requires a broadcast command so the pipeline depth is changed simultaneously in all interface modules.
  • a low latency can be programmed into the communications system by writing the new latency into the appropriate configuration registers of the interface modules.
  • the system latency can be set higher. The fact that the latency can be programmed at speeds comparable to regular bus transactions is a significant improvement over the prior art.
  • bandwidth allocation can be changed using the same method of updating the appropriate configuration registers.
  • the fact that the bandwidth allocation can be programmed at speeds comparable to a normal read or write transaction over the communications bus, either at initialization or during the run-time of the system is a significant improvement over prior art.
  • the combination of programmable bandwidth allocation and programmable bus latency provides distinct advantages; most particularly, the system designer can optimize the performance of the system for the particular configuration of initiators and targets. Furthermore, the optimization can be maintained through a variety of conditions and configurations as the bandwidth allocation and latency can be updated during operation. Furthermore, as will be described subsequently, additional flexibility and configurability is contemplated by enabling an initiator device to program the frequency at which a target and its interface module communicate. This is particularly desirable to adapt to varying subsystem timing requirements of the initiator and/or target client.
  • the bus functionability further is enhanced by adding a set of system-level control wires (referred to in Figure 1 as flags) that manage out-of- band communication.
  • Out-of-band signals are signals that do not follow the address/data-oriented pipelined bus model. Examples include device status signals, interrupt, and handshake signals.
  • these wires are also used for enhancing the performance of the basic read operations in the following manner.
  • the target interface module uses one of the out-of-band signals (referred to herein also as flags). The target interface module uses the flags to provide a precise indication back to the initiator once the target interface module has obtained the data.
  • Figure 6 illustrates the retry mechanism using the out of band signal lines.
  • the target interface module When a client cannot meet the latency of the communications bus, the target interface module issues a response packet to the initiator to indicate a "Retry". In the same time slot the bit number of a Flag wire signal is transmitted on the FlagNum lines (which are, for purposes of simplification, part of the Flag lines of Figure 1). This indicates to the initiator subsystem that it should monitor that identified flag for an indication of when to retry.
  • a Read request command is transmitted in cycle 1.
  • the bus latency is two cycles, so the data should be returned by the target interface module in cycle 3.
  • the target returns a retry response on the response wires and it returns the number of a flag wire on the FlagNum bus, in this example number 7.
  • the target interface module receives the data from its client (i.e., coupled device), it stores the data in local memory and it drives the indicated flag wire high, indicating to the initiator that the data is now available.
  • the target indicates on Flag 7 that the data is available.
  • the initiator subsystem can now re-issue the request command which in Figure 3 occurs on cycle 6.
  • the target interface module satisfies the request command (cycle 8 in Figure 3), it deasserts the Flag that has been associated with this request command.
  • the communications system preferably contains a number of flags that are allocated to the different target interface modules. Flag wires can be shared between modules or can be dedicated to a specific module. In the preferred embodiment, the allocation of the flags is stored in the configuration registers and therefore this allocation is programmable over the communications bus and at a speed comparable to normal bus transactions.
  • the flag wires can be used for a variety of purposes. For example, a different use of the flag wires is the management of handshake signals and interrupt signals. These are signals that indicate a special event in the interface modules or the subsystem. Examples of such signals are error signals or signals that indicate busy status (stop sending requests) or ready status (subsystem is ready to take accept requests). These signals are inherently out of band signals and are transferred using the flag wires.
  • the flag wires is are uniquely allocated to target interface modules and their subsystem clients for the specific purpose of interrupt or handshake signaling. In the preferred embodiment, this data allocation is stored in a configuration registers so it may be dynamically varied as system needs change. The same information is programmed in the initiator subsystem to let the initiator know the flag wires to monitor for handshake or interrupt purpose.
  • the interface modules between the bus and the subsystems contain logic to implement the bus protocols while interfacing with the client. Furthermore, it is preferred that the interface contains additional logic (245, Figure 2) to derive a client clock of the subsystem from the system bus clock.
  • the interface module through the configuration registers provides a level of programmability to enable subsystems reuse in many different systems with very different system features and performance requirements.
  • the frequency of the system bus clock is determined by system requirements, most notably the total bandwidth required on the bus to satisfy the required performance of the system.
  • the client clock frequency of a client can be independent of the system bus clock, and should be related to the functionality required of the client.
  • the interface modules therefore provide a powerful means of decoupling the system and subsystem requirements.
  • the bus latency (measured in cycles) can be increased as the bus frequency is increased, which tends to allow clients to operate a approximately consistent frequencies, since the total latency (in time) stays relatively constant. It is contemplated that the decoupling feature be extended to asynchronous connections between the interface module and the client.
  • a wireless personal communicator which requires fixed-performance real-time subsystems (to implement the wireless communications) along with low- performance, computer-derived subsystems such as a microcontroller core, embedded memory, keypad input and liquid crystal display (LCD) output.
  • Key design goals for the communicator are to provide reliable communications and as many features as possible (implemented in software on the microcontroller) at low power.
  • a second exemplary system is a set-top box for receiving digital satellite television broadcasts. The set-top box design requires much higher bit- rate data communications and higher computational performance to decompress the video and audio data, and must guarantee performance across the entire system to avoid dropping frames.
  • a third example is an Asynchronous Transfer Mode (ATM) switch.
  • the switch design likely needs a central processing unit (CPU) only to monitor performance and provide diagnostics; the normal operation would involve switching packets between identical subsystems across the shared interconnect. As in most circuit switching applications, performance guarantees are critical to proper operation of the ATM switch.
  • CPU central processing unit
  • a wireless system for video-on-demand could combine the two- way wireless subsystems from the personal communicator, the ATM packet framing engine from the ATM switch, and the video decompression subsystems from the set-top box. Combination of subsystems that were never intended to interoperate into rapidly integrated, cost-effective systems can be achieved.

Abstract

The present invention provides for an on-chip communications method with fully distributed control combining a fully-pipelined, fixed latency, synchronous bus (35) with a two-level arbitration scheme where the first level of arbitration is a framed, time division multiplexing arbitration scheme and the second level is a fairly-allocated round-robin scheme implemented using a token passing mechanism. Both the latency and the bandwidth allocation are software programmable in real-time operation of the system. The present invention also provides for a communication system where access to a shared resource (10, 15, 20) is controlled by the communications protocol. Access to and from the shared resource (10, 15, 20) from the subsystem is through a bus interface module (40, 45, 50, 55, 60). The bus interface modules (40, 45, 50, 55, 60) provide a level of indirection between the subsystem (25, 30) to be connected to the shared resource (10, 15, 20). This allows the decoupling of system performance requirements. Communication over the bus (35) is fully memory mapped.

Description

A FULLY-PIPELINED FIXED-LATENCY COMMUNICATIONS SYSTEM WITH A REALTIME DYNAMIC BANDWIDTH ALLOCATION
BACKGROUND OF THE INVENTION
1. FIELD OF THE INVENTION
The present invention relates to a bus system to couple computing devices.
2. ART BACKGROUND
As electronic computing and communications systems continue to increase in features and complexity, and to shrink in physical size and cost per function, more and more constraints are placed on the system designer. The demand for such systems is focused in high-volume, cost-sensitive markets such as consumer electronics, where rapid time-to-market is a business necessity. The only feasible approach to deliver more complex systems in less time and at lower prices is to make effective use of advanced integrated circuit technology, push functionality into software, and migrate subsystem components from one design into subsequent designs.
Re-use of subsystem components in subsequent designs have become quite common. Many other benefits are realized in addition to the time savings achieved. First, a model may be written for the proven subsystem that can provide accurate results when analyzing the requirements and performance of a new system design; the model for a new, unproven subsystem is likely to be neither as accurate as the proven subsystem, nor built in time to influence the design. Second, proven subsystems can serve as building blocks that simplify the overall design process by allowing the system designer to focus at a higher level of abstraction, while providing improved predictability in the resulting system implementation. Third, re-use of hardware subsystems protects the investment in software to control those subsystems, and allows the system software implementation to proceed as soon as the hardware building blocks have been chosen. Finally, subsystem re-use protects the investment in verification and testing. Since the desired systems are highly integrated, the required subsystems end up deeply embedded within an integrated circuit. In deeply-embedded designs, verifying the design functionality becomes very challenging and testing an individual system to prove that it is correctly built can lead to expensive delays or costly system rework. Thus, maintenance of the integrity of subsystem verification and test is likely the single biggest gain from design re-use.
Traditional approaches to design re-use have various strengths and weaknesses. An essential aspect of such approaches is the communications interface the various subsystems present to one another. One approach is to define customized point-to-point interfaces between a subsystem and each peer to which it must communicate. This interface style is particularly popular in applications where data flows from a first subsystem into the subsystem being designed, which processes the data and outputs the data to a second subsystem. This customized approach offers protocol simplicity, guaranteed performance, and isolation from dependencies on unrelated subsystems. However, customized interfaces by their very nature are inflexible. If a new application needs an existing subsystem to be interfaced to one that does not share its interface, design re-work is required. If the application requires the subsystem to communicate with different subsystems, multiple copies of subsystems or interfaces may be required, leading to system inefficiencies. Furthermore, if an updated algorithm is to be implemented in the subsystem, the new algorithm will need to be wedged into the old customized interfaces to allow communication with existing peer subsystems. A second approach is to define a system using standardized interfaces. It is common to see standardized interfaces paired with various forms of shared interconnect to form a communications subsystem. Many standardized interfaces are based on pre-established computer bus protocols. One significant advantage to this approach is flexibility in communications patterns: the ability for every agent to communicate with every other agent in the system. Computer buses allow flexibility in system design, since as many different agents may be connected together as required by the system, as long as the bus has sufficient performance. A final advantage is the natural mapping of address/data transfers, which are fundamental to processor-memory communications, into systems where complex algorithms that are implemented in software need to communicate with hardware subsystems. Along with shared interconnect comes a requirement to allocate the communications resources among the various initiator devices and target subsystems. In the case of computer buses, resource allocation is typically referred to as arbitration.
A principal disadvantage of standardized bus interfaces is the uncertainty in transfer delay (normally called latency) that results from arbitration. Latency uncertainty causes trouble for subsystems that must satisfy real-time constraints, since data that does not arrive in time to meet a deadline can result in improper system behavior. A second disadvantage is a lack of total available transfer opportunities (normally called bandwidth) that results from the computer system heritage. Historically, as well as presently, a bus typically is designed to support the peak transfer rate of the processor connected to it, and transfer opportunities not claimed by the processor are made available to other agents. Systems with combined bandwidths in excess of the peak processor bandwidth therefore must resort to multilevel bus schemes, or a mix of buses and dedicated connections, to separate the bandwidth. A final disadvantage is the inability to effectively support communication that does not map easily into address/data transfers (for instance, handshaking signals between two agents) or communications originated by an agent that is not an initiator (for instance, an interrupt signal driven by a target subsystem to signal an initiator device that data is available).
In summary, existing communication approaches do not meet the requirements for effective design re-use. What is needed is a new structure that allows maximum subsystem re-use in systems that span a wide range of performance characteristics. If a computer bus could be extended to remove its performance and communications style limitations, it might well serve as the basis for many highly integrated systems.
Furthermore, subsystem communications requirements vary greatly. Some subsystems, such as input/output devices like keyboards, infrared remote controllers, and LED displays, have very low bandwidth requirements and are very tolerant of variable latency. Other subsystems, such as a RISC CPU or a digital signal processor, desire high available bandwidth and low latency to memory, but can tolerate bandwidth limits and variable latency at the expense of slower program execution. Still other subsystems, such as a Tl telephony link or an SVGA display, have moderate to high bandwidth requirements, but cannot tolerate uncertainties in bandwidth nor latency. Such real-time subsystems need guaranteed bandwidth and latency, and there is no advantage to providing any extra performance.
Therefore, it is desirable that such a communication mechanism would allow subsystems of widely varying performance characteristics to interoperate, and greatly improve the re-use of the subsystems by providing a standardized interface. However, this does not address the issue that the operating frequency of the communication mechanism must be variable to support the required bandwidth of the system under design. If a standardized interface is at the communication interface, changing frequencies of the communication interface requires changing the operating frequencies of the subsystems. Increasing the frequency of a subsystem beyond its requirements wastes power, can lead to algorithmic problems (some subsystems need to operate at fixed frequencies), and can prevent design re-use altogether when the new frequency is higher than the previously-designed subsystem can operate. What is needed is a method to de-couple the frequency of the communication interface from the operating frequencies of the various client subsystems, so that each may operate based on their own requirements.
SUMMARY OF THE INVENTION
This invention streamlines the design of very complex systems by providing highly-predictable communications capabilities that allow higher degrees of design re-use, simpler performance analysis, flexible communications capabilities, and shorter design times.
In one embodiment, the single chip computer bus system includes at least one initiator device and at least one target subsystem. The bus cycles are divided up into recurring frames. At least one initiator device is pre-allocated certain request frames to transmit a request and a certain reply frame a predetermined number of clock cycles after the request frame. When a request is to be issued, the initiator device issues the request during the pre-allocated frame. The requests contain an address identifying the target subsystem and a command indicating the action to be taken by the target. Each target subsystem coupled to the bus monitors the request packets issued and responds to those request packets containing the address that corresponds to the target subsystem address. The target responds by performing the command noted in the packet and transmitting a response packet the predetermined number of clock cycles after the request packet was issued.
Preferably, this is implemented in a system that unifies communication between subsystems having widely varying constraints. For example, in one embodiment, cycles are pre-allocated to initiators with high bandwidth and low latency requirements. The remaining initiators arbitrate for available frames, such as on a round-robin basis. BRIEF DESCRIPTION OF THE DRAWINGS
The objects, features and advantages of the present invention will be apparent from the following detailed description in which:
Figure 1 illustrates one embodiment of a single chip communication system that operates in accordance with the teachings of the present invention.
Figure 2a illustrates a simplified block diagram of one embodiment of an initiator interface module.
Figures 2b, 2c, 2d, 2e, 2f, 2g, 2h describe signals into and out of the interface module of Figure 2a.
Figure 3a is a simplified flow diagram of one embodiment of the process for setting up arbitration for access to the bus.
Figure 3b describes one embodiment of the process for issuing requests and replying to requests.
Figures 4a and 4b illustrate one embodiment of the arbitration states for the second level of arbitration.
Figure 5 is an illustrative timing diagram showing bus transactions in accordance with the teachings of the present invention.
Figure 6 illustrates a retry process in accordance with the teachings of the present invention.
DETAILED DESCRIPTION
The present invention provides an innovative communications protocol and bus structure on a single semiconductor device to support different subsystems having different timing constraints. In particular, as will be described below, the system and method of the present invention provides an efficient mechanism for providing a high bandwidth and low latency communications path between two or more subsystems as well as provides an efficient mechanism for providing a minimum guaranteed bandwidth to any subsystem that needs real-time performance. Furthermore, the system and method of the present invention provides an efficient mechanism for decoupling the requirements of the subsystems from the system level requirements. This enables coupled subsystems to be re-used in systems with widely varying system latency and bandwidth requirements without modification of the subsystem. In addition, the system and method of the present invention provides an efficient mechanism to implement address/data-style memory mapped communication, dedicated point-to-point communication, and one-to-many broadcast-style communication. The proposed communications method governs the communication between different subsystems that are part of a larger system and that are communicating to each other through a shared communication bus. The system can consist of one or more integrated circuits or chips, and a subsystem can be a logic block on an integrated circuit containing many logic blocks or an integrated circuit implementing a single logic function.
Figure 1 shows one embodiment of a system that operates in accordance with the teachings of the present invention. The system is coupled to three target subsystems 10, 15 and 20 and two initiator subsystems 25, 30 communicating to each other over the communications bus 35. A target subsystem 10, 15, 20 is defined herein as a subsystem that can only receive and respond to requests. Typical target subsystems are memory subsystems and input/output (I O) devices. An initiator subsystem 25, 30 is a subsystem that can receive and respond to requests, and can also issue requests. Typical initiator subsystems are central processing unit (CPU) cores, digital signal processor (DSP) cores direct memory access (DMA) engines, co-processors, etc. In one embodiment, an interface unit 40, 45, 50, 55, 60 is connected between each subsystem 10, 15, 20, 25, 30 and the bus 35. These interface modules 40, 45, 50, 55, 60 contain logic which issues and/or replies to packets in accordance with the bus protocol. In addition, as will be explained below, it is preferred that the interface synchronizes the subsystem and bus clocks for efficient data transfer. Finally, in the present invention, there is no need for a dedicated controller unit to manage the transactions over the communications bus as it is preferred that the functionality is distributed among the interface units 40, 45, 50, 55, 60. Furthermore, the interface units can be merged into the logic of the subsystem or can be implemented as a separate block (40, 45, 50, 55, 60) coupled to bus.
In one embodiment, the communications bus 35 in Figure 1 consists of the following wires: command/address lines 65, assert lines 70, clock lines 75, data/response lines 80 and flag lines 85. The system also includes token ring lines, e.g., lines 90, 95 for implementing a second level of arbitration as will be described below.
The address/command lines 65 are used to transmit the address that identifies the target subsystem. Preferably, the address lines further identify a resource within the target. Preferably, the specific allocation of lines between the address of a specific target subsystem and a resource within a subsystem can be varied and different for each target. Furthermore, it is preferred that the allocation is programmable.
The assert line 70 is preferably a single line used by an initiator subsystem to indicate that it is going to use the packet preallocated to it to transmit a request. If the assert line is not active, then the initiator holding the token will gain access to that packet. The clock line 75 carries the clock from which the subsystems synchronize operation across the bus. The data/response lines 80 are used to transmit data and convey addressed target responses to requests. The flag lines 85 are general purpose lines available for specialized communication, thus adding further to the flexibility and robustness of the bus system. This particularly is advantageous to support communication that does not map easily into address/data transfers. For instance, as will be explained subsequently, these general purpose lines can be used to support handshaking signals between two subsystems or a communication originated by a subsystem that is not an initiator, such as an interrupt signal driven by a target subsystem to signal an initiator subsystem that data is available.
Figures 2a and 2b illustrate a simplified block diagram of the connections to one embodiment of an initiator interface module. The module contains logic and memory necessary to perform the functions described herein, including issuing requests (e.g., logic 250), interfacing to the client (clock/synchronizer 245), maintaining a configuration (configuration registers 240) and updating the configuration in accordance with commands received (e.g., logic 250). In order to receive commands to update the configuration of operation of the module, the initiator also functions as a target and therefore includes address/decode logic 242. For purposes of simplification of discussion, initiators are described as having embedded targets as their primary function is that of an initiator; however it maintains the functionality of the target as described herein. However flexibility is maintained as it is further contemplated that a subsystem may at different times function as a target or an initiator. The table of Figure 2b summarizes the signals. In particular, "cycle" indicates the phase of the bus clock (sbclock) the signal is active. Since the bus is pipelined, multiple transactions can be simultaneously in progress. However, since it is preferred that all transaction have the same latency, each transaction precisely follows the same template. Cycle SO is referred to herein as the arbitration cycle. Cycle S 1 is referred to herein as the command/address cycle. Cycle SM, which occurs a predetermined number of cycles after SI, is the data/response cycle. Cycle SN is the error cycle.
Referring to Figure 2a, the clock signal line sbclock 205, receives the bus clock. If the client operates synchronously to the bus or at some derived frequency, the clock circuitry 242 performs any clock division necessary to drive the interface to the client. The signals sbTokenln (round robin token grant in) and sbTokenOut (round robin token grant out) 207, 209 are used in the second level of arbitration which is implemented in a token ring to indicate an incoming token and outgoing token, respectively. sbAssert 211 is asserted by owners of preallocated packet to state the intention to use the packet. The signal sbCmd 213 identifies the type of transaction as illustrated in Figure 2c. The signal sbAddr 215 conveys the address of the target. All commands except the idle and broadcast command use this address mechanism. The sb Width 217 signal indicates the data width, sbData 219, conveys data, sbResp 221 conveys response info such as that depicted in Figure 2d. The sbFlagNum 223, as will be described later, identifies particular flags of sbFlag 225 for the receiver of the signal to monitor. The signals sbFlag 225 are configurable flag lines that can be used a variety of ways depending upon implementation.
The interface module which interfaces with a target client further contains target subsystem signal lines 230 to interface with a target client. The signals are discussed in Figure 2e. These include the transaction type, exemplary types illustrated in Figure 2f. In addition, the interface module which interfaces with an initiator client further contains initiator subsystem signal lines 235 as illustrated in Figure 2g, including initiator client transaction types, such as those illustrated in Figure 2h.
Communications are achieved by combining the following: framed operation, time-division multiplex access, synchronous operation, pipelined operation, fixed-latency operation, and split-transactions. In a time division multiplex access (TDMA) system, bus cycles are divided into frames. Device transfers between predetermined source and destination devices are pre-assigned frames. A synchronous system is one in which devices perform transfers based upon the same clock. In the present embodiment, a pipelined bus is a bus in which the address and command fields associated with a transaction are transmitted over the bus on a clock cycle before the data and response are transmitted allowing overlapping of the data/response of a first transaction with the address/command portion of a later transaction. In a fully pipelined bus, a new transaction can be initiated on every cycle of the clock. Thus, additional transactions can be initiated before completion of a prior transaction. Furthermore, in the present invention it has been determined that certain performance benefits can be achieved by fixing the latency between initiation of a transaction and completion of a transaction. Thus, in a fully pipelined, fixed latency, split transaction bus, the latency between a command and an associated data transfer is a known guaranteed value. This characteristic is extremely important for a bus that communicates with a processor or digital signal processing cores, which themselves operate in a fixed latency fully pipelined manner.
In prior art communication systems that are TDMA, the frames are typically allocated for prespecified point-to-point communications, such as point- to-point voice communications. Any changes to the point-to-point communications happens slowly and typically due to the changes in either end of the point-to-point communications link. The present invention overcomes these shortfalls by allocating only the right to initiate a transfer/operation command. This enables the initiator subsystem to easily communicate with a multiplicity of target subsystems on a real-time basis. The initiator subsystem issues requests that include address information identifying the address of the target subsystem; thus, the initiator can communicate during a frame with one or more target subsystems based upon the command and address used. The communication system of the present invention provides a very high performance, low latency communication path between subsystems over a shared interconnect. In order to support the required system bandwidth, it may be necessary to operate the communication bus at a higher operating frequency than that of the individual subsystems. Differences between the bus frequency and the subsystem frequency can present significant challenges in the subsystem design, depending on the patterns of communication required by each subsystem. In particular, a single subsystem will likely be unable to keep up with a faster communications bus over long sequences of traffic to or from the subsystem; the problem is that the subsystem should not need to deal with data more rapidly than its own peak bandwidth requirements.
The system and method of the present invention further attempts to minimize usage of burst transfers. A burst is a sequence of data transfers occurring on consecutive bus cycles, typically involving implied address incrementation. Traditional computer busses emphasize such burst traffic as a way to improve performance by reducing the time inefficiencies resulting from bus arbitration and address transfers. The system cost implications of emphasizing burst transfers over a communications bus that operates at a higher frequency than the sending and receiving subsystems are that of expensive storage locations at the sender and receiver. The sender cannot initiate its transfer until it has buffered up a burst length's worth of data, and the receiver must provide a buffer into which to transfer the data. Furthermore, the very process of waiting for the last data to be created at the sender before transmitting the first data adds latency to the overall transfer.
It can therefore be seen that a better solution is one that allows data to be transferred at the natural operating rate of the sender or receiver. Thus, a time- interleaved set of individual transfers would replace a series of sequential burst transfers. This approach reduces storage and improves efficiency at both ends of the transfer. In addition, an appropriately configured time-interleaving scheme can deliver seamless integration of traditional computing traffic with demanding multimedia and communications traffic characterized by hard real-time deadlines. Delivering all of these features requires enhancements to the fundamental arbitration scheme of the communication bus.
In one embodiment of the system of the present invention, a two level arbitration is used. This generally is described with reference to Figures 3a and 3b. Referring to Figure 3a, the bandwidth allocation is performed. The allocation can be performed at initialization of the system. Furthermore, as noted herein, reallocation can be performed dynamically during operation by updating the configuration registers of the subsystems using available write bus transactions. To establish the bandwidth allocation, at step 302, the packets are preallocated to the different initiator subsystems. Depending upon the initiator subsystems coupled to the bus system and the needs of the initiator subsystems, varying numbers of packets may be allocated to different initiator subsystems. For example, one initiator subsystem may have high bandwidth requirements and is preallocated a large percentage of the available packets. Another initiator subsystem may be preallocated a small percentage of available packets, while still another initiator subsystem may not be preallocated any packets. Furthermore, it may be decided to not preallocate some packets such that there are always packets available for round robin arbitration.
At step 304, the token ring network is established to implement a second level of arbitration, for example a round robin arbitration. Typically the members to the token ring may be initiator subsystems that were not preallocated packets; however, initiator subsystems that have been preallocated packets may also be members of the token ring network.
Once the two levels of arbitration have been established, the bus system can begin operation. Referring to Figure 3b, at step 306, it is determined whether the initiator preallocated a particular upcoming packet is going to use the packet. For example, this can be determined by monitoring a particular bus line, such as the assert line of the present embodiment. If asserted, the initiator preallocated the packet has access to perform a transaction, step 308 and a request is issued, step 310. If the initiator does not require access, then the packet is available for a second level of arbitration, which is performed via the token ring network. Preferably, as shown in Figure 3b, the second level of arbitration (step 312, 314) is performed concurrently with the first level of arbitration (i.e., steps 306, 308) such that the token bearing initiator issues a request immediately after the preallocated initiator declines access. However, it is contemplated that the arbitration processes may also be implemented in a sequential manner.
Referring again to Figure 3b, at step 312, if the current holder of the token requires the packet, the token bearing initiator maintains the token and is granted access, at step 314, if the preallocated initiator declines access. Otherwise, the token is passed to the next initiator subsystem in the ring to permit the next initiator to gain access. This process continues until the token is located at an initiator that requires access. In the present embodiment, if the receiver of the token is again the initial holder of the token indicating therefore the token was passed around the ring without an initiator subsystem requiring access, the initial holder will maintain the token, and if granted access at step 314, will issue a dummy request to enable the arbitration to be completed.
The request issued by an initiator subsystem contains an address of the intended target. Preferably, the request further contains an address of a particular resource in the target. The target subsystems monitor each request issued to determine whether the address corresponds to the address of the target, step 322, and if a correspondence (e.g., a match) is determined, the target subsystem responds to the request a predetermined number of clock cycles later, step 324. For example, if the request is a write operation, the target subsystem receives the data and issues a completion response the predetermined number of clock cycles after the issuance of the request. If the request is a read operation, the target subsystem provides the read data the predetermined number of clock cycles after the issuance of the request. In one embodiment, a 1 : 1 match between the issued address and target address is contemplated. However, various levels of correspondence, such as comparisons performed in conjunction with masks to match one or more devices are also contemplated.
As noted above, when a packet bus cycle is allocated to a subsystem, this subsystem has the right to issue a request during the slots associated with that bus cycle. This allows a pre-allocation of the bandwidth to the different subsystems. The advantages of a TDMA are guaranteed and predictable bandwidth, since the bus cycles are pre-allocated to the different subsystems in advance. This feature is critical in supporting clients with fixed real time bandwidth requirements. The combination of these two concepts, time division multiplex access and fixed-latency pipelined transactions, provides for a very high-performance communications protocol that guarantees both latency and bandwidth.
As briefly described above, the first level of arbitration is implemented in the following way. In one embodiment, the total available number of bus cycles are divided up into recurring frames. For purposes of discussion, the frame is divided up in a number of packets, which corresponds to a number of cycles. For example, a 256 cycle frame is divided up in 128 2-cycle packets. Thus, a frame could be composed of 128 2-cycle packets, and in such a design the frame would repeat every 256 bus cycles. The first level arbitration scheme revolves around the allocation of these packets to the different subsystems. Since only initiator subsystems can issue commands, the packets should only be allocated to the initiator subsystems.
In the present embodiment, every initiator subsystem contains a synchronous counter, which runs in lock step with all other counters. The output of this counter is used to index a programmable logic circuit whose output determines if the packet associated with the output of the current counter value is allocated to this particular interface module. By implementing a counter in each initiator interface module and maintaining the counters in lock step, the system implements a fully distributed TDMA time wheel, preferably as long as no two initiator interface modules are allocated the same packet. The programmable logic circuits in the interface modules are pre-programmed at system initialization with information regarding allocation and/or may be re- programmed during run-time of the system by commands issued over the communications bus to the different initiator subsystems. This allows for dynamic re-allocation of the bandwidth in the system.
The allocation of the packet occurs preferably during a bus cycle before the occurrence of the packet; for example, the bus cycle immediately prior to the occurrence of the packet. Although a variety of implementations may be used, the most versatile implementation of the programmable logic circuit is as a lookup table in the form of a small random access memory (RAM). The depth of such a RAM is equal to the number of packets in the frame. Thus, the TDMA time wheel is implemented simply by indexing into the RAM with the frame counter. Preferably, the RAM where the tables are stored are user-visible and accessible for reading and writing over the communication bus by software using the same read/write commands as used in standard data transactions. This allows the bandwidth allocation of the system to be changed at transaction speeds comparable to normal read/write transactions over the bus, and to change the allocation to unevenly distribute bandwidth as needed to support certain applications.
In certain situations, subsystems may not use one or more of the preallocated cycles. This may occur as although some system traffic is predictable, even the predictable traffic may not be uniform. For instance, frame buffer traffic transferred to a display is very predictable, but the traffic flow is interrupted while the display performs horizontal and vertical retrace operations. Furthermore, some initiator subsystems may have very low performance requirements, and pre-allocating any bandwidth to such a subsystem may lead to a lot of unusable cycles. Therefore, it is desirable to enhance the time-division multiplexed access protocol with the second level of arbitration.
At the second level of arbitration, packets that are not pre-allocated or packets that were pre-allocated but not used are contended for by the other initiator subsystems on the bus in a fair round-robin scheme using a token passing mechanism. This second level of arbitration ensures that no packet remains unused if any subsystem in the system can use the bus. The mechanism of the second level of arbitration retains all the advantages of the previously described communications method, but increases the overall efficiency of use of the available system bandwidth.
In the present embodiment, the second level of arbitration is implemented as a fair round-robin scheme using a token passing mechanism. All initiators that participate in the second level of arbitration are connected in a ring to pass the token around. Typically, those initiators that can tolerate a unpredictable latency participate in the second level of arbitration. In addition, it is contemplated that subsystems that participate in the first level of arbitration can also participate in the second level of arbitration. A token is allocated to one initiator subsystem at initialization. The token signifies a conditional right to issue bus commands during the bus cycles associated with the current packet. This right is conditional because it is dependent on the pre-allocated owner of the packet not asserting its right to use the packet or on the packet being unallocated. When a packet remains un-allocated after the above-described first level of arbitration, the packet is arbitrated for using the second level of arbitration. Thus, the TDMA scheme actually allocates a right of first refusal: the pre-allocated owner must assert its right to use its packet, or else the packet is made available to the initiator that ends up with the token.
On every arbitration cycle (i.e. every packet), the token may remain at the initiator that held it on the previous cycle, or it may pass around the ring to the next initiator that is requesting the token, unlike existing token ring systems, in one embodiment of the present invention the token may be passed completely around the ring back to the initiator that started it. The token's action depends upon system conditions and fairness constraints as set forth in the system's rules. For example, it is unfair for one initiator to use the token for more than one packet, since other initiator subsystems may be waiting. It is equally unfair for an initiator subsystem to be forced to give up the token before having a chance to use it. Thus, in one embodiment, the token-passing rules are structured such that an initiator subsystem that requests and receives the token may keep it until getting an opportunity to initiate precisely one packet's worth of commands across the communication bus. That opportunity may occur in the arbitration cycle in which the token arrives, or it may not occur for a large number of cycles, depending on the usage decisions made by the devices to which the packets are pre-allocated to owners. Finally, once the initiator with the token has a chance to use it, the initiator starts the token around the ring by driving its TokenOut signal to logic one. If no other initiator requests the token, it will pass completely around the ring and return to its sender.
In order to prevent the token from circulating around the ring indefinitely, it is preferred that the initiator subsystem that starts the token must temporarily break the ring. Thus, the initiator subsystem that starts the token on any given arbitration cycle can end up with it again by the end of that arbitration cycle. It should be noted that it is desirable that the communication bus signals always be actively driven so that the wires do not float to an undetermined state where improper bus operation and increased power dissipation may occur. As such, the initiator that ends up with the token should initiate a bus command whenever no pre-allocated packet owner asserts its right of first refusal. In such a case, the initiator with the token should initiate a useless command (a harmless read or write, or a NOP/Idle command) if it has no useful commands to issue. If the initiator that owns the token at the beginning of the arbitration cycle has not had an opportunity to use the token and has commands to issue, it can keep the token. If not, the initiator pulls a signal wire high (e.g., Token Out) indicating to other initiator subsystems that the token is available for the contention through the second level of arbitration. The initiator subsystem that has the token (i.e. the initiator subsystem that used the token last) passes the token on to the next initiator subsystem in the ring, which then, in turn, can appropriate the token if it can use the cycle. If not, the token is passed around until the token is allocated, in which case that specific initiator subsystem keeps the token as a starting point for the next token ring arbitration. If no initiator subsystem uses the token, it comes back to the initiator that initially had the token. All packet arbitration, both the first level and second level, occurs in the same cycle. This second level of arbitration does not provide guaranteed bandwidth, but does provide for guaranteed access. Using the second level of arbitration, no initiator system can command the bus for more than 1 packet if another initiator also wants access to the bus.
Figure 4a illustrates the states of one embodiment of the second level of arbitration and the table of Figure 4b sets forth a brief explanation of the states. In this embodiment the propagation delay around the ring is minimized by two implementation aspects. First, the interface module generating the token does so purely as a function of the current state. Furthermore, the interface modules along the way decide to pass the token or not purely as a function of the current state. These two features get the token into the ring as quickly as possible and prevent any delay-inducing uncertainty in how long the token will need to propagate through an intermediate interface module.
The input to the states are ReqToken, Tokenin, Assert, Reset_n, First and the outputs are TokenOut, Grant Token and the next state. ReqToken indicates that the interface is requesting the token, Tokenin indicates that the token is available to this initiator, Assert is used to indicate that the preallocated initiator will use its packet, reset_n is used to force the system into a reset state, TokenOut is used to pass the token to the next initiator in the ring, GrantToken is used to tell the initiator that it should prepare to issue a request and Next State indicates the next state of the particular interface module of the initiator subsystem.
In state NOTOKEN, the interface module of the subsystem neither desires nor has the token. Thus, if the token arrives on Tokenin it is immediately passed on via TokenOut. The state machine leaves NOTOKEN once it samples the interface module requesting the token. In state WANT TOKEN, the interface module does not have, but desires, the token. It therefore breaks the ring, such that when the token arrives on Tokenin this module will grab it. Once the token arrives, the interface module will use it (by noticing GrantToken) if no preallocated slot owner drove Assert. If Assert is active, then the module will keep the token until it gets an opportunity to us it, so the state machine transitions to HAVETOKEN.
In state HAVETOKEN, the module has the token and is waiting for an opportunity to use it. As soon as no module raises Assert, the slot is available and this module will use it (by noticing GrantToken). When either this occurs or the interface module stops asserting ReqToken (presumably because its request was satisfied via a pre-allocated slot), the state machine transitions to GENTOKEN. Since this module has the token, there is no need to pay attention to the Tokenin line (e.g., it must be zero). In state GENTOKEN, the module will start the token around the ring at the beginning of the arbitration cycle. The module must break the ring in case no module is in WANTTOKEN (i.e., is actively requesting the token). If the token makes it all the way around the ring back to this module, the interface module will receive GrantToken and be required to initiate a command, independently of whether the device requested the token. While the state machine is in reset (Reset_n is driven low), the starting state is set to NOTOKEN for all interface modules except the one labeled First. The First device is unique in that it starts off with the token at reset, thus ensuring that there is precisely one token in the ring at any time. Furthermore, it should be noted that in the present embodiment, the Don't Care entries in the ReqToken column (other than during reset) guarantee that precisely one device always initiates a command. The interface module that ends up with the token at the end of arbitration must initiate a command if no pre-allocated slot owner drives the Assert line high.
Using the above described allocation methodology, a variety of clients having varying requirements can be supported. The same system can be programmed to meet particular client needs; thus, the proper portion of preallocated clients and packets can be programmed according to system needs.
The minimum set of commands that are issued over the bus are a read command and a write command. The communications method can naturally be extended to support other commands, including specially designed control commands to control the subsystems coupled to the bus. Figure 5 shows a timing diagram of exemplary transactions across such a fully pipelined, fixed- latency bus that operates in accordance with the teachings of the present invention. Every transaction consists of an arbitration part, a command/address part, and a data/response part. Since the bus is a fixed-latency bus, there is a fixed delay (in number of clock cycles) between the two transactions, i.e., request and response. In Figure 5, the latency between arbitration and command/address is one cycle, the latency between command/address and data is assumed to be 2 two cycles, and a packet (the arbitration unit) is one bus cycle. In cycle 1, initiator subsystem I-D asserts its right to use its pre-allocated packet, so in cycle time slot 2, I-D issues a command and address on the bus to a target. For explanation purposes, the command is labeled Command 1. All target interface modules contain address decoding logic to decide if a transaction is addressed to their client subsystem. In time slot 4, the addressed target interface module responds to the request command. In case of a read request command, the read data is sent from the target to the initiator. In case of a write request command, the data to be written is sent from the initiator to the target in cycle 4.
A response signal that is driven by the target in cycle 4 from the target over the data response lines indicates if the transaction was successful or not. In the one embodiment, there are 4 possible responses: "Valid",""Busy","Retry" and "No Response". The "Valid" response indicates that the transaction was successful. The "Busy" response indicates that the target module could respond to the command. The "Retry" response is only allowed for read commands, and indicates that the target is attempting to satisfy the command but could not do so within the fixed bus latency. The initiator must retry its request later. In case of a "No Response", the address provided with the command did not match any of the targets, and therefore no target responded.
Continuing to reference Figure 5, a second arbitration cycle begins in clock cycle 2 and another transaction is issued in cycle 3. In this situation, no pre-allocated initiator asserted its right to use the bus cycle, so initiator subsystem I-E, which ended up with the token in cycle 2, issues a command and address in cycle 3. The data and response signals associated with this the command are issued transfer in cycle 5, a bus cycle latency later.
As noted earlier, the interface modules preferably are programmable to provide for dynamic reallocation of the packets. In one embodiment, the interface modules for each subsystem contain a set of configuration registers (240, Figure 2) to store the configuration data of the modules and the bus. Preferably, the configuration registers store data identifying packets preallocated to the corresponding subsystem. The configuration registers have addresses in the system address space, and therefore can be read from and written to using available bus transactions. Preferably, there are 2 sets of configuration registers, buffered and unbuffered configuration registers. Data written to an unbuffered register is visible to the interface module and to the rest of the system immediately after a write operation to the register. Data written to a buffered register is stored in a holding register location and only becomes visible after that data has been transferred from the holding register to the configuration register. The transfer is initiated through a broadcast command over the bus. A broadcast command is a command that affects all interface modules simultaneously. Thus, buffered registers typically are used when it is desirable to update several interface modules simultaneously before the new data can be allowed to be seen by the rest of the system.
One example of the use of buffered registers is the assignment of the address map in the system. Each target interface or module has a locally stored mask and match data value that specifies to which address(es) the target will respond. The address on the bus is compared to the match field and the result is masked. If all bits in the masked result are 0, this specific interface module is addressed. The mask/match data is stored in a configuration register and can be written to by regular bus commands. Proper operation of the system requires that target address ranges must be uniquely assigned in the system. If address changes are to be implemented, for example, the swapping of addresses, the new assignments should take place simultaneously in all affected interface modules. Otherwise, overlapping address-spaces may exist during the transition from one address map to the other.
As noted earlier, this invention is not limited to the use of buffered and unbuffered configuration registers. Furthermore, in systems where some or none of the data in the configuration registers needs to change during the lifetime of the system, all or some of the data can be stored in read only memory (ROM). ROM typically requires less die area, and prevents undesired writes to configuration locations that resulted in improper system configuration. Preferably, a typical system contains interface modules having a combination of buffered configuration registers, unbuffered configuration registers, and ROM. Information included in the configuration registers consists of data regarding packets preallocated to initiator subsystem, latency, frequency of communication between interface module and the coupled address device of the target for purposes of request address matching.
An important feature of the above is the fact that configuration data can be written during real-time operation of the system using available normal read and write commands as the registers are part of the system address space. This allows the system designer to change many significant features of the communication system via simple software or even pre-computed command sequences stored in a ROM.
One important and unique enhancement in the present invention over the prior art is the fact that the bus latency is programmable at initialization of the system or during the run-time of the system. In the preferred embodiment, the latency is programmable over the communications bus. Because of the fully pipelined nature of the communication bus, such an operation requires a broadcast command so the pipeline depth is changed simultaneously in all interface modules. When low-latency responses are important to the system (such as access to memory), a low latency can be programmed into the communications system by writing the new latency into the appropriate configuration registers of the interface modules. On the other hand, when most of the critical communications over the bus is between long-latency subsystems (systems that can not respond very quickly), the system latency can be set higher. The fact that the latency can be programmed at speeds comparable to regular bus transactions is a significant improvement over the prior art.
Not only can the latency be changed, but also the bandwidth allocation can be changed using the same method of updating the appropriate configuration registers. The fact that the bandwidth allocation can be programmed at speeds comparable to a normal read or write transaction over the communications bus, either at initialization or during the run-time of the system is a significant improvement over prior art.
The combination of programmable bandwidth allocation and programmable bus latency provides distinct advantages; most particularly, the system designer can optimize the performance of the system for the particular configuration of initiators and targets. Furthermore, the optimization can be maintained through a variety of conditions and configurations as the bandwidth allocation and latency can be updated during operation. Furthermore, as will be described subsequently, additional flexibility and configurability is contemplated by enabling an initiator device to program the frequency at which a target and its interface module communicate. This is particularly desirable to adapt to varying subsystem timing requirements of the initiator and/or target client.
Preferably, the bus functionability further is enhanced by adding a set of system-level control wires (referred to in Figure 1 as flags) that manage out-of- band communication. Out-of-band signals are signals that do not follow the address/data-oriented pipelined bus model. Examples include device status signals, interrupt, and handshake signals. In the preferred embodiment, these wires are also used for enhancing the performance of the basic read operations in the following manner.
For example, in the operation of the system as illustrated by Figure 5, if the target subsystem can respond to a read request command in time to meet the bus latency requirement, a "Valid" response is transmitted over the response bus. In the case where the target subsystem cannot respond in time to meet the bus latency requirement, a "Retry" response is transmitted. Rather than requiring the initiator to estimate when the desired data is available, which results in additional wasted bus cycles or increased latency, the target interface module uses one of the out-of-band signals (referred to herein also as flags). The target interface module uses the flags to provide a precise indication back to the initiator once the target interface module has obtained the data. Figure 6 illustrates the retry mechanism using the out of band signal lines. When a client cannot meet the latency of the communications bus, the target interface module issues a response packet to the initiator to indicate a "Retry". In the same time slot the bit number of a Flag wire signal is transmitted on the FlagNum lines (which are, for purposes of simplification, part of the Flag lines of Figure 1). This indicates to the initiator subsystem that it should monitor that identified flag for an indication of when to retry.
In the example of Figure 6, a Read request command is transmitted in cycle 1. The bus latency is two cycles, so the data should be returned by the target interface module in cycle 3. However, the data is not available yet, so the target returns a retry response on the response wires and it returns the number of a flag wire on the FlagNum bus, in this example number 7. When the target interface module receives the data from its client (i.e., coupled device), it stores the data in local memory and it drives the indicated flag wire high, indicating to the initiator that the data is now available. In Figure 6, after cycle 4, the target indicates on Flag 7 that the data is available. At this time, the initiator subsystem can now re-issue the request command which in Figure 3 occurs on cycle 6. When the target interface module satisfies the request command (cycle 8 in Figure 3), it deasserts the Flag that has been associated with this request command.
The communications system preferably contains a number of flags that are allocated to the different target interface modules. Flag wires can be shared between modules or can be dedicated to a specific module. In the preferred embodiment, the allocation of the flags is stored in the configuration registers and therefore this allocation is programmable over the communications bus and at a speed comparable to normal bus transactions.
The flag wires can be used for a variety of purposes. For example, a different use of the flag wires is the management of handshake signals and interrupt signals. These are signals that indicate a special event in the interface modules or the subsystem. Examples of such signals are error signals or signals that indicate busy status (stop sending requests) or ready status (subsystem is ready to take accept requests). These signals are inherently out of band signals and are transferred using the flag wires. For these purposes, the flag wires is are uniquely allocated to target interface modules and their subsystem clients for the specific purpose of interrupt or handshake signaling. In the preferred embodiment, this data allocation is stored in a configuration registers so it may be dynamically varied as system needs change. The same information is programmed in the initiator subsystem to let the initiator know the flag wires to monitor for handshake or interrupt purpose.
As noted above, the interface modules between the bus and the subsystems contain logic to implement the bus protocols while interfacing with the client. Furthermore, it is preferred that the interface contains additional logic (245, Figure 2) to derive a client clock of the subsystem from the system bus clock. The interface module through the configuration registers provides a level of programmability to enable subsystems reuse in many different systems with very different system features and performance requirements. The frequency of the system bus clock is determined by system requirements, most notably the total bandwidth required on the bus to satisfy the required performance of the system. The client clock frequency of a client can be independent of the system bus clock, and should be related to the functionality required of the client. The interface modules therefore provide a powerful means of decoupling the system and subsystem requirements.
In a traditional fixed-latency bus system, increasing the available bandwidth requires increasing the operating frequency of the bus and all of its clients, since the clients would need to respond in less time to meet the latency. With the system of the present invention, the bus latency (measured in cycles) can be increased as the bus frequency is increased, which tends to allow clients to operate a approximately consistent frequencies, since the total latency (in time) stays relatively constant. It is contemplated that the decoupling feature be extended to asynchronous connections between the interface module and the client.
The requirement for frequency decoupling is primarily based on the rich variety of systems that can benefit from this approach. One example system is a wireless personal communicator, which requires fixed-performance real-time subsystems (to implement the wireless communications) along with low- performance, computer-derived subsystems such as a microcontroller core, embedded memory, keypad input and liquid crystal display (LCD) output. Key design goals for the communicator are to provide reliable communications and as many features as possible (implemented in software on the microcontroller) at low power. A second exemplary system is a set-top box for receiving digital satellite television broadcasts. The set-top box design requires much higher bit- rate data communications and higher computational performance to decompress the video and audio data, and must guarantee performance across the entire system to avoid dropping frames. A third example is an Asynchronous Transfer Mode (ATM) switch. The switch design likely needs a central processing unit (CPU) only to monitor performance and provide diagnostics; the normal operation would involve switching packets between identical subsystems across the shared interconnect. As in most circuit switching applications, performance guarantees are critical to proper operation of the ATM switch.
The opportunities presented by design re-use for a variety of systems are now illustrated. A wireless system for video-on-demand could combine the two- way wireless subsystems from the personal communicator, the ATM packet framing engine from the ATM switch, and the video decompression subsystems from the set-top box. Combination of subsystems that were never intended to interoperate into rapidly integrated, cost-effective systems can be achieved.
The invention has been described in conjunction with the preferred embodiment. It is evident that numerous alternatives, modifications, variations and uses will be apparent to those skilled in the art in light of the foregoing description.

Claims

CLAIMSWhat is claimed is:
1. A computer bus system comprising: a synchronous bus operative during a number of bus cycles, said number of bus cycles divided up into recurring frames, each frame further divided into packets comprising at least one clock cycle; at least one initiator subsystem coupled to the bus, the at least one initiator configured to have at least one packet preallocated to the at least one initiator subsystem, said initiator subsystem configured to send out a request during a clock cycle within the at least one preallocated packet, said request comprising a command identifying an operation to be performed by a target subsystem and an address of a target subsystem; at least one target subsystem, said target subsystem configured to receive the address of the request and determine if the address corresponds to the address of the target subsystem, wherein if the address of the request corresponds to the address of the target subsystem, said target subsystem configured to respond to the request on a second clock cycle, said second clock cycle occurring a predetermined number of cycles after said first clock cycle.
2. The computer bus system as defined in claim 1, wherein the at least one initiator comprises a plurality of initiator subsystems and wherein at least two of the plurality of said initiator subsystem having at least one of said packets pre-allocated to it.
3. The computer bus system as defined in claim 1, wherein every pre-allocated packet is allocated to at most one initiator.
4. The computer bus system as defined in claim 1, where the decision on which initiator subsystem has access to the bus on said first cycle occurs during a cycle which occurs before said first cycle.
5. The computer bus system as defined in claim 4, wherein when a packet is pre-allocated to an initiator subsystem but not used by said initiator subsystem or when a packet is not pre-allocated, said initiator subsystem arbitrating access for the packet in accordance with a fair round robin scheme.
6. The computer bus system as defined in claim 1, where said initiator subsystem and target subsystem comprise an addressable memory configured to configuration data, said configuration data controlling said predetermined number of cycles and said pre-allocation of packets, and where said addressable memory can be accessed over the bus by requests similar to those used for transmitting data between initiator subsystems and target subsystems.
7. The computer bus system as defined in claim 1, where said predetermined number of cycles is programmable in software in real-time operation at speeds comparable to normal transactions over the bus.
8. The computer bus system as defined in claim 1, where said pre- allocation of packets to said initiator subsystem is programmable in software in real-time operation at speeds comparable to normal transactions over the bus.
9. The computer bus system as defined in claim 1, where the bus includes a set of flag wires, each flag wire being allocated to one subsystem, and said flag wires being used for transmitting control information between the clients and said control information being unrelated to either the address space of the clients or the pre-allocation of the time-segments.
10. The computer system as set forth in claim 9, wherein said control information being unrelated to either the address space of the clients or the pre- allocation of the packets.
11. The computer bus system as defined in claim 9, where the bus further comprises a set of flag number wires, indicating the number at of at least one of said flag wires that conveys information.
12. The computer bus system as set forth in claim 11, wherein the flag wires are used to manage transactions that cannot be responded in time to meet determined a latency.
13. The computer bus system as set froth in claim 12, wherein if the data cannot be returned by a target subsystem at the request of an initiator subsystem on said second cycle, said target subsystem configured by transmitting a number of a flag wire over said flag number wires on said second cycle; when said target subsystem has the data available or can determine that said data will soon be available, said target module configured to assert a signal on one of said flag wire that corresponds to the number transmitted on said flag number wires; and said initiator subsystem, is configured to detect the transition on said flag wire, and re-issue said transaction, wherein said target subsystem is configured to respond to the request and de-asserts the signal on said flag wire.
14. The computer bus system as set forth in claim 1 , wherein the subsystem comprises an interface module and a client module, the client clock configured to be synchronous to the clock of the bus, the client clock being derived from the bus clock and the ratio of the bus clocks to the client clock is programmable in software, hence decoupling the clock frequency of the client from said bus clock frequency.
15. A computer bus system as set forth in claim 1, wherein the subsystem comprises an interface module and a client module clock frequency of the client configured to be unrelated to the bus clock frequency, said interface module comprising logic configured to synchronize the client clock signal to the bus clock signal.
16. The bus system as set forth in claim 15, wherein the logic comprises programmable latency.
17. A method for transmitting data between clients over a shared bus consisting of the following steps:
An initiator sending a request on a first clock cycle, said request comprising of at least a command and a target address, said first cycle being part of a packet that is pre-allocated to the initiator, said packet comprising at least on clock-cycle, a plurality of packets comprising a recurring frame; at least one target receiving said request in said first cycle and addressed one target of the at least one target responding to said request if the address corresponds to the address space of said target; and said addressed target responding to said request on a second cycle, said second cycle occurring a pre-determined number of cycles after said first cycle.
18. The method of transmitting data as defined in claim 17, further comprising the steps of when a clock cycle is pre-allocated to an initiator but not used or a clock-cycle is not pre-allocated to any initiator, executing a second level of arbitration wherein said unused clock cycle is allocated to another initiator.
19. A method of changing system level parameters in a computer bus system comprising a bus coupling at least two clients, comprising of at least one initiator and one target, said method comprising the following steps: defining, in each client, a first addressable memory; defining, in each client, a second addressable memory containing configuration data used by said client; an initiator writing new configuration data to said first memory in said clients by issuing write commands over the bus addressed to said first memory in said clients; and said initiator issuing a single command to all clients as the result of which the configuration data is transferred from said holding memory to said second memory wherein the new configuration data is used by the clients.
20. In a single chip computer bus system comprising a bus coupling at least one initiator client and at least one target client, a method for communication across the bus comprising steps of: pre-allocating at least one packet to said initiator client, the at least one packet determined by dividing the bus cycles into recurring frames, each frame divided up into packets consisting of at least one clock cycle; the initiator sending out a request on a first cycle , said first cycle being part of a time-segment that is pre-allocated to said initiator and said request comprising at least a command field and an address field, said address field identifying a target client; the target client receiving the address located in the address field and if an address of the target corresponds to the address located in the address field, said target client receiving said request during said first cycle and responding to said request on a second cycle, said second cycle occurring a predetermined number of cycles after said first cycle.
PCT/US1998/015006 1997-09-05 1998-07-22 A fully-pipelined fixed-latency communications system with a real-time dynamic bandwidth allocation WO1999013405A1 (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
EP98936927A EP1027657B1 (en) 1997-09-05 1998-07-22 A fully-pipelined fixed-latency communications system with a real-time dynamic bandwidth allocation
KR1020007002388A KR100572372B1 (en) 1997-09-05 1998-07-22 Fully pipelined fixed standby communication system with real-time dynamic bandwidth allocation
DE69832410T DE69832410T2 (en) 1997-09-05 1998-07-22 PIPELINE COMMUNICATION SYSTEM WITH FIXED LATENCY TIME USING DYNAMIC REAL-TIME BANDWIDTH ALLOCATION
AU85764/98A AU8576498A (en) 1997-09-05 1998-07-22 A fully-pipelined fixed-latency communications system with real-time dynamic bandwidth allocation
AT98936927T ATE310277T1 (en) 1997-09-05 1998-07-22 FIXED LATENCY PIPELINE COMMUNICATIONS SYSTEM USING REAL-TIME DYNAMIC BANDWIDTH ALLOCATION
JP2000511115A JP4259751B2 (en) 1997-09-05 1998-07-22 Fully pipelined fixed latency communication system that dynamically allocates bandwidth in real time

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US08/924,368 US5948089A (en) 1997-09-05 1997-09-05 Fully-pipelined fixed-latency communications system with a real time dynamic bandwidth allocation
US08/924,368 1997-09-05

Publications (1)

Publication Number Publication Date
WO1999013405A1 true WO1999013405A1 (en) 1999-03-18

Family

ID=25450136

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1998/015006 WO1999013405A1 (en) 1997-09-05 1998-07-22 A fully-pipelined fixed-latency communications system with a real-time dynamic bandwidth allocation

Country Status (9)

Country Link
US (1) US5948089A (en)
EP (1) EP1027657B1 (en)
JP (1) JP4259751B2 (en)
KR (1) KR100572372B1 (en)
CN (1) CN1150460C (en)
AT (1) ATE310277T1 (en)
AU (1) AU8576498A (en)
DE (1) DE69832410T2 (en)
WO (1) WO1999013405A1 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7739437B2 (en) 2005-01-31 2010-06-15 Panasonic Corporation Access arbiter and arbitrable condition verification device
US7913014B2 (en) 2004-09-28 2011-03-22 Nxp B.V. Data processing system and method for memory arbitration
EP2538340A1 (en) * 2008-11-05 2012-12-26 MediaTek, Inc Shared resource arbitration
US8667199B2 (en) 2007-03-22 2014-03-04 Arm Limited Data processing apparatus and method for performing multi-cycle arbitration
CN105760607A (en) * 2016-02-22 2016-07-13 烽火通信科技股份有限公司 Simulation assembly for simulating bus effective bandwidth based on token bucket and method

Families Citing this family (96)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6895510B1 (en) * 1997-11-24 2005-05-17 International Business Machines Corporation Mutual internet authentication between a client and server utilizing a dummy IOP request
US6138187A (en) * 1998-08-21 2000-10-24 International Business Machines Corporation Method and system for increasing spatial reuse in a serial storage architecture subsystem
US6182183B1 (en) 1998-11-13 2001-01-30 Sonics, Inc. Communications system and method with multilevel connection identification
US6487606B1 (en) * 1998-11-18 2002-11-26 Nortel Networks Limited System and method for delivering messages through a totem communications system
TW388817B (en) * 1998-11-20 2000-05-01 Via Tech Inc Method reducing latency of writing data in memory
JP3953243B2 (en) * 1998-12-29 2007-08-08 インターナショナル・ビジネス・マシーンズ・コーポレーション Synchronization method and apparatus using bus arbitration control for system analysis
US6449671B1 (en) * 1999-06-09 2002-09-10 Ati International Srl Method and apparatus for busing data elements
GB2352144A (en) * 1999-07-16 2001-01-17 Texas Instruments Ltd Data transfer between memory nodes
US6801985B1 (en) * 1999-09-10 2004-10-05 Texas Instruments Incorporated Data bus using synchronous fixed latency loop including read address and data busses and write address and data busses
DE19946716A1 (en) * 1999-09-29 2001-04-12 Infineon Technologies Ag Process for operating a processor bus
US6701397B1 (en) 2000-03-21 2004-03-02 International Business Machines Corporation Pre-arbitration request limiter for an integrated multi-master bus system
DE10022479B4 (en) * 2000-05-09 2004-04-08 Infineon Technologies Ag Arrangement for the transmission of signals between a data processing device and a functional unit in a main memory system of a computer system
US7325221B1 (en) 2000-08-08 2008-01-29 Sonics, Incorporated Logic system with configurable interface
US6785284B1 (en) * 2000-08-10 2004-08-31 Infineon Technologies North America Corp. Interleavement for transport of frames and cells
US7165094B2 (en) * 2001-03-09 2007-01-16 Sonics, Inc. Communications system and method with non-blocking shared interface
US6785753B2 (en) * 2001-06-01 2004-08-31 Sonics, Inc. Method and apparatus for response modes in pipelined environment
US20030004699A1 (en) * 2001-06-04 2003-01-02 Choi Charles Y. Method and apparatus for evaluating an integrated circuit model
ATE297568T1 (en) * 2001-09-26 2005-06-15 Siemens Ag METHOD FOR ARBITRATION OF ACCESS TO A DATA BUS
US6804738B2 (en) * 2001-10-12 2004-10-12 Sonics, Inc. Method and apparatus for scheduling a resource to meet quality-of-service restrictions
US7099922B2 (en) * 2002-01-23 2006-08-29 International Business Machines Corporation Method and system for simultaneous management of multiple tokens on a communication ring
US7315551B2 (en) * 2002-03-15 2008-01-01 Lockheed Martin Corporation Synchronous low voltage differential I/O buss
US7356633B2 (en) * 2002-05-03 2008-04-08 Sonics, Inc. Composing on-chip interconnects with configurable interfaces
US7254603B2 (en) * 2002-05-03 2007-08-07 Sonics, Inc. On-chip inter-network performance optimization using configurable performance parameters
US7194566B2 (en) * 2002-05-03 2007-03-20 Sonics, Inc. Communication system and method with configurable posting points
US7302691B2 (en) * 2002-05-10 2007-11-27 Sonics, Incorporated Scalable low bandwidth multicast handling in mixed core systems
US6880133B2 (en) * 2002-05-15 2005-04-12 Sonics, Inc. Method and apparatus for optimizing distributed multiplexed bus interconnects
JP4025593B2 (en) * 2002-07-11 2007-12-19 富士通株式会社 Broadcast communication data delivery apparatus and broadcast communication system
US7426182B1 (en) * 2002-08-28 2008-09-16 Cisco Technology, Inc. Method of managing signal processing resources
US7032046B2 (en) * 2002-09-30 2006-04-18 Matsushita Electric Industrial Co., Ltd. Resource management device for managing access from bus masters to shared resources
US7243264B2 (en) * 2002-11-01 2007-07-10 Sonics, Inc. Method and apparatus for error handling in networks
US6976106B2 (en) * 2002-11-01 2005-12-13 Sonics, Inc. Method and apparatus for speculative response arbitration to improve system latency
US7266786B2 (en) * 2002-11-05 2007-09-04 Sonics, Inc. Method and apparatus for configurable address mapping and protection architecture and hardware for on-chip systems
US7603441B2 (en) * 2002-12-27 2009-10-13 Sonics, Inc. Method and apparatus for automatic configuration of multiple on-chip interconnects
US20040210696A1 (en) * 2003-04-18 2004-10-21 Meyer Michael J. Method and apparatus for round robin resource arbitration
US7149829B2 (en) * 2003-04-18 2006-12-12 Sonics, Inc. Various methods and apparatuses for arbitration among blocks of functionality
US7139854B2 (en) * 2003-06-10 2006-11-21 Texas Instruments Incorporated Pipelining access to serialization tokens on a bus
US7194658B2 (en) * 2003-07-24 2007-03-20 Sonics, Inc. Various methods and apparatuses for interfacing of a protocol monitor to protocol checkers and functional checkers
US20050066097A1 (en) * 2003-09-04 2005-03-24 Matsushita Electric Industrial Co., Ltd. Resource management apparatus
US7296105B2 (en) * 2003-10-03 2007-11-13 Sonics, Inc. Method and apparatus for configuring an interconnect to implement arbitration
US9087036B1 (en) 2004-08-12 2015-07-21 Sonics, Inc. Methods and apparatuses for time annotated transaction level modeling
US8504992B2 (en) * 2003-10-31 2013-08-06 Sonics, Inc. Method and apparatus for establishing a quality of service model
US7665069B2 (en) * 2003-10-31 2010-02-16 Sonics, Inc. Method and apparatus for establishing a quality of service model
US20050175027A1 (en) * 2004-02-09 2005-08-11 Phonex Broadband Corporation System and method for requesting and granting access to a network channel
US7543088B2 (en) * 2004-03-11 2009-06-02 Sonics, Inc. Various methods and apparatuses for width and burst conversion
US7475168B2 (en) * 2004-03-11 2009-01-06 Sonics, Inc. Various methods and apparatus for width and burst conversion
JP4480427B2 (en) * 2004-03-12 2010-06-16 パナソニック株式会社 Resource management device
US20050210329A1 (en) * 2004-03-18 2005-09-22 Newisys, Inc. Facilitating system diagnostic functionality through selective quiescing of system component sensor devices
US7747771B1 (en) 2004-06-30 2010-06-29 Oracle America, Inc. Register access protocol in a multihreaded multi-core processor
US7739436B2 (en) * 2004-11-01 2010-06-15 Sonics, Inc. Method and apparatus for round robin resource arbitration with a fast request to grant response
US8032676B2 (en) * 2004-11-02 2011-10-04 Sonics, Inc. Methods and apparatuses to manage bandwidth mismatches between a sending device and a receiving device
US7277975B2 (en) * 2004-11-02 2007-10-02 Sonics, Inc. Methods and apparatuses for decoupling a request from one or more solicited responses
US7155554B2 (en) * 2004-11-02 2006-12-26 Sonics, Inc. Methods and apparatuses for generating a single request for block transactions over a communication fabric
JP4907166B2 (en) * 2005-01-12 2012-03-28 パナソニック株式会社 Resource management device
JP4419868B2 (en) * 2005-02-25 2010-02-24 ソニー株式会社 Information processing apparatus and method, memory control apparatus and method, recording medium, and program
US20060225015A1 (en) * 2005-03-31 2006-10-05 Kamil Synek Various methods and apparatuses for flexible hierarchy grouping
US20070067531A1 (en) * 2005-08-22 2007-03-22 Pasi Kolinummi Multi-master interconnect arbitration with time division priority circulation and programmable bandwidth/latency allocation
EP1955175A1 (en) * 2005-09-09 2008-08-13 Freescale Semiconductor, Inc. Interconnect and a method for designing an interconnect
US7694249B2 (en) * 2005-10-07 2010-04-06 Sonics, Inc. Various methods and apparatuses for estimating characteristics of an electronic system's design
JP4974508B2 (en) * 2005-10-28 2012-07-11 キヤノン株式会社 Bus master device, bus arbitration device, and bus arbitration method
EP2002334A1 (en) 2006-03-31 2008-12-17 British Telecommunications Public Limited Company Xml-based transfer and a local storage of java objects
EP2002336A1 (en) * 2006-03-31 2008-12-17 British Telecommunications Public Limited Company Server computer component
US20080082708A1 (en) * 2006-09-29 2008-04-03 Kar Leong Wong Token hold off for chipset communication
US8868397B2 (en) 2006-11-20 2014-10-21 Sonics, Inc. Transaction co-validation across abstraction layers
US20080120082A1 (en) * 2006-11-20 2008-05-22 Herve Jacques Alexanian Transaction Co-Validation Across Abstraction Layers
US8020124B2 (en) * 2006-11-20 2011-09-13 Sonics, Inc. Various methods and apparatuses for cycle accurate C-models of components
CN101232425B (en) * 2007-01-26 2010-04-21 京信通信系统(中国)有限公司 Bus type distributed control system master-salve station opposite terminal control method
US7814243B2 (en) * 2007-06-01 2010-10-12 Sonics, Inc. Shared storage for multi-threaded ordered queues in an interconnect
US8108648B2 (en) * 2007-06-25 2012-01-31 Sonics, Inc. Various methods and apparatus for address tiling
US8438320B2 (en) * 2007-06-25 2013-05-07 Sonics, Inc. Various methods and apparatus for address tiling and channel interleaving throughout the integrated system
US9495290B2 (en) * 2007-06-25 2016-11-15 Sonics, Inc. Various methods and apparatus to support outstanding requests to multiple targets while maintaining transaction ordering
US8229723B2 (en) * 2007-12-07 2012-07-24 Sonics, Inc. Performance software instrumentation and analysis for electronic design automation
TWI337517B (en) * 2008-03-04 2011-02-11 Inventec Corp Trace carrier
US8135878B1 (en) * 2008-04-07 2012-03-13 Tellabs San Jose Inc. Method and apparatus for improving throughput on a common bus
US8073820B2 (en) 2008-04-07 2011-12-06 Sonics, Inc. Method and system for a database to monitor and analyze performance of an electronic design
US8032329B2 (en) * 2008-09-04 2011-10-04 Sonics, Inc. Method and system to monitor, debug, and analyze performance of an electronic design
JP5460156B2 (en) * 2009-07-14 2014-04-02 キヤノン株式会社 Data processing device
US9342471B2 (en) * 2010-01-29 2016-05-17 Mosys, Inc. High utilization multi-partitioned serial memory
US20110213949A1 (en) * 2010-03-01 2011-09-01 Sonics, Inc. Methods and apparatus for optimizing concurrency in multiple core systems
US8972995B2 (en) 2010-08-06 2015-03-03 Sonics, Inc. Apparatus and methods to concurrently perform per-thread as well as per-tag memory access scheduling within a thread and across two or more threads
US8601288B2 (en) 2010-08-31 2013-12-03 Sonics, Inc. Intelligent power controller
US8438306B2 (en) 2010-11-02 2013-05-07 Sonics, Inc. Apparatus and methods for on layer concurrency in an integrated circuit
US9405700B2 (en) 2010-11-04 2016-08-02 Sonics, Inc. Methods and apparatus for virtualization in an integrated circuit
US8514889B2 (en) 2011-08-26 2013-08-20 Sonics, Inc. Use of common data format to facilitate link width conversion in a router with flexible link widths
US8798038B2 (en) 2011-08-26 2014-08-05 Sonics, Inc. Efficient header generation in packetized protocols for flexible system on chip architectures
US8711867B2 (en) 2011-08-26 2014-04-29 Sonics, Inc. Credit flow control scheme in a router with flexible link widths utilizing minimal storage
US8868941B2 (en) 2011-09-19 2014-10-21 Sonics, Inc. Apparatus and methods for an interconnect power manager
US9910454B2 (en) 2012-06-07 2018-03-06 Sonics, Inc. Synchronizer with a timing closure enhancement
US9378023B2 (en) * 2012-06-13 2016-06-28 International Business Machines Corporation Cross-pipe serialization for multi-pipeline processor
WO2015060820A1 (en) * 2013-10-22 2015-04-30 Hewlett-Packard Development Company, L.P. Hybrid circuit-packet switch
US10152112B2 (en) 2015-06-10 2018-12-11 Sonics, Inc. Power manager with a power switch arbitrator
US10268614B2 (en) 2016-04-19 2019-04-23 Nokia Of America Corporation Method and apparatus for a segmented on-chip digital interface block
US10745425B2 (en) 2016-07-01 2020-08-18 Shenzhen Yhlo Biotech Co., Ltd. Chemiluminescence enhancer and chemiluminescence immunodetection kit
US10540186B1 (en) 2017-04-18 2020-01-21 Amazon Technologies, Inc. Interception of identifier from client configurable hardware logic
US11245550B2 (en) * 2017-12-24 2022-02-08 Technion Research & Development Foundation Limited Message authentication based on a physical location on a bus
US10496594B1 (en) 2018-06-01 2019-12-03 Nxp Usa, Inc. Inter-processor communication method for access latency between system-in-package (SIP) dies
US10496593B1 (en) * 2018-06-01 2019-12-03 Nxp Usa, Inc. Inter-processor communication and signaling system and method

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4979096A (en) * 1986-03-08 1990-12-18 Hitachi Ltd. Multiprocessor system
US5414813A (en) * 1990-02-13 1995-05-09 Kabushiki Kaisha Toshiba Direct transfer from a receive buffer to a host in a token-passing type network data transmission system
US5465330A (en) * 1990-06-14 1995-11-07 Mitsubishi Denki Kabushiki Kaisha Network address managing method and system
US5553245A (en) * 1994-05-11 1996-09-03 Macronix International Co., Ltd. Automatic configuration of multiple peripheral interface subsystems in a computer system
US5701420A (en) * 1994-07-20 1997-12-23 Intel Corporation Method for initializing an array of configurable components
US5727169A (en) * 1990-06-12 1998-03-10 Sgs-Thomson Microelectronics, S.A. Electronically configurable connection device

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4199661A (en) * 1978-05-05 1980-04-22 Control Data Corporation Method and apparatus for eliminating conflicts on a communication channel
EP0077328A4 (en) * 1981-04-27 1985-06-26 Textron Inc Multi-master processor bus.
US5430848A (en) * 1992-08-14 1995-07-04 Loral Fairchild Corporation Distributed arbitration with programmable priorities

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4979096A (en) * 1986-03-08 1990-12-18 Hitachi Ltd. Multiprocessor system
US5414813A (en) * 1990-02-13 1995-05-09 Kabushiki Kaisha Toshiba Direct transfer from a receive buffer to a host in a token-passing type network data transmission system
US5727169A (en) * 1990-06-12 1998-03-10 Sgs-Thomson Microelectronics, S.A. Electronically configurable connection device
US5465330A (en) * 1990-06-14 1995-11-07 Mitsubishi Denki Kabushiki Kaisha Network address managing method and system
US5553245A (en) * 1994-05-11 1996-09-03 Macronix International Co., Ltd. Automatic configuration of multiple peripheral interface subsystems in a computer system
US5701420A (en) * 1994-07-20 1997-12-23 Intel Corporation Method for initializing an array of configurable components

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7913014B2 (en) 2004-09-28 2011-03-22 Nxp B.V. Data processing system and method for memory arbitration
US7739437B2 (en) 2005-01-31 2010-06-15 Panasonic Corporation Access arbiter and arbitrable condition verification device
US8667199B2 (en) 2007-03-22 2014-03-04 Arm Limited Data processing apparatus and method for performing multi-cycle arbitration
EP2538340A1 (en) * 2008-11-05 2012-12-26 MediaTek, Inc Shared resource arbitration
EP2538338A1 (en) * 2008-11-05 2012-12-26 MediaTek, Inc Shared resource arbitration
EP2544095A3 (en) * 2008-11-05 2013-05-01 MediaTek, Inc Shared resource arbitration
CN105760607A (en) * 2016-02-22 2016-07-13 烽火通信科技股份有限公司 Simulation assembly for simulating bus effective bandwidth based on token bucket and method

Also Published As

Publication number Publication date
DE69832410D1 (en) 2005-12-22
CN1301364A (en) 2001-06-27
KR20010023734A (en) 2001-03-26
KR100572372B1 (en) 2006-04-19
EP1027657B1 (en) 2005-11-16
EP1027657A1 (en) 2000-08-16
EP1027657A4 (en) 2003-04-09
DE69832410T2 (en) 2006-08-03
US5948089A (en) 1999-09-07
ATE310277T1 (en) 2005-12-15
JP2001516100A (en) 2001-09-25
CN1150460C (en) 2004-05-19
AU8576498A (en) 1999-03-29
JP4259751B2 (en) 2009-04-30

Similar Documents

Publication Publication Date Title
US5948089A (en) Fully-pipelined fixed-latency communications system with a real time dynamic bandwidth allocation
US6725313B1 (en) Communications system and method with multilevel connection identification
JP4638216B2 (en) On-chip bus
US5623672A (en) Arrangement and method of arbitration for a resource with shared user request signals and dynamic priority assignment
US6393506B1 (en) Virtual channel bus and system architecture
EP0737925B1 (en) Pipelined distributed bus arbitration system
US6148357A (en) Integrated CPU and memory controller utilizing a communication link having isochronous and asynchronous priority modes
US5507002A (en) Peripheral component interconnect special cycle protocol using soft message IDS
US20050204085A1 (en) Resource management device
US5566177A (en) Priority-based arbitrator on a token-based communication medium
US20030088722A1 (en) System and method for managing priorities in a PCI bus system
EP1820109B1 (en) Time-based weighted round robin arbiter
US7698485B2 (en) Round-robin bus protocol
JP2004246862A (en) Resource management device
US7069363B1 (en) On-chip bus
JPH04349558A (en) Architecture of workstation having selectable central processing unit
JP2002189705A (en) Multiprocessor arrangement
JPH1125036A (en) Arbitration system and method for arbitorating access
KR100335371B1 (en) Local memory arbiter and interface apparatus
JP3240863B2 (en) Arbitration circuit
KR100243868B1 (en) Arbiter logic in main computer system
JPH08171531A (en) Access conflict control method
JPH04238559A (en) Competition processing control system
JP2002073536A (en) Device and method for arbitrating bus
EP1459191A1 (en) Communication bus system

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 98810932.8

Country of ref document: CN

AK Designated states

Kind code of ref document: A1

Designated state(s): AL AM AT AT AU AZ BA BB BG BR BY CA CH CN CU CZ CZ DE DK DK EE EE ES FI GB GE GH GM HR HU ID IL IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT UA UG US UZ VN YU ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW SD SZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 1020007002388

Country of ref document: KR

WWE Wipo information: entry into national phase

Ref document number: 1998936927

Country of ref document: EP

REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

WWP Wipo information: published in national office

Ref document number: 1998936927

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: CA

WWP Wipo information: published in national office

Ref document number: 1020007002388

Country of ref document: KR

WWG Wipo information: grant in national office

Ref document number: 1998936927

Country of ref document: EP

WWG Wipo information: grant in national office

Ref document number: 1020007002388

Country of ref document: KR