US20070201365A1 - Video packet multiplexer with intelligent packet discard - Google Patents

Video packet multiplexer with intelligent packet discard Download PDF

Info

Publication number
US20070201365A1
US20070201365A1 US11/337,403 US33740306A US2007201365A1 US 20070201365 A1 US20070201365 A1 US 20070201365A1 US 33740306 A US33740306 A US 33740306A US 2007201365 A1 US2007201365 A1 US 2007201365A1
Authority
US
United States
Prior art keywords
packets
packet
multiplexer
congestion
queue
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/337,403
Inventor
Frederick Skoog
John Blanton
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Alcatel Lucent SAS
Original Assignee
Alcatel USA Sourcing Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Alcatel USA Sourcing Inc filed Critical Alcatel USA Sourcing Inc
Priority to US11/337,403 priority Critical patent/US20070201365A1/en
Assigned to ALCATEL USA SOURCING, L.P. reassignment ALCATEL USA SOURCING, L.P. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BLANTON, JOHN, SKOOG, FREDERICK
Priority to EP07000860.2A priority patent/EP1811725B1/en
Priority to CN200710007345.XA priority patent/CN101026572B/en
Publication of US20070201365A1 publication Critical patent/US20070201365A1/en
Assigned to ALCATEL-LUCENT USA INC. reassignment ALCATEL-LUCENT USA INC. MERGER (SEE DOCUMENT FOR DETAILS). Assignors: ALCATEL USA SOURCING, INC.
Assigned to ALCATEL USA SOURCING, INC. reassignment ALCATEL USA SOURCING, INC. CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: ALCATEL USA SOURCING, L.P.
Priority to US13/112,753 priority patent/US8861346B2/en
Assigned to ALCATEL LUCENT reassignment ALCATEL LUCENT ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ALCATEL-LUCENT USA INC.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/11Identifying congestion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/12Avoiding congestion; Recovering from congestion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/19Flow control; Congestion control at layers above the network layer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/24Traffic characterised by specific attributes, e.g. priority or QoS
    • H04L47/2416Real-time traffic
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/28Flow control; Congestion control in relation to timing considerations
    • H04L47/283Flow control; Congestion control in relation to timing considerations in response to processing delays, e.g. caused by jitter or round trip time [RTT]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/28Flow control; Congestion control in relation to timing considerations
    • H04L47/286Time to live
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/32Flow control; Congestion control by discarding or delaying data units, e.g. packets or frames

Definitions

  • This invention relates in general to network communications and, more particularly, to a method and apparatus for discarding packets.
  • packets of data may be lost for a variety of reasons. Some packets are randomly lost due to uncontrollable errors - for example, errors caused by noise on a transmission line, synchronization-issues, etc.-Some-packets are lost due to congestion, i.e., it is not possible for a network element to transmit all received packets in a timely manner.
  • Current discard mechanisms for IP QoS (quality of service) algorithms implement random selection schemes to determine which packets to discard without regard to the relative effect on the eventual output.
  • missing packets cause the destination device to request a retransmission of the missing information. This is not very feasible, however, in a network that has multicasting of real-time streams such as audio or video. Normally, there will not be enough time available for requesting and receiving the retransmitted packets, unless buffers at the destination device are very large.
  • the destination device waits for a certain amount of time before declaring a packet as lost.
  • some decoders may request retransmission, other decoders may correct the problem to the extent possible by error concealment techniques.
  • Error concealment techniques will in most cases result in degradation of output quality and are incapable of correcting some errors; further, the degree of the output error will be different depending upon the type of data in the lost packet, some of which will be more difficult to conceal than others. Thus, if packets must be discarded, some types of packets will be better candidates for discarding than others.
  • a multiplexer for broadband subscriber access comprises a memory queue for storing packets of video information and traffic management circuitry for selectively discarding one or more packets from the memory queue responsive to a detection of congestion in the memory queue based on parameters associated with existing packets currently in the queue.
  • the present invention provides significant advantages over the prior art. By selectively discarding packets from the queue, the degradation of the resulting video during times of congestion is greatly improved.
  • FIG. 1 illustrates a block diagram of a IP video delivery system
  • FIG. 2 illustrates a block diagram of a multiplexer of FIG. 1 ;
  • FIG. 3 illustrates how congestion can occur at the multiplexer because of aggregated data rates can exceed expected average aggregated data rates
  • FIG. 4 illustrates a block diagram of a traffic management system of the multiplexer of FIG. 3 ;
  • FIG. 5 illustrates timing considerations in the transmission of a packet between a source and its destination
  • FIG. 6 illustrates operation of the timing recognition function of the traffic management system of FIG. 4 ;
  • FIG. 7 illustrates operation of the timing recognition with respect to varying time periods
  • FIG. 8 through 10 illustrate operation of the timing recognition with respect to cyclic packet spacings
  • FIG. 11 illustrates a implementation for determining a best correlation of cyclic spacings
  • FIG. 12 is a graph showing simulations of predicted values of packet spacings with actual values
  • FIG. 13 a illustrates an embodiment for receiving information about upcoming packets
  • FIG. 13 b illustrates storage of virtual and actual packets
  • FIG. 14 illustrates the concept of stale packets
  • FIG. 15 illustrates a packet discard policy
  • FIG. 16 illustrates an ordered set of packets in a queue for describing an alternative discard method
  • FIG. 17 illustrates a flow chart describing the alternative discard method
  • FIG. 18 illustrates another discard method
  • FIG. 19 illustrates an example using the discard method of FIG. 18 .
  • FIG. 1 shows a block diagram of an IP video network 10 for sending video programming to a site 12 .
  • Sources such as video head ends, or VHEs
  • VHEs video head ends
  • the IP video receivers 22 translate the video packets to video for video monitors 24 .
  • the data must pass through a public/private network 26 which may include a plurality of routers, including edge router 28 .
  • the output of edge router 28 is received by multiplexer 30 (which could be, for example, a DSLAM access element), where the data for multiple video channels is placed on twisted pair lines 31 .
  • a modem 32 (such as a DSL modem) on the user site communicates between the multiplexer 30 and the IP video receivers 22 through on-site router 34 .
  • the VHE sources 20 stream video information to the IP video receivers 22 .
  • the video data is typically sent as a multicast transmission.
  • on-demand video unicast transmission may be used.
  • on-demand video generally has a longer buffer, since the delay from source 20 to viewing is not important as broadcast video servers and, thus, on-demand video has a lower priority than live broadcast video services.
  • the site 12 may have several IP video receivers 22 each receiving multiple streams of programming. For example, each IP video receiver 22 could receive two video data streams. If there were three IP video receivers 22 in the site 12 , and each receiver 22 was receiving two video streams, then the link 31 between the multiplexer 30 and the modem 32 would be carrying video packets for six different data streams.
  • Modern day video protocols compress the video stream by periodically sending a full frame (compressed) of video data, followed by differential frames which indicate the changes between frames, rather than the frame itself. Accordingly, a scene which has a rapidly changing image will require a higher bandwidth than a frame that is relatively still.
  • the total available bandwidth between the video heads 20 and the IP receivers 22 for a site 12 is generally fixed by the bandwidth of link 31 , in view of the technology used by the multiplexer 30 and modem 32 .
  • the number of data streams supported by the link 31 is determined by an average bandwidth for each received channel (link 31 can be carrying other data traffic such as Internet traffic, which has a lower priority than the live video data streams (lowest priority) and voice (VOIP—voice over Internet protocol, which generally has the highest priority).
  • link 31 can be carrying other data traffic such as Internet traffic, which has a lower priority than the live video data streams (lowest priority) and voice (VOIP—voice over Internet protocol, which generally has the highest priority).
  • VOIP voice over Internet protocol
  • FIG. 2 illustrates a block diagram of the multiplexer 30 supporting N different data streams.
  • N For a system designed to provide viewing up to two data streams over three receivers 22 , N would equal six.
  • An input stage 40 receives various video streams and sorts packets for different data streams into respective FIFO (first in, first out) memories 42 .
  • An output stage 44 multiplexes packets from the FIFO memories 42 onto the link 31 .
  • router 34 directs packets to the proper receiver 22 .
  • Traffic Management System 46 controls the multiplexing of the packets from memories 42 onto the link 31 , as described in greater detail below.
  • the congestion problem is illustrated in FIG. 3 .
  • the traffic management system 46 must make intelligent decisions about which packets to discard to minimize any adverse effects on data service to the end user.
  • data packets come from Source A and Source B.
  • Each source implements a policy to provide data at a known average rate.
  • the data from the two sources must be merged onto link 31 , which has a capacity to accommodate the combined average data rates.
  • Limited buffering is available from the FIFO memories 42 ; however, it is desirable to keep the FIFO memories as small as possible; otherwise a noticeable delay will occur when live video channels are switched.
  • the multiplexer 30 has the memory capacity to buffer additional packets, it may need to drop packets because of timing considerations associated with its FIFO 42 . For example, a multiplexer 30 may have a requirement that all packets will be sent within 200 ms of receiving that packet. If the condition cannot be met for an incoming packet, the multiplexer will either need to not place the incoming packet in the FIFO 42 or drop packets already in the FIFO 42 .
  • the multiplexer 30 is designed to minimize the effect of dropping packets.
  • a critical aspect of the problem is that all packets are time-critical. For each data stream all packets are generated on a single server (VHE 20 ). Once generated, each packet has a strict “use by” time. A packet that becomes stale in transit to the end user becomes unusable. To conserve shared link bandwidth, stale packets must be discarded without being transmitted over the link 31 .
  • multiplexer 30 conforms to a policy that requires the minimum degradation of service to the end user when packets are discarded. This goal is accomplished in two basic ways: (1) the multiplexer 30 discards the minimum amount of data necessary to avoid congestion and (2) the multiplexer 30 makes use of a priority scheme to ensure the least useful packets are preferentially discarded.
  • the multiplexer 30 can use a self-clocking mechanism that uses packet arrival times to determine packet use-by times.
  • the multiplexer 30 can use priority codes associated with each packet to determine which packets to drop first.
  • the priority scheme can also identify packet interdependencies, so that when a superior packet is dropped, packets that depend on that packet can be identified and also be dropped.
  • a look-ahead feature can identify the nature of future traffic and allow the multiplexer 30 to make better decisions about which packets to drop. For example, when knowledge of future traffic indicates that a high priority packet will arrive, the multiplexer 30 can drop packets currently in its queues to eliminate congestion in advance. Further, a traffic analysis function 30 can examine all queued traffic plus all known future traffic and determine the optimum set of packets to drop.
  • FIG. 4 illustrates a block diagram of the traffic management system 46 .
  • Traffic management system 46 includes a timing recognition function 50 , a priority recognition function 52 and a discard control function 54 .
  • the timing recognition function 50 derives information which can be used to determine the time at which a particular packet will become stale.
  • the priority recognition function determines a priority for a particular packet, either explicitly or implicitly.
  • the discard control function 54 determines which packets should be discarded.
  • FIGS. 5-12 illustrate the timing recognition function 50 .
  • FIG. 5 illustrates a video packet source that generates packets 60 and an end user that consumes the packets 60 . If a packet is generated at time T 1 it must be available to the end user at time T 2 . If the user receives the packet after T 2 the packet cannot be used.
  • the packet stream passes through the multiplexer 30 (any other packet streams are not shown in the diagram). In order to accomplish its task of merging multiple packet streams the multiplexer 30 will need to delay some of the packets 60 .
  • the end-to-end time budget for the packets includes some delay by the multiplexer 30 .
  • One of the considerations for the multiplexer design is that packets must not be delayed so much that the end-to-end schedule is compromised. Complicating the situation is jitter 61 in the packet delay timing. Jitter is represented in the diagram as a variable delay function in the packet transport.
  • the source 20 cooperates by generating (and transmitting) packets 60 according to a predictable schedule.
  • the simplest such schedule involves transmitting packets at equal time intervals, which may be agreeable with a video service that sends one packet for each video frame.
  • other transmission timing schemes will work as well. What is necessary is that the multiplexer 30 can detect the timing pattern and “lock onto” it. In this way, the packets are “self clocking.” They provide the timing signal the multiplexer 30 uses to determine the appropriate time schedule for each packet.
  • the operation of the timing recognition function 50 is illustrated in FIG. 6 .
  • This diagram illustrates the case of packets equally spaced in time. Packets from the source 20 arrive at the multiplexer 30 after a variable delay due to the jitter function.
  • the timing recognition function 50 implements a synchronizer function based on the assumption packets never arrive early. When a packet 60 arrives the synchronizer starts a timer. When a subsequent packet arrives the time between packets is noted.
  • the timing schedule is assumed from packet A.
  • Packet B arrives behind schedule and is assumed to be delayed. The schedule is not changed. Packet C arrives ahead of schedule, and the schedule is shifted backwards, as shown in the lower part of the diagram.
  • the multiplexer 30 adds its time budget to the arrival schedule to compute the stale time for the packets.
  • a refinement to the time synchronizer function allows it to accommodate a varying time period.
  • a varying time period would be symptomatic of clock drift between the multiplexer 30 and the packet source 20 .
  • the packet delay is bounded and maintains an average value.
  • the packet delay (e 1 , e 2 . . . ) average increases with time.
  • the synchronizer function detects the increasing delay and accommodates it by increasing the time period p of the expected packets.
  • the synchronizing function computes a running average of the packet delay over a fixed number of the most recent delay times.
  • p i is the projected period for computing the schedule for the next packet.
  • is the measured time between the two most recent packets, and 0 ⁇ 1. When ⁇ is small (close to zero) the period will be adjusted to accommodate a slowly varying packet rate.
  • the multiplexer 30 will synchronize with the packet stream, provided the packet spacing is cyclic or otherwise predictable.
  • synchronization can be accomplished by pattern matching, as shown in FIG. 8 . Bars in the top row represent the arrival of packets at the multiplexer, and the spacing between packets is represented by the height of the vertical bars in the second row. The bottom row represents the expected pattern of packet spacings. Spacings between packets are recorded and matched against the expected pattern.
  • the multiplexer will need to assess the match (correlation) between the expected pattern and the actual pattern and determine the point of best correlation.
  • the method for detecting the traffic pattern involves comparing inter-packet spacings. In the simplest method, the current inter-packet spacing is recorded and compared with subsequent inter-packet spacings. This operation is performed with differing offsets between the spacings used in the comparison and is illustrated in FIGS. 9 and 10 .
  • FIG. 9 illustrates various timing offsets from a given packet.
  • FIG. 10 illustrates finding an offset that correlates to the most likely period of a pattern of spacings. The offset that produces the best correlation in terms of absolute differences ( ⁇ in FIG. 10 ) is the period of the pattern.
  • FIG. 11 A possible implementation scheme for determining ⁇ is shown in FIG. 11 .
  • the simulation compared predicted values of the packet spacing with actual values and plotted the correlation (absolute value of the difference).
  • the plots in FIG. 12 are from a simulation that used 20 circular buffers. Plots for the correlation function for lengths of 1, 2, 4, 8, and 10 are included. The actual packet sequence length was 10, and the plot shows the correlation function for that length converged, while the correlation for the other lengths did not.
  • the plots represent every 10 packets, so it is apparent the 10-length pattern became discernible from the other patterns after about 30 packets.
  • the discard control function 54 of the traffic management system 46 of multiplexer 30 uses priority information associated with the packets to assist in determining which packets to drop. This information is obtained from the priority detection function 52 .
  • the packet source provides a priority indicator to communicate dropping preferences to the multiplexer. Making use of several priority levels, the dropping priorities could be coded according to the following scheme:
  • the discard control function 54 bases its packet dropping decisions on information about packets in its queues 42 . It is desired to minimize the number of packets queued at the multiplexer for several reasons:
  • FIG. 13 a illustrates the concept.
  • the diagram in the top of the figure shows queued packets 60 q actually stored in one of the queues 42 in the multiplexer 30 and unqueued packets 60 u that have not yet arrived.
  • the multiplexer 30 has information about the queued packets 60 q , but without information about the unqueued packets 60 u , it cannot take them into account in determining which packets to drop.
  • the virtual packets 60 h have been stored in a queue 42 in the multiplexer 30 .
  • Virtual packets are “header packets” that contain information about the actual packets, but not the payload.
  • a header packet 60 h is sent in advance of each data packet (including payload), and the multiplexer 30 queues the header packets, both header packets 60 q for data packets already at the multiplexer 30 and header packets 60 h for data packets that have not arrived.
  • the discard control function 54 is now able to compute a dropping solution as though all packets, actual and virtual, were queued at the multiplexer. Since the header packets 60 h are much smaller than the data packets, more of them can be queued without seriously impacting memory requirements at the multiplexer 30 .
  • header packets 60 h does not alleviate the problem of time delay. Since the header packets 60 h must be sent in advance of the data packets, the data packets are effectively delayed at the source past the time they could have been transmitted.
  • the packet discard function 54 will minimize data loss by observing priorities related to image dependence and packet age.
  • a primary rule is that shared link bandwidth will not be wasted by transmitting stale data.
  • the secondary rule is that when packets are dropped in accordance with the primary rule, impact on video quality will be minimized.
  • the first rule is enforced by identifying packets with stale data and dropping those packets.
  • packets 60 near the head of the queue can be transmitted before their data become stale. However, some of the data in packet a will become stale before it can be transmitted, because packets ahead of it in the queue are large and will take a long time to transmit. If some packets ahead of packet a in the queue are dropped, then packet x can be transmitted before its data become stale. The fact that packet x will become stale (unless other packets are dropped) does not imply packet y will become stale. This is because the schedule of packet y may allow it to be delivered much later.
  • the discard control function 54 may implement the packet dropping policy through the use of a packet database 70 , as illustrated in FIG. 15 .
  • the database 70 will record corresponding entries 72 of pertinent information required by the traffic management system 46 :
  • Packet priority can be obtained directly from packet header information, or it can be inferred by the size of the packet and/or by the position of the packet in a cyclic traffic pattern.
  • the packet delivery schedule is the time when the last data from the packet must be transmitted through the shared link.
  • the delivery schedule is inferred from the packet arrival time and its position within a cyclic traffic pattern.
  • the discard control function 54 determines whether a packet 60 will become stale by computing the transmission times for the data in the packet 60 and for the data ahead of the packet in its queue 42 . To the first approximation the queue time of a packet uses the assumption the queue will be apportioned a certain amount of shared link bandwidth and that the queue will be served on an equitable basis.
  • a preferred embodiment for discarding packets is:
  • the invention described in connection with FIGS. 1 through 15 provides significant advantages over the prior art in maximizing video quality or other packet services to an end user.
  • Second, the invention may compute packet delivery schedules (used to identify stale packets) from packet sequence and packet arrival times without relying on packet internal information.
  • Third, the invention may identify packet sequence patterns and synchronize the packet schedule computations with the arriving packet stream.
  • the invention may infer packet priority from packet size or from a packet's position in a sequence without relying on packet internal information.
  • Fifth, the invention may infer packet interdependencies from packet priority sequences and drop additional packets according to their dependencies.
  • the invention can use out-of-band header packets or predictions of future packet traffic based on observed traffic patterns to provide advance information about future traffic.
  • the invention can use that information in addition to currently queued packets to perform its traffic management computations. This minimizes the number of packets that need to be queued.
  • the invention can minimize loss of quality data by making an early identification of potentially stale, high priority packets. It works to preserve high priority packets by dropping lower priority packets.
  • the invention performs its task without undue coupling with the packet source. This minimizes the impact of any changes in the design of the packet source.
  • FIGS. 16 and 17 illustrate an alternative method of determining which packets to discard.
  • FIG. 16 illustrates a sequence of packets 60 in a queue 42 .
  • the earliest packet is number “1” and the latest packet is number “6”.
  • FIG. 17 is a flow chart describing operation of a discard process for discarding packets 60 in the queues 42 .
  • step 70 starting with packet “1”, a packet n is marked for deletion if it cannot be transmitted over link 31 in a time which is within its “time-to-live” (TTL(n)). Index n is incremented until all such packets are discarded.
  • TTL(n) time-to-live
  • steps 70 and 72 is to remove all packets 60 from the queues 42 if the packet cannot be received at the site 12 in time for use by an IP receiver 22 .
  • a total transit time is determined for packet n, taking in consideration the non-discarded packets ahead of packet n (i.e., older packets across all queues 42 ) is computed; this time is denoted TTT(n).
  • TTT(n)>TTL(n) is not met, i.e., if there will be time to transmit packet n after transmitting all preceding packets, then n is incremented for evaluation of the next packet in step 78 . Otherwise, steps are taken to discard lower priority packets ahead of packet n.
  • the total transit time of packets which precede packet n but have a lower priority is determined. This value is denoted TTTlp(n).
  • step 82 If the condition TTT(n) ⁇ TTTlp(n)>TTL(n) is not met in step 82 , i.e., if packet n will not be transmitted in time even if all lower priority packets are discarded, then packet n is marked for discard and index n is incremented in step 84 .
  • step 82 If the condition TTT(n) ⁇ TTTlp(n)>TTL(n) is met in step 82 , i.e., packet n can be transmitted in time by discarding one or more of the preceding lower priority packets, then a minimum set of packets is chosen in step 86 by discarding packets in order of increasing size starting at the smallest. This could be done as shown in 17 by discarding the smallest preceding non-discarded packet, and recomputing in step 80 , or by determining the minimum time needed for transmitting packet n on time, and summing the transmit times for the preceding packets in order of size, starting at the smallest packet, until the minimum time is reached.
  • Discarded packets are simply not transmitted by the output stage 44 once they reach the front of the queue 42 .
  • the output stage 44 examines each queue 42 in turn in cyclic order using a pointer to determine which queue has the oldest (non-discarded) packet a the head of the queue. If two or more packets have the “oldest” time, then the output stage 44 serves the first queue encountered with the oldest packet. The pointer is advanced one position past the queue served, and the cycle is repeated.
  • the output stage 44 uses all the FIFOs 42 to create a much larger FIFO.
  • This embodiment of the invention works to maximize video quality or other packet service to an end user with a number of advantages.
  • FIGS. 18 and 19 illustrates another embodiment for operation of the discard control function 54 .
  • packets are designated in terms of priority, either explicitly by marking at the source 20 or implicitly as described above.
  • Critical video packets would be marked with a high priority indicating a need for a high probability for delivery to the decoder. This priority would be associated with I and P Frames generated using H.264 video compression as an example.
  • Less important packets, that if lost could potentially be concealed by the decoder or would have minimum impact on video display quality, would be marked with a lower priority indicating the need for a lower probability for delivery to the decoder.
  • This lower priority would be associated with Bs and B frames generated using H.264 video compression as an example.
  • discard control function 54 would first select low priority packets to be dropped to achieve sufficient bandwidth requirement reduction to ensure that higher priority packets would have bandwidth to be transmitted toward the subscriber.
  • Low priority packets associated with all video streams being multiplexed for a single access subscriber line would be candidates for the discardable packet selection process in step 90 .
  • the idea is that it would be better to discard a low priority frame from-another queue within the multiplex group than to discard a higher priority packet from the video stream that was causing the overload condition.
  • the smallest number of low priority packets could be selected similar to step 86 of FIG. 17 .
  • FIG. 19 represents a traffic management system with queuing and scheduling functions that accommodates multiple video packet streams that are composed of both high priority video data packets and low priority video data packets. Also shown is the existence of a new high priority packet (packet 1 f ) that has arrived at the traffic manager. Packet 1 f needs to enter Channel Queue 1 and be served by the output stage 44 before it ages pasts a delay period that would cause the packet to arrive at the subscribers video decoder too late to be decoded and displayed. Since this new packet is High Priority, it has to be allowed to enter queue 1 .
  • the selection process would select which packet or packets to discard, following the above stated rules for this example, in the following order: 2a, na, 1b, nb, 1c, 2c, 2d, nd, 1e, 1a, 1d.
  • the packet dropping process would stop when sufficient bandwidth is gained to be able to transmit the new video packet within the specified maximum age for video packets within the traffic manager.
  • This method for discarding packets delivers a high level of video display quality to the video viewer, with minimum processing.

Abstract

A IP video delivery system (10) includes a multiplexer (30) for transmitting multiple data streams of packets over link (31) to a site having one or more receivers (22). During times of congestion, the multiplexer will discard packets from its internal queues. Packets are intelligently chosen to minimize the effects on the output from the receivers (22) by taking into account timing information related to the packets and priority.

Description

    STATEMENT OF FEDERALLY SPONSORED RESEARCH OR DEVELOPMENT
  • The U.S. Government has a paid-up license in this invention and the right in limited circumstances to require the patent owner to license others on reasonable terms as provided for by the terms of Award No. 70NANB3H3053 awarded by National Institute of Standards and Technology.
  • BACKGROUND OF THE INVENTION
  • 1. Technical Field
  • This invention relates in general to network communications and, more particularly, to a method and apparatus for discarding packets.
  • 2. Description of the Related Art
  • In a digital information delivery network, between a source device and a destination device, packets of data may be lost for a variety of reasons. Some packets are randomly lost due to uncontrollable errors - for example, errors caused by noise on a transmission line, synchronization-issues, etc.-Some-packets are lost due to congestion, i.e., it is not possible for a network element to transmit all received packets in a timely manner. Current discard mechanisms for IP QoS (quality of service) algorithms implement random selection schemes to determine which packets to discard without regard to the relative effect on the eventual output.
  • For some data transfer protocols, missing packets cause the destination device to request a retransmission of the missing information. This is not very feasible, however, in a network that has multicasting of real-time streams such as audio or video. Normally, there will not be enough time available for requesting and receiving the retransmitted packets, unless buffers at the destination device are very large.
  • When an expected packet in a packet stream is not received at the destination device, the destination device waits for a certain amount of time before declaring a packet as lost. Once a packet is declared as lost, some decoders may request retransmission, other decoders may correct the problem to the extent possible by error concealment techniques. Error concealment techniques will in most cases result in degradation of output quality and are incapable of correcting some errors; further, the degree of the output error will be different depending upon the type of data in the lost packet, some of which will be more difficult to conceal than others. Thus, if packets must be discarded, some types of packets will be better candidates for discarding than others.
  • Accordingly, there is a need for a method and apparatus for identifying and discarding packets to minimize output errors.
  • BRIEF SUMMARY OF THE INVENTION
  • In the present invention, a multiplexer for broadband subscriber access comprises a memory queue for storing packets of video information and traffic management circuitry for selectively discarding one or more packets from the memory queue responsive to a detection of congestion in the memory queue based on parameters associated with existing packets currently in the queue.
  • The present invention provides significant advantages over the prior art. By selectively discarding packets from the queue, the degradation of the resulting video during times of congestion is greatly improved.
  • BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS
  • For a more complete understanding of the present invention, and the advantages thereof, reference is now made to the following descriptions taken in conjunction with the accompanying drawings, in which:
  • FIG. 1 illustrates a block diagram of a IP video delivery system;
  • FIG. 2 illustrates a block diagram of a multiplexer of FIG. 1;
  • FIG. 3 illustrates how congestion can occur at the multiplexer because of aggregated data rates can exceed expected average aggregated data rates;
  • FIG. 4 illustrates a block diagram of a traffic management system of the multiplexer of FIG. 3;
  • FIG. 5 illustrates timing considerations in the transmission of a packet between a source and its destination;
  • FIG. 6 illustrates operation of the timing recognition function of the traffic management system of FIG. 4;
  • FIG. 7 illustrates operation of the timing recognition with respect to varying time periods;
  • FIG. 8 through 10 illustrate operation of the timing recognition with respect to cyclic packet spacings;
  • FIG. 11 illustrates a implementation for determining a best correlation of cyclic spacings;.
  • FIG. 12 is a graph showing simulations of predicted values of packet spacings with actual values;
  • FIG. 13 a illustrates an embodiment for receiving information about upcoming packets;
  • FIG. 13 b illustrates storage of virtual and actual packets;
  • FIG. 14 illustrates the concept of stale packets;
  • FIG. 15 illustrates a packet discard policy;
  • FIG. 16 illustrates an ordered set of packets in a queue for describing an alternative discard method;
  • FIG. 17 illustrates a flow chart describing the alternative discard method;
  • FIG. 18 illustrates another discard method; and
  • FIG. 19 illustrates an example using the discard method of FIG. 18.
  • DETAILED DESCRIPTION OF THE INVENTION
  • The present invention is best understood in relation to FIGS. 1-19 of the drawings, like numerals being used for like elements of the various drawings.
  • FIG. 1 shows a block diagram of an IP video network 10 for sending video programming to a site 12. Sources (such as video head ends, or VHEs) 20 provide the programming by streaming video information in packets. The packets are ultimately received by one or more IP video receivers 22 at the site 12. The IP video receivers 22 translate the video packets to video for video monitors 24. To get to the IP video receivers 22, the data must pass through a public/private network 26 which may include a plurality of routers, including edge router 28. The output of edge router 28 is received by multiplexer 30 (which could be, for example, a DSLAM access element), where the data for multiple video channels is placed on twisted pair lines 31. A modem 32 (such as a DSL modem) on the user site communicates between the multiplexer 30 and the IP video receivers 22 through on-site router 34.
  • In operation, the VHE sources 20 stream video information to the IP video receivers 22. For live video broadcasts, such as a live television signal, the video data is typically sent as a multicast transmission. For on-demand video, unicast transmission may be used. At the receiver side, on-demand video generally has a longer buffer, since the delay from source 20 to viewing is not important as broadcast video servers and, thus, on-demand video has a lower priority than live broadcast video services. The site 12 may have several IP video receivers 22 each receiving multiple streams of programming. For example, each IP video receiver 22 could receive two video data streams. If there were three IP video receivers 22 in the site 12, and each receiver 22 was receiving two video streams, then the link 31 between the multiplexer 30 and the modem 32 would be carrying video packets for six different data streams.
  • Modern day video protocols compress the video stream by periodically sending a full frame (compressed) of video data, followed by differential frames which indicate the changes between frames, rather than the frame itself. Accordingly, a scene which has a rapidly changing image will require a higher bandwidth than a frame that is relatively still. The total available bandwidth between the video heads 20 and the IP receivers 22 for a site 12 is generally fixed by the bandwidth of link 31, in view of the technology used by the multiplexer 30 and modem 32.
  • With a fixed bandwidth in which to transfer all packets for all data streams for a site 12, the number of data streams supported by the link 31 is determined by an average bandwidth for each received channel (link 31 can be carrying other data traffic such as Internet traffic, which has a lower priority than the live video data streams (lowest priority) and voice (VOIP—voice over Internet protocol, which generally has the highest priority). However, the data rates for the separate N data flows are not constant. At times, multiple channels may be simultaneously using more than their average bandwidth, resulting in congestion on link 31.
  • FIG. 2 illustrates a block diagram of the multiplexer 30 supporting N different data streams. For a system designed to provide viewing up to two data streams over three receivers 22, N would equal six. An input stage 40 receives various video streams and sorts packets for different data streams into respective FIFO (first in, first out) memories 42. An output stage 44 multiplexes packets from the FIFO memories 42 onto the link 31. At the site 12, router 34 directs packets to the proper receiver 22. Traffic Management System 46 controls the multiplexing of the packets from memories 42 onto the link 31, as described in greater detail below.
  • The congestion problem is illustrated in FIG. 3. When the combined data rates from the N sources exceed the capacity of the link 31 and the capacity of the multiplexer 30 to buffer the overage in its FIFO memories 42, the traffic management system 46 must make intelligent decisions about which packets to discard to minimize any adverse effects on data service to the end user. The situation is illustrated in FIG. 3, which employs only two data sources (N=2). In FIG. 3, data packets come from Source A and Source B. Each source implements a policy to provide data at a known average rate. The data from the two sources must be merged onto link 31, which has a capacity to accommodate the combined average data rates. Limited buffering is available from the FIFO memories 42; however, it is desirable to keep the FIFO memories as small as possible; otherwise a noticeable delay will occur when live video channels are switched. When the combined data rates from the sources exceeds the average for too long, the capacity to buffer the excess data is exceeded, and some packets must be dropped. Even if the multiplexer 30 has the memory capacity to buffer additional packets, it may need to drop packets because of timing considerations associated with its FIFO 42. For example, a multiplexer 30 may have a requirement that all packets will be sent within 200 ms of receiving that packet. If the condition cannot be met for an incoming packet, the multiplexer will either need to not place the incoming packet in the FIFO 42or drop packets already in the FIFO 42.
  • In operation, the multiplexer 30 is designed to minimize the effect of dropping packets. A critical aspect of the problem is that all packets are time-critical. For each data stream all packets are generated on a single server (VHE 20). Once generated, each packet has a strict “use by” time. A packet that becomes stale in transit to the end user becomes unusable. To conserve shared link bandwidth, stale packets must be discarded without being transmitted over the link 31.
  • In operation, multiplexer 30 conforms to a policy that requires the minimum degradation of service to the end user when packets are discarded. This goal is accomplished in two basic ways: (1) the multiplexer 30 discards the minimum amount of data necessary to avoid congestion and (2) the multiplexer 30 makes use of a priority scheme to ensure the least useful packets are preferentially discarded.
  • To accomplish these goals, four mechanisms may be used. First, in the absence of explicit time schedule information associated with the packets, the multiplexer 30 can use a self-clocking mechanism that uses packet arrival times to determine packet use-by times. The multiplexer 30 can use priority codes associated with each packet to determine which packets to drop first. The priority scheme can also identify packet interdependencies, so that when a superior packet is dropped, packets that depend on that packet can be identified and also be dropped.
  • A look-ahead feature can identify the nature of future traffic and allow the multiplexer 30 to make better decisions about which packets to drop. For example, when knowledge of future traffic indicates that a high priority packet will arrive, the multiplexer 30 can drop packets currently in its queues to eliminate congestion in advance. Further, a traffic analysis function 30 can examine all queued traffic plus all known future traffic and determine the optimum set of packets to drop.
  • FIG. 4 illustrates a block diagram of the traffic management system 46. Traffic management system 46 includes a timing recognition function 50, a priority recognition function 52 and a discard control function 54. The timing recognition function 50 derives information which can be used to determine the time at which a particular packet will become stale. The priority recognition function determines a priority for a particular packet, either explicitly or implicitly. The discard control function 54 determines which packets should be discarded. These functions are described in greater detail below. These functions could be implemented on a general purpose processor, DSP (digital signal processor) or other processing device.
  • FIGS. 5-12 illustrate the timing recognition function 50. FIG. 5 illustrates a video packet source that generates packets 60 and an end user that consumes the packets 60. If a packet is generated at time T1 it must be available to the end user at time T2. If the user receives the packet after T2 the packet cannot be used. The packet stream passes through the multiplexer 30 (any other packet streams are not shown in the diagram). In order to accomplish its task of merging multiple packet streams the multiplexer 30 will need to delay some of the packets 60. The end-to-end time budget for the packets includes some delay by the multiplexer 30. One of the considerations for the multiplexer design is that packets must not be delayed so much that the end-to-end schedule is compromised. Complicating the situation is jitter 61 in the packet delay timing. Jitter is represented in the diagram as a variable delay function in the packet transport.
  • To enable the multiplexer 30 to discern the end-to-end schedule, the source 20 cooperates by generating (and transmitting) packets 60 according to a predictable schedule. The simplest such schedule involves transmitting packets at equal time intervals, which may be agreeable with a video service that sends one packet for each video frame. However, other transmission timing schemes will work as well. What is necessary is that the multiplexer 30 can detect the timing pattern and “lock onto” it. In this way, the packets are “self clocking.” They provide the timing signal the multiplexer 30 uses to determine the appropriate time schedule for each packet.
  • The operation of the timing recognition function 50 is illustrated in FIG. 6. This diagram illustrates the case of packets equally spaced in time. Packets from the source 20 arrive at the multiplexer 30 after a variable delay due to the jitter function. The timing recognition function 50 implements a synchronizer function based on the assumption packets never arrive early. When a packet 60 arrives the synchronizer starts a timer. When a subsequent packet arrives the time between packets is noted. In the upper part of the diagram the timing schedule is assumed from packet A. Packet B arrives behind schedule and is assumed to be delayed. The schedule is not changed. Packet C arrives ahead of schedule, and the schedule is shifted backwards, as shown in the lower part of the diagram. Once the schedule has been synchronized with the packet stream the multiplexer 30 adds its time budget to the arrival schedule to compute the stale time for the packets.
  • A refinement to the time synchronizer function allows it to accommodate a varying time period. A varying time period would be symptomatic of clock drift between the multiplexer 30 and the packet source 20. In the top part of FIG. 7 the packet delay is bounded and maintains an average value. In the bottom part of the diagram the packet delay (e1, e2 . . . ) average increases with time. The synchronizer function detects the increasing delay and accommodates it by increasing the time period p of the expected packets.
  • To accommodate a varying packet rate, the synchronizing function computes a running average of the packet delay over a fixed number of the most recent delay times. The packet schedule will be adjusted using the following formula:
    p i+1=(1−α)p i+αΔ
  • In this formula pi is the projected period for computing the schedule for the next packet. Δ is the measured time between the two most recent packets, and 0<α<1. When α is small (close to zero) the period will be adjusted to accommodate a slowly varying packet rate.
  • Even when the packet spacing is not uniform, the multiplexer 30 will synchronize with the packet stream, provided the packet spacing is cyclic or otherwise predictable. When the packet spacing is cyclic and known in advance, synchronization can be accomplished by pattern matching, as shown in FIG. 8. Bars in the top row represent the arrival of packets at the multiplexer, and the spacing between packets is represented by the height of the vertical bars in the second row. The bottom row represents the expected pattern of packet spacings. Spacings between packets are recorded and matched against the expected pattern.
  • Since the incoming packets are subject to delay jitter, there will not be a perfect match between the expected pattern and the actual pattern. The multiplexer will need to assess the match (correlation) between the expected pattern and the actual pattern and determine the point of best correlation. The correlation between the expected pattern and the actual pattern is computed by the following formula: c = a A i - E i
  • This is the sum of the absolute values of the differences between the actual (A) and expected (E) values over the region α. The region for which the sum is minimum is the best match. When the spacing pattern is not known in advance, the multiplexer 30 will detect it. The method for detecting the traffic pattern involves comparing inter-packet spacings. In the simplest method, the current inter-packet spacing is recorded and compared with subsequent inter-packet spacings. This operation is performed with differing offsets between the spacings used in the comparison and is illustrated in FIGS. 9 and 10. FIG. 9 illustrates various timing offsets from a given packet. FIG. 10 illustrates finding an offset that correlates to the most likely period of a pattern of spacings. The offset that produces the best correlation in terms of absolute differences (δ in FIG. 10) is the period of the pattern. By recording the pattern and using it to match with future traffic, the multiplexer can synchronize with the packet arrival times.
  • A possible implementation scheme for determining δ is shown in FIG. 11. As packets arrive the packet spacing is measured, and the values are entered into multiple circular buffers. There is one buffer for each possible sequence length and each buffer has its own pointer, which is advanced after each packet. Packet spacing values are added to the buffers using the following formula:
    δi+1=(1−α)δi+αΔ
  • In the formula 0<α<1. This causes each δ to converge to the average of the most recent values of Δ. When the length of a circular buffer equals the length of the cyclic packet sequence, the values of the δs will converge to the Δs in the packet spacing sequence. If the jitter function is well-behaved, the averages of the Δs will be the specified packet spacing values, and the δs will be an accurate representation of the spacing pattern. When the length of a buffer does not match the length of the packet sequence, the value of the δs will tend to converge to the average packet spacing over the whole sequence.
  • This scheme has been tested by simulation. The simulation compared predicted values of the packet spacing with actual values and plotted the correlation (absolute value of the difference). The plots in FIG. 12 are from a simulation that used 20 circular buffers. Plots for the correlation function for lengths of 1, 2, 4, 8, and 10 are included. The actual packet sequence length was 10, and the plot shows the correlation function for that length converged, while the correlation for the other lengths did not. The convergence of the correlation function involved a low band pass filter (α=0.1), which resulted in a slow convergence. The plots represent every 10 packets, so it is apparent the 10-length pattern became discernible from the other patterns after about 30 packets.
  • To compensate for clock drift with cyclic traffic patterns, the previous formula for adjusting the expected spacing is modified. Instead of a single spacing p the formula now uses a set of spacings pj, where 2≦j≦P, and P is the period of the cycle. The formula becomes the following: ( p j ) i + 1 = ( 1 - α ) ( p j ) i + αΔ
  • The discard control function 54 of the traffic management system 46 of multiplexer 30 uses priority information associated with the packets to assist in determining which packets to drop. This information is obtained from the priority detection function 52. In a first embodiment of the priority detection function, the packet source provides a priority indicator to communicate dropping preferences to the multiplexer. Making use of several priority levels, the dropping priorities could be coded according to the following scheme:
      • 1. Priority 0 is the highest priority. No priority 0 packet will be dropped unless there are no other packets to drop.
      • 2. If the multiplexer encounters a sequence of packet priorities of the form a a . . . a, if one of the packets is dropped, all such packets in the sequence must be dropped.
      • 3. A sequence of the form a a . . . b . . . a a, where a=b+1(i.e., a is lower priority than b), indicates the a packets are dependent on the b packet. If the b packet is dropped, then the a packets must be dropped as well.
      • 4. A sequence of the form b c d . . . e, where b<c<d, and e<d indicates that dropping the b packet requires that c, d . . . packets be dropped, but it does not require that the e packet be dropped.
      • 5. Similar rules can be encoded in the packet priority scheme in order to extend the logic to more complex dropping policies.
  • The discard control function 54 bases its packet dropping decisions on information about packets in its queues 42. It is desired to minimize the number of packets queued at the multiplexer for several reasons:
      • 1. Queuing packets requires memory hardware.
      • 2. Longer queues represent more delay.
      • 3. Longer queues represent more packets that have to be flushed when one packet stream is stopped and replaced by another (as when a different channel is selected by a video customer).
  • Some of these difficulties can be alleviated by queuing information about packets without actually queuing the packets. FIG. 13 a illustrates the concept. The diagram in the top of the figure shows queued packets 60 q actually stored in one of the queues 42 in the multiplexer 30 and unqueued packets 60 u that have not yet arrived. The multiplexer 30 has information about the queued packets 60 q, but without information about the unqueued packets 60 u, it cannot take them into account in determining which packets to drop.
  • In FIG. 13 b, the virtual packets 60 h have been stored in a queue 42 in the multiplexer 30. Virtual packets are “header packets” that contain information about the actual packets, but not the payload. A header packet 60 h is sent in advance of each data packet (including payload), and the multiplexer 30 queues the header packets, both header packets 60 q for data packets already at the multiplexer 30 and header packets 60 h for data packets that have not arrived. The discard control function 54 is now able to compute a dropping solution as though all packets, actual and virtual, were queued at the multiplexer. Since the header packets 60h are much smaller than the data packets, more of them can be queued without seriously impacting memory requirements at the multiplexer 30.
  • The use of header packets 60 h does not alleviate the problem of time delay. Since the header packets 60 h must be sent in advance of the data packets, the data packets are effectively delayed at the source past the time they could have been transmitted.
  • The packet discard function 54 will minimize data loss by observing priorities related to image dependence and packet age. A primary rule is that shared link bandwidth will not be wasted by transmitting stale data. The secondary rule is that when packets are dropped in accordance with the primary rule, impact on video quality will be minimized.
  • The first rule is enforced by identifying packets with stale data and dropping those packets. In FIG. 14, packets 60 near the head of the queue can be transmitted before their data become stale. However, some of the data in packet a will become stale before it can be transmitted, because packets ahead of it in the queue are large and will take a long time to transmit. If some packets ahead of packet a in the queue are dropped, then packet x can be transmitted before its data become stale. The fact that packet x will become stale (unless other packets are dropped) does not imply packet y will become stale. This is because the schedule of packet y may allow it to be delivered much later.
  • The discard control function 54 may implement the packet dropping policy through the use of a packet database 70, as illustrated in FIG. 15. When packets 60 arrive at the multiplexer the database 70 will record corresponding entries 72 of pertinent information required by the traffic management system 46:
      • 1. Size of the packet.
      • 2. Packet priority.
      • 3. Delivery schedule.
  • Packet priority can be obtained directly from packet header information, or it can be inferred by the size of the packet and/or by the position of the packet in a cyclic traffic pattern. The packet delivery schedule is the time when the last data from the packet must be transmitted through the shared link. The delivery schedule is inferred from the packet arrival time and its position within a cyclic traffic pattern. When the decision is made to drop a packet it is identified in the database.
  • The discard control function 54 determines whether a packet 60 will become stale by computing the transmission times for the data in the packet 60 and for the data ahead of the packet in its queue 42. To the first approximation the queue time of a packet uses the assumption the queue will be apportioned a certain amount of shared link bandwidth and that the queue will be served on an equitable basis. A preferred embodiment for discarding packets is:
      • 1. Remove packets only from the front of the queues.
      • 2. First remove the front of the queues all packets that are marked for dropping.
      • 3. Examine the remainder of the packets in each queue and determine which packets will become stale by computing their delivery schedule. In computing the expected delivery times, ignore all packets that have been marked
      • 4. For each high priority packet that will become stale, examine lower priority packets in front of it and determine if the high priority packet can be salvaged by dropping the lower priority packet. If this can be accomplished, mark the lower priority packet or packets for dropping. In marking packets for dropping, drop packets closest to the front of the queue first and observe packet dependencies. That is, drop dependent packets as necessary. If a high priority packet cannot be salvaged by dropping lower priority packets, then mark the high priority packet for dropping.
      • 5. Repeat step 4 until there are no more high priority packets that will become stale.
      • 6. From the front of one of the queues, take the packet that has the earliest delivery schedule. In performing this step, observe any established policy for ensuring fair service for the queues. This will allow serving a packet with a slightly later delivery schedule in some cases.
  • The invention described in connection with FIGS. 1 through 15 provides significant advantages over the prior art in maximizing video quality or other packet services to an end user. First, it identifies potentially stale packets in advance and preserves service quality by discarding low priority packets in order to preserve high priority packets. Second, the invention may compute packet delivery schedules (used to identify stale packets) from packet sequence and packet arrival times without relying on packet internal information. Third, the invention may identify packet sequence patterns and synchronize the packet schedule computations with the arriving packet stream. Fourth, the invention may infer packet priority from packet size or from a packet's position in a sequence without relying on packet internal information. Fifth, the invention may infer packet interdependencies from packet priority sequences and drop additional packets according to their dependencies. Sixth, the invention can use out-of-band header packets or predictions of future packet traffic based on observed traffic patterns to provide advance information about future traffic. Sixth, the invention can use that information in addition to currently queued packets to perform its traffic management computations. This minimizes the number of packets that need to be queued. Seventh, the invention can minimize loss of quality data by making an early identification of potentially stale, high priority packets. It works to preserve high priority packets by dropping lower priority packets. Eighth, the invention performs its task without undue coupling with the packet source. This minimizes the impact of any changes in the design of the packet source.
  • FIGS. 16 and 17 illustrate an alternative method of determining which packets to discard. For reference, FIG. 16 illustrates a sequence of packets 60 in a queue 42. The earliest packet is number “1” and the latest packet is number “6”.
  • FIG. 17 is a flow chart describing operation of a discard process for discarding packets 60 in the queues 42. In step 70, starting with packet “1”, a packet n is marked for deletion if it cannot be transmitted over link 31 in a time which is within its “time-to-live” (TTL(n)). Index n is incremented until all such packets are discarded. The purpose of steps 70 and 72 is to remove all packets 60 from the queues 42 if the packet cannot be received at the site 12 in time for use by an IP receiver 22.
  • In step 74, a total transit time is determined for packet n, taking in consideration the non-discarded packets ahead of packet n (i.e., older packets across all queues 42) is computed; this time is denoted TTT(n). In step 76, if the condition TTT(n)>TTL(n) is not met, i.e., if there will be time to transmit packet n after transmitting all preceding packets, then n is incremented for evaluation of the next packet in step 78. Otherwise, steps are taken to discard lower priority packets ahead of packet n. In step 80, the total transit time of packets which precede packet n but have a lower priority is determined. This value is denoted TTTlp(n). If the condition TTT(n)−TTTlp(n)>TTL(n) is not met in step 82, i.e., if packet n will not be transmitted in time even if all lower priority packets are discarded, then packet n is marked for discard and index n is incremented in step 84.
  • If the condition TTT(n)−TTTlp(n)>TTL(n) is met in step 82, i.e., packet n can be transmitted in time by discarding one or more of the preceding lower priority packets, then a minimum set of packets is chosen in step 86 by discarding packets in order of increasing size starting at the smallest. This could be done as shown in 17 by discarding the smallest preceding non-discarded packet, and recomputing in step 80, or by determining the minimum time needed for transmitting packet n on time, and summing the transmit times for the preceding packets in order of size, starting at the smallest packet, until the minimum time is reached.
  • Discarded packets are simply not transmitted by the output stage 44 once they reach the front of the queue 42. For selecting from the various queues 42, the output stage 44 examines each queue 42 in turn in cyclic order using a pointer to determine which queue has the oldest (non-discarded) packet a the head of the queue. If two or more packets have the “oldest” time, then the output stage 44 serves the first queue encountered with the oldest packet. The pointer is advanced one position past the queue served, and the cycle is repeated. The output stage 44, in essence, uses all the FIFOs 42 to create a much larger FIFO.
  • This embodiment of the invention works to maximize video quality or other packet service to an end user with a number of advantages. First, it ensures that useless packets (packets that cannot be transmitted prior to their “use by” time) are not transmitted through the shared link 31. Second, it minimizes the loss of video quality by preferentially dropping low priority packets and the smallest of the low priority packets. Third, it ensures that a high priority packet is not dropped due to being behind low priority packets in the queue. Fourth, it ensures that the oldest packet in all the queues 42 is preferentially transmitted. Fifth, it ensures all queues 42 are served fairly by providing serving the least recently served queue when all other factors are equal.
  • FIGS. 18 and 19 illustrates another embodiment for operation of the discard control function 54. In this embodiment for a multi-queue multiplexer 30, it is assumed that packets are designated in terms of priority, either explicitly by marking at the source 20 or implicitly as described above. Critical video packets would be marked with a high priority indicating a need for a high probability for delivery to the decoder. This priority would be associated with I and P Frames generated using H.264 video compression as an example. Less important packets, that if lost could potentially be concealed by the decoder or would have minimum impact on video display quality, would be marked with a lower priority indicating the need for a lower probability for delivery to the decoder. This lower priority would be associated with Bs and B frames generated using H.264 video compression as an example.
  • During periods of congestion, i.e., when the traffic management system determines that a critical priority packet will not be transmitted in time, discard control function 54 would first select low priority packets to be dropped to achieve sufficient bandwidth requirement reduction to ensure that higher priority packets would have bandwidth to be transmitted toward the subscriber. Low priority packets associated with all video streams being multiplexed for a single access subscriber line would be candidates for the discardable packet selection process in step 90. The idea is that it would be better to discard a low priority frame from-another queue within the multiplex group than to discard a higher priority packet from the video stream that was causing the overload condition. The smallest number of low priority packets could be selected similar to step 86 of FIG. 17. If discarding all low priority packets is not enough to provide timely transmission of a high priority packet, then preceding high priority packets in the same queue will be made available for discard in step 92. Thus, a high priority video packet entering the multiplexing queue during congestion periods would have precedence over older packets of the same priority only within the same video stream. The idea is that if multiple high priority video packets are in the multiplex queue at the same time the transmission of the newest high priority will have the best result as far as video display quality is concerned. Also, it is likely that discarding older high priority packets in deference to a new high priority packet will have less objection effects than selecting a high priority packet from another stream, which could suffer objectionable video artifacts. It is more likely that the congested video stream would correct its video quality by the transmission of the new high priority packet.
  • An example is given in FIG. 19. FIG. 19 represents a traffic management system with queuing and scheduling functions that accommodates multiple video packet streams that are composed of both high priority video data packets and low priority video data packets. Also shown is the existence of a new high priority packet (packet 1 f) that has arrived at the traffic manager. Packet 1 f needs to enter Channel Queue 1 and be served by the output stage 44 before it ages pasts a delay period that would cause the packet to arrive at the subscribers video decoder too late to be decoded and displayed. Since this new packet is High Priority, it has to be allowed to enter queue 1. If the total time required to transmit packet 1 f and all preceding packets is greater than the time-to-live for packet 1 f, then one or more packets in the queues should be dropped to make room for packet 1 f. The selection process would select which packet or packets to discard, following the above stated rules for this example, in the following order: 2a, na, 1b, nb, 1c, 2c, 2d, nd, 1e, 1a, 1d. The packet dropping process would stop when sufficient bandwidth is gained to be able to transmit the new video packet within the specified maximum age for video packets within the traffic manager.
  • This method for discarding packets delivers a high level of video display quality to the video viewer, with minimum processing.
  • Although the Detailed Description of the invention has been directed to certain exemplary embodiments, various modifications of these embodiments, as well as alternative embodiments, will be suggested to those skilled in the art. The invention encompasses any modifications or alternative embodiments that fall within the scope of the Claims.

Claims (20)

1. A multiplexer for broadband subscriber access, comprising:
a memory queue for storing packets of video information;
traffic management circuitry for selectively discarding one or more packets from the memory queue responsive to a detection of congestion in the memory queue based on parameters associated with existing packets currently in the queue.
2. The multiplexer of claim 1 wherein said traffic management circuitry detects congestion where a current transit time through the multiplexer is greater than a time-to-live parameter associated with an incoming packet.
3. The multiplexer of claim 2 wherein said time-to-live parameter is determined by observing the arrival of incoming packets.
4. The multiplexer of claim 1 wherein the traffic management circuitry detects congestion based on information associated with information on packets to be sent in the future.
5. The multiplexer of claim 4 wherein information on packets to be sent in the future comprises header information from packets to be sent in the future.
6. The multiplexer of claim 5 wherein said header information is stored in a virtual queue.
7. The multiplexer of claim 1 wherein said traffic management circuitry discards packets having interdependencies with previously discarded packets.
8. The multiplexer of claim 7 wherein the packets having interdependencies comprises packets containing differential video data.
9. The multiplexer of claim 1 wherein said memory queue comprises a plurality of first-in, first-out (FIFO) memories associated with respective data streams, and wherein said traffic management circuitry includes circuitry for discarding low priority packets from any of the FIFO memories responsive to a detection of congestion.
10. The multiplexer of claim 9 wherein said traffic management circuitry includes circuitry for discarding high priority packets only from the FIFO memory associated with an incoming packet causing the congestion.
11. A method of multiplexing packets for broadband subscriber access, comprising the steps of:
storing packets of video information in a memory queue;
detecting congestion situations;
selectively discarding one or more packets from the memory queue responsive to a detection of congestion in the memory queue based on parameters associated with existing packets currently in the queue.
12. The method of claim 11 wherein said detecting step comprises the step of detecting situations where a current transit time is greater than a time-to-live parameter associated with an incoming packet.
13. The method of claim 12 wherein said step of step of detecting situations where a current transit time is greater than a time-to-live parameter associated with an incoming packet comprises the step of observing the arrival of incoming packets to estimate the time-to-live parameter.
14. The method of claim 11 wherein the detecting step comprises the step of detecting congestion based on information associated with information on packets to be sent in the future.
15. The method of claim 14 wherein said step of detecting congestion based on information associated with information on packets to be sent in the future comprises the step of receiving packets including header information from packets to be sent in the future.
16. The method of claim 15 and further comprising the steps of storing the header information in a virtual queue.
17. The method of claim 11 wherein said discarding step comprises the step of discarding packets having interdependencies with previously discarded packets.
18. The method of claim 17 wherein the packets having interdependencies are packets containing differential video data.
19. The method of claim 11 wherein said memory queue comprises a plurality of first-in, first-out (FIFO) memories associated with respective data streams, and wherein said discarding step comprises the step of discarding low priority packets from any of the FIFO memories responsive to a detection of congestion.
20. The method of claim 19 wherein said traffic management circuitry further includes circuitry for discarding high priority packets only from a FIFO memory associated with an incoming packet causing the congestion.
US11/337,403 2006-01-23 2006-01-23 Video packet multiplexer with intelligent packet discard Abandoned US20070201365A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US11/337,403 US20070201365A1 (en) 2006-01-23 2006-01-23 Video packet multiplexer with intelligent packet discard
EP07000860.2A EP1811725B1 (en) 2006-01-23 2007-01-17 Packet discard in a multiplexer
CN200710007345.XA CN101026572B (en) 2006-01-23 2007-01-23 There is the video packet multiplexer of intelligent packet discard
US13/112,753 US8861346B2 (en) 2006-01-23 2011-05-20 Video packet multiplexer with intelligent packet discard

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US11/337,403 US20070201365A1 (en) 2006-01-23 2006-01-23 Video packet multiplexer with intelligent packet discard

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US13/112,753 Continuation US8861346B2 (en) 2006-01-23 2011-05-20 Video packet multiplexer with intelligent packet discard

Publications (1)

Publication Number Publication Date
US20070201365A1 true US20070201365A1 (en) 2007-08-30

Family

ID=37963854

Family Applications (2)

Application Number Title Priority Date Filing Date
US11/337,403 Abandoned US20070201365A1 (en) 2006-01-23 2006-01-23 Video packet multiplexer with intelligent packet discard
US13/112,753 Active 2027-04-07 US8861346B2 (en) 2006-01-23 2011-05-20 Video packet multiplexer with intelligent packet discard

Family Applications After (1)

Application Number Title Priority Date Filing Date
US13/112,753 Active 2027-04-07 US8861346B2 (en) 2006-01-23 2011-05-20 Video packet multiplexer with intelligent packet discard

Country Status (3)

Country Link
US (2) US20070201365A1 (en)
EP (1) EP1811725B1 (en)
CN (1) CN101026572B (en)

Cited By (47)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070183416A1 (en) * 2006-02-07 2007-08-09 Mark Gooch Per-port penalty queue system for re-prioritization of network traffic sent to a processor
US20080008188A1 (en) * 2006-05-25 2008-01-10 Proximetry, Inc. Systems and methods for wireless resource management with quality of service (qos) management
US20080075123A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Multiplexed Data Stream Timeslot Map
US20080074996A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Aggregated Link Traffic Protection
US20080075127A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Bandwidth Reuse in Multiplexed Data Stream
US20080075069A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Multi-Network Compatible Data Architecture
US20080075124A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Multi-Component Compatible Data Architecture
US20080075122A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Network Clock Synchronization Floating Window and Window Delineation
US20080075121A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Multi-Frame Network Clock Synchronization
US20080075002A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Multiplexed Data Stream Circuit Architecture
US20080075110A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Multiplexed Data Stream Payload Format
US20080101226A1 (en) * 2006-10-31 2008-05-01 Albrecht Alan R Buffer management in a network device
US20080112344A1 (en) * 2006-11-13 2008-05-15 Fujitsu Limited Stale data removal using latency count in a wimax scheduler
US20080144505A1 (en) * 2006-11-03 2008-06-19 Michael Anthony Isnardi Method and Apparatus for Bitrate Reduction
US20080181114A1 (en) * 2007-01-26 2008-07-31 Futurewei Technologies, Inc. Closed-Loop Clock Synchronization
US20080291935A1 (en) * 2007-05-25 2008-11-27 Campion Nicholas F Methods, Systems, and Computer Program Products for Selectively Discarding Packets
US20100142524A1 (en) * 2007-07-02 2010-06-10 Angelo Garofalo Application data flow management in an ip network
US7768919B1 (en) * 2003-04-28 2010-08-03 Verizon Laboratories Inc. Selective packet discard apparatus and method
US20110204717A1 (en) * 2010-02-22 2011-08-25 Cisco Technology, Inc. System and method for providing collaborating power controllers
US8023437B1 (en) 2006-06-28 2011-09-20 Insors Integrated Communications Methods, systems and program products for a distributed communications configuration
US20110252155A1 (en) * 2010-04-12 2011-10-13 Shyam Parekh Queue management unit and method for streaming video packets in a wireless network
US20110285863A1 (en) * 2010-05-23 2011-11-24 James Burke Live television broadcasting system for the internet
US8121990B1 (en) 2006-06-28 2012-02-21 Insors Integrated Communications Methods, systems and program products for communicating file modification information
US8144632B1 (en) * 2006-06-28 2012-03-27 Insors Integrated Communications Methods, systems and program products for efficient communications during data sharing event
US20120147746A1 (en) * 2010-12-14 2012-06-14 Cisco Technology, Inc. System and method for optimizing packet routing in a mesh network
US20120198509A1 (en) * 2011-01-27 2012-08-02 International Business Machines Corporation Systems and methods for managed video services at edge-of-the-network
US8295310B2 (en) 2006-09-25 2012-10-23 Futurewei Technologies, Inc. Inter-packet gap network clock synchronization
US8395652B1 (en) 2006-06-28 2013-03-12 Insors Integrated Communications Data network collaboration systems having a shared file
US8412773B1 (en) 2006-06-28 2013-04-02 Insors Integrated Communications Methods, systems and program products for initiating a process on data network
US8458283B1 (en) 2006-06-28 2013-06-04 Insors Integrated Communications Methods and program products for efficient communication of shared file modifications during a collaboration event
US8494009B2 (en) 2006-09-25 2013-07-23 Futurewei Technologies, Inc. Network clock synchronization timestamp
US8516050B1 (en) 2006-06-28 2013-08-20 Insors Integrated Communications Methods and program products for communicating file modifications during a collaboration event
US8730800B2 (en) 2008-11-17 2014-05-20 Huawei Technologies Co., Ltd. Method, apparatus, and system for transporting video streams
US20140241376A1 (en) * 2013-02-27 2014-08-28 Apple Inc. Apparatus and Method for Controlling Transaction Flow in Integrated Circuits
US8976705B2 (en) 2010-12-14 2015-03-10 Cisco Technology, Inc. System and method for providing configuration data in a mesh network
US20150085132A1 (en) * 2013-09-24 2015-03-26 Motorola Solutions, Inc Apparatus for and method of identifying video streams transmitted over a shared network link, and for identifying and time-offsetting intra-frames generated substantially simultaneously in such streams
US20150112909A1 (en) * 2013-10-17 2015-04-23 Qualcomm Incorporated Congestion avoidance in networks of spiking neurons
US20160087866A1 (en) * 2014-09-22 2016-03-24 Avaya Inc. Adaptive management of a media buffer
CN105656807A (en) * 2015-12-25 2016-06-08 盛科网络(苏州)有限公司 Network chip multi-channel data transmission method and transmission device
US20170155930A1 (en) * 2014-07-04 2017-06-01 Samsung Electronics Co., Ltd. Devices and methods for transmitting/receiving data in communication system
US9923836B1 (en) * 2014-11-21 2018-03-20 Sprint Spectrum L.P. Systems and methods for configuring a delay based scheduler for an access node
US20180288730A1 (en) * 2017-04-03 2018-10-04 Nxp B.V. Range determining module and associated methods and apparatus
US10505859B2 (en) * 2016-11-10 2019-12-10 The Government Of The United States Of America, As Represented By The Secretary Of The Navy Packet deadlines in a queue to control the age of information
US10560359B2 (en) * 2016-12-23 2020-02-11 Cisco Technology, Inc. Method and device for reducing multicast flow joint latency
US10887151B2 (en) 2018-10-05 2021-01-05 Samsung Eletrônica da Amazônia Ltda. Method for digital video transmission adopting packaging forwarding strategies with path and content monitoring in heterogeneous networks using MMT protocol, method for reception and communication system
US11228996B2 (en) * 2018-09-28 2022-01-18 Qualcomm Incorporated Delivery time windows for low latency communications
US20220141148A1 (en) * 2020-11-04 2022-05-05 Rajeshwar PATIL System and method for congestion management in computer networks

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101247595B1 (en) 2009-06-12 2013-03-26 시그너스 브로드밴드, 인코포레이티드 Systems and methods for intelligent discard in a communication network
US8627396B2 (en) 2009-06-12 2014-01-07 Cygnus Broadband, Inc. Systems and methods for prioritization of data for intelligent discard in a communication network
US8531961B2 (en) 2009-06-12 2013-09-10 Cygnus Broadband, Inc. Systems and methods for prioritization of data for intelligent discard in a communication network
US9706257B2 (en) 2009-09-14 2017-07-11 At&T Intellectual Property I, L.P. Viewing control management across multiple access points
US9325502B2 (en) * 2009-11-13 2016-04-26 At&T Intellectual Property I, L.P. Identity management for transactional content
CA2786200C (en) * 2011-09-23 2015-04-21 Cygnus Broadband, Inc. Systems and methods for prioritization of data for intelligent discard in a communication network
US20150312651A1 (en) * 2014-04-28 2015-10-29 Honeywell International Inc. System and method of optimized network traffic in video surveillance system
CN104104685B (en) * 2014-07-23 2018-01-02 天脉聚源(北京)教育科技有限公司 A kind of picture retransmission method and device
US10554571B2 (en) * 2015-08-18 2020-02-04 Avago Technologies International Sales Pte. Limited Packet-to-packet timing reconstruction for channel bonding
CN108174305B (en) * 2016-12-07 2019-03-12 北京云中融信网络科技有限公司 The message treatment method and device of chatroom is broadcast live
GB2568460B (en) * 2017-10-16 2022-11-16 Displaylink Uk Ltd Encoding and transmission of display data
CN109669917B (en) * 2018-12-24 2020-07-10 北京华大九天软件有限公司 Waveform storage method based on priority grouping

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5444706A (en) * 1989-12-27 1995-08-22 Kabushiki Kaisha Toshiba Packet exchange network and method of controlling discard of packets in packet exchange network
US6031940A (en) * 1996-11-27 2000-02-29 Teralogic, Inc. System and method for efficiently encoding video frame sequences
US20020110134A1 (en) * 2000-12-15 2002-08-15 Glenn Gracon Apparatus and methods for scheduling packets in a broadband data stream
US6751194B1 (en) * 1999-05-31 2004-06-15 Nec Corporation Packet multiplexer for priority control
US6967924B1 (en) * 1996-01-29 2005-11-22 Hitachi, Ltd. Packet switching device and cell transfer control method
US6990071B2 (en) * 2000-03-30 2006-01-24 Network Physics, Inc. Method for reducing fetch time in a congested communication network
US7499402B2 (en) * 2004-04-13 2009-03-03 Radioframe Networks, Inc. Network delay control

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6360271B1 (en) * 1999-02-02 2002-03-19 3Com Corporation System for dynamic jitter buffer management based on synchronized clocks
US6091709A (en) * 1997-11-25 2000-07-18 International Business Machines Corporation Quality of service management for packet switched networks
JP3601994B2 (en) * 1998-09-17 2004-12-15 沖電気工業株式会社 ATM cell multiplexing apparatus and ATM cell multiplexing method
US6882623B1 (en) * 2000-02-08 2005-04-19 Native Networks Technologies Ltd. Multi-level scheduling method for multiplexing packets in a communications network
KR100528861B1 (en) * 2003-03-04 2005-11-16 삼성전자주식회사 Packet forwarding system having control packet processor that allocates the bandwidth for control packet adaptively and a method processing thereof
CN1324862C (en) * 2004-01-06 2007-07-04 北京邮电大学 Method for regulating congest window in communication network
US7805482B2 (en) * 2005-12-30 2010-09-28 Senactive It-Dienstleistungs Gmbh Method of correlating events in data packet streams

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5444706A (en) * 1989-12-27 1995-08-22 Kabushiki Kaisha Toshiba Packet exchange network and method of controlling discard of packets in packet exchange network
US6967924B1 (en) * 1996-01-29 2005-11-22 Hitachi, Ltd. Packet switching device and cell transfer control method
US6031940A (en) * 1996-11-27 2000-02-29 Teralogic, Inc. System and method for efficiently encoding video frame sequences
US6751194B1 (en) * 1999-05-31 2004-06-15 Nec Corporation Packet multiplexer for priority control
US6990071B2 (en) * 2000-03-30 2006-01-24 Network Physics, Inc. Method for reducing fetch time in a congested communication network
US20020110134A1 (en) * 2000-12-15 2002-08-15 Glenn Gracon Apparatus and methods for scheduling packets in a broadband data stream
US7499402B2 (en) * 2004-04-13 2009-03-03 Radioframe Networks, Inc. Network delay control

Cited By (82)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7768919B1 (en) * 2003-04-28 2010-08-03 Verizon Laboratories Inc. Selective packet discard apparatus and method
US20070183416A1 (en) * 2006-02-07 2007-08-09 Mark Gooch Per-port penalty queue system for re-prioritization of network traffic sent to a processor
US20080008188A1 (en) * 2006-05-25 2008-01-10 Proximetry, Inc. Systems and methods for wireless resource management with quality of service (qos) management
US9565396B2 (en) 2006-06-28 2017-02-07 Iocom Uk Limited Methods, systems and program products for initiating a process on data network
US9436700B2 (en) 2006-06-28 2016-09-06 Iocom Uk Limited Methods and program products for communicating file modifications during a collaboration event
US8516050B1 (en) 2006-06-28 2013-08-20 Insors Integrated Communications Methods and program products for communicating file modifications during a collaboration event
US8458283B1 (en) 2006-06-28 2013-06-04 Insors Integrated Communications Methods and program products for efficient communication of shared file modifications during a collaboration event
US8412773B1 (en) 2006-06-28 2013-04-02 Insors Integrated Communications Methods, systems and program products for initiating a process on data network
US8395652B1 (en) 2006-06-28 2013-03-12 Insors Integrated Communications Data network collaboration systems having a shared file
US8144632B1 (en) * 2006-06-28 2012-03-27 Insors Integrated Communications Methods, systems and program products for efficient communications during data sharing event
US8121990B1 (en) 2006-06-28 2012-02-21 Insors Integrated Communications Methods, systems and program products for communicating file modification information
US8023437B1 (en) 2006-06-28 2011-09-20 Insors Integrated Communications Methods, systems and program products for a distributed communications configuration
US8401010B2 (en) 2006-09-25 2013-03-19 Futurewei Technologies, Inc. Multi-component compatible data architecture
US8295310B2 (en) 2006-09-25 2012-10-23 Futurewei Technologies, Inc. Inter-packet gap network clock synchronization
US20080075123A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Multiplexed Data Stream Timeslot Map
US20080074996A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Aggregated Link Traffic Protection
US9106439B2 (en) 2006-09-25 2015-08-11 Futurewei Technologies, Inc. System for TDM data transport over Ethernet interfaces
US9019996B2 (en) 2006-09-25 2015-04-28 Futurewei Technologies, Inc. Network clock synchronization floating window and window delineation
US20080075069A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Multi-Network Compatible Data Architecture
US8982912B2 (en) 2006-09-25 2015-03-17 Futurewei Technologies, Inc. Inter-packet gap network clock synchronization
US7809027B2 (en) 2006-09-25 2010-10-05 Futurewei Technologies, Inc. Network clock synchronization floating window and window delineation
US7813271B2 (en) 2006-09-25 2010-10-12 Futurewei Technologies, Inc. Aggregated link traffic protection
US7961751B2 (en) 2006-09-25 2011-06-14 Futurewei Technologies, Inc. Multiplexed data stream timeslot map
US7986700B2 (en) 2006-09-25 2011-07-26 Futurewei Technologies, Inc. Multiplexed data stream circuit architecture
US8976796B2 (en) 2006-09-25 2015-03-10 Futurewei Technologies, Inc. Bandwidth reuse in multiplexed data stream
US20080075127A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Bandwidth Reuse in Multiplexed Data Stream
US8837492B2 (en) 2006-09-25 2014-09-16 Futurewei Technologies, Inc. Multiplexed data stream circuit architecture
US20080075124A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Multi-Component Compatible Data Architecture
US8660152B2 (en) 2006-09-25 2014-02-25 Futurewei Technologies, Inc. Multi-frame network clock synchronization
US20080075110A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Multiplexed Data Stream Payload Format
US20080075002A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Multiplexed Data Stream Circuit Architecture
US20080075122A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Network Clock Synchronization Floating Window and Window Delineation
US8588209B2 (en) 2006-09-25 2013-11-19 Futurewei Technologies, Inc. Multi-network compatible data architecture
US8289962B2 (en) 2006-09-25 2012-10-16 Futurewei Technologies, Inc. Multi-component compatible data architecture
US8494009B2 (en) 2006-09-25 2013-07-23 Futurewei Technologies, Inc. Network clock synchronization timestamp
US8532094B2 (en) 2006-09-25 2013-09-10 Futurewei Technologies, Inc. Multi-network compatible data architecture
US8340101B2 (en) 2006-09-25 2012-12-25 Futurewei Technologies, Inc. Multiplexed data stream payload format
US20080075121A1 (en) * 2006-09-25 2008-03-27 Futurewei Technologies, Inc. Multi-Frame Network Clock Synchronization
US8054847B2 (en) * 2006-10-31 2011-11-08 Hewlett-Packard Development Company, L.P. Buffer management in a network device
US20080101226A1 (en) * 2006-10-31 2008-05-01 Albrecht Alan R Buffer management in a network device
US20080144505A1 (en) * 2006-11-03 2008-06-19 Michael Anthony Isnardi Method and Apparatus for Bitrate Reduction
US8355403B2 (en) * 2006-11-13 2013-01-15 Fujitsu Semiconductor Limited Stale data removal using latency count in a WiMAX scheduler
US20080112344A1 (en) * 2006-11-13 2008-05-15 Fujitsu Limited Stale data removal using latency count in a wimax scheduler
US8605757B2 (en) 2007-01-26 2013-12-10 Futurewei Technologies, Inc. Closed-loop clock synchronization
US7787498B2 (en) * 2007-01-26 2010-08-31 Futurewei Technologies, Inc. Closed-loop clock synchronization
US20080181114A1 (en) * 2007-01-26 2008-07-31 Futurewei Technologies, Inc. Closed-Loop Clock Synchronization
US20080291935A1 (en) * 2007-05-25 2008-11-27 Campion Nicholas F Methods, Systems, and Computer Program Products for Selectively Discarding Packets
US7646781B2 (en) * 2007-05-25 2010-01-12 International Business Machines Corporation Methods, systems, and computer program products for selectively discarding packets
US20100142524A1 (en) * 2007-07-02 2010-06-10 Angelo Garofalo Application data flow management in an ip network
US20150071073A1 (en) * 2007-07-02 2015-03-12 Telecom Italia S.P.A. Application data flow management in an ip network
US9935884B2 (en) * 2007-07-02 2018-04-03 Telecom Italia S.P.A. Application data flow management in an IP network
US8891372B2 (en) * 2007-07-02 2014-11-18 Telecom Italia S.P.A. Application data flow management in an IP network
US8730800B2 (en) 2008-11-17 2014-05-20 Huawei Technologies Co., Ltd. Method, apparatus, and system for transporting video streams
US8941261B2 (en) 2010-02-22 2015-01-27 Cisco Technology, Inc. System and method for providing collaborating power controllers
US20110204717A1 (en) * 2010-02-22 2011-08-25 Cisco Technology, Inc. System and method for providing collaborating power controllers
US20110252155A1 (en) * 2010-04-12 2011-10-13 Shyam Parekh Queue management unit and method for streaming video packets in a wireless network
US8443097B2 (en) * 2010-04-12 2013-05-14 Alcatel Lucent Queue management unit and method for streaming video packets in a wireless network
US20110285863A1 (en) * 2010-05-23 2011-11-24 James Burke Live television broadcasting system for the internet
US8976705B2 (en) 2010-12-14 2015-03-10 Cisco Technology, Inc. System and method for providing configuration data in a mesh network
US8605591B2 (en) * 2010-12-14 2013-12-10 Cisco Technology, Inc. System and method for optimizing packet routing in a mesh network
US20120147746A1 (en) * 2010-12-14 2012-06-14 Cisco Technology, Inc. System and method for optimizing packet routing in a mesh network
US20120324524A1 (en) * 2011-01-27 2012-12-20 International Business Machines Corporation Managed video services at edge-of-the-network
US20120198509A1 (en) * 2011-01-27 2012-08-02 International Business Machines Corporation Systems and methods for managed video services at edge-of-the-network
US8898718B2 (en) * 2011-01-27 2014-11-25 International Business Machines Corporation Systems and methods for managed video services at edge-of-the-network
US9270610B2 (en) * 2013-02-27 2016-02-23 Apple Inc. Apparatus and method for controlling transaction flow in integrated circuits
US20140241376A1 (en) * 2013-02-27 2014-08-28 Apple Inc. Apparatus and Method for Controlling Transaction Flow in Integrated Circuits
US20150085132A1 (en) * 2013-09-24 2015-03-26 Motorola Solutions, Inc Apparatus for and method of identifying video streams transmitted over a shared network link, and for identifying and time-offsetting intra-frames generated substantially simultaneously in such streams
US9544534B2 (en) * 2013-09-24 2017-01-10 Motorola Solutions, Inc. Apparatus for and method of identifying video streams transmitted over a shared network link, and for identifying and time-offsetting intra-frames generated substantially simultaneously in such streams
US20150112909A1 (en) * 2013-10-17 2015-04-23 Qualcomm Incorporated Congestion avoidance in networks of spiking neurons
US10701408B2 (en) * 2014-07-04 2020-06-30 Samsung Electronics Co., Ltd. Devices and methods for transmitting/receiving data in communication system
US20170155930A1 (en) * 2014-07-04 2017-06-01 Samsung Electronics Co., Ltd. Devices and methods for transmitting/receiving data in communication system
US20160087866A1 (en) * 2014-09-22 2016-03-24 Avaya Inc. Adaptive management of a media buffer
US9769043B2 (en) * 2014-09-22 2017-09-19 Avaya Inc. Adaptive management of a media buffer
US9923836B1 (en) * 2014-11-21 2018-03-20 Sprint Spectrum L.P. Systems and methods for configuring a delay based scheduler for an access node
CN105656807A (en) * 2015-12-25 2016-06-08 盛科网络(苏州)有限公司 Network chip multi-channel data transmission method and transmission device
US10505859B2 (en) * 2016-11-10 2019-12-10 The Government Of The United States Of America, As Represented By The Secretary Of The Navy Packet deadlines in a queue to control the age of information
US10560359B2 (en) * 2016-12-23 2020-02-11 Cisco Technology, Inc. Method and device for reducing multicast flow joint latency
US10383085B2 (en) * 2017-04-03 2019-08-13 Nxp B.V. Range determining module and associated methods and apparatus
US20180288730A1 (en) * 2017-04-03 2018-10-04 Nxp B.V. Range determining module and associated methods and apparatus
US11228996B2 (en) * 2018-09-28 2022-01-18 Qualcomm Incorporated Delivery time windows for low latency communications
US10887151B2 (en) 2018-10-05 2021-01-05 Samsung Eletrônica da Amazônia Ltda. Method for digital video transmission adopting packaging forwarding strategies with path and content monitoring in heterogeneous networks using MMT protocol, method for reception and communication system
US20220141148A1 (en) * 2020-11-04 2022-05-05 Rajeshwar PATIL System and method for congestion management in computer networks

Also Published As

Publication number Publication date
CN101026572A (en) 2007-08-29
EP1811725A3 (en) 2008-06-25
EP1811725A2 (en) 2007-07-25
CN101026572B (en) 2015-10-14
US20110222400A1 (en) 2011-09-15
US8861346B2 (en) 2014-10-14
EP1811725B1 (en) 2014-10-22

Similar Documents

Publication Publication Date Title
US8861346B2 (en) Video packet multiplexer with intelligent packet discard
US7609709B2 (en) Video aware traffic management
EP1122931B1 (en) Real-time media content synchronization and transmission in packet network apparatus and method
US7079486B2 (en) Adaptive threshold based jitter buffer management for packetized data
EP1606717B1 (en) Method and device for multimedia streaming
US8437255B2 (en) Estimating wireless processing device queue length and estimating signal reception quality in a wireless network
US7307989B2 (en) Window flow control with common time reference
US7701884B2 (en) Network communications bandwidth control
US20060268692A1 (en) Transmission of electronic packets of information of varying priorities over network transports while accounting for transmission delays
JP2005526422A (en) Communication system and communication technique for transmission from source to destination
CN112019392A (en) Multi-feature stream access method for TSN (traffic service network)
US7768919B1 (en) Selective packet discard apparatus and method
WO2010138913A1 (en) Systems and methods for video statistical multiplexing adapting to internet protocol networks
EP1931068A1 (en) Method of adaptively dejittering packetized signals buffered at the receiver of a communication network node
US20160173395A1 (en) Timing transport method in a communication network
CN111163362A (en) Video receiving method and system capable of self-adapting retransmission waiting time
EP2291955B1 (en) Packet latency estimation
Narbutt et al. Adaptive playout buffering for audio/video transmission over the internet
Sehgal et al. Fast distortion-buffer optimized streaming of multimedia
Sreeraj et al. Optimizing the jitter losses using adaptive jitter buffer at the receiver

Legal Events

Date Code Title Description
AS Assignment

Owner name: ALCATEL USA SOURCING, L.P., TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SKOOG, FREDERICK;BLANTON, JOHN;REEL/FRAME:017147/0984;SIGNING DATES FROM 20060118 TO 20060120

AS Assignment

Owner name: ALCATEL USA SOURCING, INC., TEXAS

Free format text: CHANGE OF NAME;ASSIGNOR:ALCATEL USA SOURCING, L.P.;REEL/FRAME:026159/0315

Effective date: 20061231

Owner name: ALCATEL-LUCENT USA INC., NEW JERSEY

Free format text: MERGER;ASSIGNOR:ALCATEL USA SOURCING, INC.;REEL/FRAME:026155/0216

Effective date: 20081101

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO PAY ISSUE FEE

AS Assignment

Owner name: ALCATEL LUCENT, FRANCE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ALCATEL-LUCENT USA INC.;REEL/FRAME:033616/0042

Effective date: 20140826