US20120011191A2 - Data segmentation, request and transfer method - Google Patents

Data segmentation, request and transfer method Download PDF

Info

Publication number
US20120011191A2
US20120011191A2 US12/713,939 US71393910A US2012011191A2 US 20120011191 A2 US20120011191 A2 US 20120011191A2 US 71393910 A US71393910 A US 71393910A US 2012011191 A2 US2012011191 A2 US 2012011191A2
Authority
US
United States
Prior art keywords
range
server
network interfaces
requests
data resource
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US12/713,939
Other versions
US20110213827A1 (en
US8516147B2 (en
Inventor
Dominik KASPAR
Kristian EVENSEN
Paal Engelstad
Audun HANSEN
Carsten Griwodz
Pal HALVORSEN
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Celerway Communication As
Original Assignee
Simula Innovation AS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to US12/713,939 priority Critical patent/US8516147B2/en
Application filed by Simula Innovation AS filed Critical Simula Innovation AS
Assigned to SIMULA INNOVATION AS, TELENOR ASA reassignment SIMULA INNOVATION AS ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HANSEN, AUDUN F., EVENSEN, KRISTIAN R., ENGELSTAD, PAAL E., GRIWODZ, CARSTEN, HALVORSEN, PAL, KASPAR, DOMINIK
Assigned to SIMULA INNOVATION SA reassignment SIMULA INNOVATION SA ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: TELENOR ASA
Publication of US20110213827A1 publication Critical patent/US20110213827A1/en
Publication of US20120011191A2 publication Critical patent/US20120011191A2/en
Priority to US13/946,610 priority patent/US20130304875A1/en
Publication of US8516147B2 publication Critical patent/US8516147B2/en
Application granted granted Critical
Assigned to SIMULA INNOVATION AS reassignment SIMULA INNOVATION AS CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE'S NAME PREVIOUSLY RECORDED ON REEL 025588 FRAME 0453. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT OF ASSIGNOR'S INTEREST. Assignors: TELENOR ASA
Assigned to CELERWAY COMMUNICATION AS reassignment CELERWAY COMMUNICATION AS ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SIMULA INNOVATION AS
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/75Media network packet handling
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/14Multichannel or multilink protocols
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/75Media network packet handling
    • H04L65/756Media network packet handling adapting media to device capabilities
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W76/00Connection management
    • H04W76/10Connection setup
    • H04W76/15Setup of multiple wireless link connections
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W28/00Network traffic management; Network resource management
    • H04W28/02Traffic management, e.g. flow control or congestion control
    • H04W28/06Optimizing the usage of the radio link, e.g. header compression, information sizing, discarding information
    • H04W28/065Optimizing the usage of the radio link, e.g. header compression, information sizing, discarding information using assembly or disassembly of packets
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W88/00Devices specially adapted for wireless communication networks, e.g. terminals, base stations or access point devices
    • H04W88/02Terminal devices
    • H04W88/06Terminal devices adapted for operation in multiple networks or having at least two operational modes, e.g. multi-mode terminals

Definitions

  • the present invention relates to a system and method of using range requests and pipelining to segment a data resource and transfer it over multiple network interfaces.
  • TCP is purely end-to-end and allows only one IP address (one interface) on each side.
  • SCTP supports multiple IP addresses per endpoint, but uses them only for failover.
  • a common approach is therefore to provide specialized libraries that transparently partition application-layer data into multiple transport streams.
  • the implementation of such middleware requires software modifications to clients and servers.
  • the present invention provides, inter alia, a method of requesting information.
  • the method includes a step of generating a plurality of range requests using a client having a plurality of network interfaces, the range requests each corresponding to a data range of a data resource. Also included in the method are the steps of transmitting, from the client having the plurality of network interfaces, a plurality of range requests to a server via the plurality of network interfaces, and receiving, at the client and from the server, a plurality of segments via the plurality of network interfaces such that each segment is received via a network interface which transmitted the corresponding range request, each segment including a portion of the data resource. Also included in the method is the step of reassembling the data resource using the plurality of segments.
  • the method further includes a step of generating, using the client having the plurality of network interfaces, at least one additional range request, each additional range request corresponding to a data range of the data resource, a step of determining an upper time limit for each at least one additional range request, a step of transmitting, from the client having a plurality of network interfaces, the at least one additional range request to the server via at least one of the plurality of network interfaces before the determined upper time limit, a step of receiving, at the client and from the server, a plurality of additional segments via the plurality of network interfaces such that each additional segment is received via a network interface which transmitted the corresponding additional range request, each additional segment including a portion of the data resource, and a step of reassembling the data resource using at least the additional plurality of segments.
  • the present invention also includes a method of transferring information.
  • the method includes a step of segmenting, using a client having a plurality of network interfaces, a data resource into a plurality of segments based on metadata of the data resource. Also included in the method are the steps of generating a plurality of range requests, the range requests each including a segment of the plurality of segments which corresponds to a data range of the data resource and transmitting, from the client having the plurality of network interfaces, the plurality of range requests to a server via the plurality of network interfaces for reassembly of the data resource using the plurality of segments at the server.
  • the device includes a plurality of network interfaces and a generation unit that generates a plurality of range requests, the range requests each corresponding to a data range of a data resource. Also included in the device is a communications unit that transmits the plurality of range requests to a server via the plurality of network interfaces, the server segmenting the data resource based on the range requests into a plurality of segments, where the communications unit further receives the plurality of segments, which are transmitted by the server, via the plurality of network interfaces such that each segment is received via a network interface which transmitted the corresponding range request. Moreover, the device includes a reassembly unit that reassembles the data resource using the plurality of segments received by the communications unit.
  • FIG. 1A illustrates a system according to an embodiment of the present invention including a client having multiple interfaces which communicate multiple range requests to a server;
  • FIG. 1B shows a block diagram illustrating an exemplary structure of the client according to an embodiment of the present invention
  • FIG. 1C shows a flow diagram illustrating a method according to an embodiment of the present invention
  • FIG. 1D shows a block diagram illustrating an exemplary structure of the server according to an embodiment of the present invention
  • FIG. 1E shows a flow diagram illustrating a method according to an embodiment of the present invention
  • FIG. 2A illustrates a system according to another embodiment of the present invention including a client having multiple interfaces which communicate multiple range requests to a server;
  • FIG. 2B illustrates a system according to another embodiment of the present invention including a client having multiple interfaces which communicate multiple data bearing range requests to a device;
  • FIG. 2C shows a block diagram illustrating an exemplary structure of the client according to another embodiment of the present invention.
  • FIG. 2D shows a flow diagram illustrating a method according to an embodiment of the present invention
  • FIG. 3 illustrates a data resource according to an embodiment of the present invention that has multiple sections each corresponding to a range of the data resource;
  • FIG. 4 illustrates a pipelining process according to an embodiment of the present invention, the pipelining distributing multiple requests from the client interface to the server;
  • FIG. 5 illustrates a pipelining process according to another embodiment of the present invention, the pipelining distributing multiple requests from a plurality of client interfaces to the server;
  • FIG. 6 shows a time diagram illustrating comparison between the time elapsed in a sequential process and a pipelined process
  • FIG. 7 illustrates experimental results obtained by downloading a file using two different interfaces
  • FIG. 8 illustrates a pipelining pattern when executing pipelining range requests using multiple interfaces during the start-up phase
  • FIG. 9 shows the performance gain of multi-segment pipelining in an exemplary embodiment
  • FIG. 10 illustrates an exemplary embodiment providing results from a performance comparison of using a single interface versus using multiple interfaces
  • FIG. 11 illustrates the received byte ranges during a progressing download according to an exemplary embodiment of the present invention
  • FIG. 12 illustrates an example of a startup latency during which data is downloaded but not yet played back at the desired playback bitrate according to an embodiment of the present invention
  • FIG. 13 illustrates a comparison of the throughput gain when using multiple interfaces and the throughput when using the fastest interface alone
  • FIG. 14 illustrates a comparison of the start-up latency at various segment sizes between two exemplary embodiments of the claimed invention.
  • FIG. 15 shows a block diagram illustrating a hardware configuration of a device of the present invention.
  • a data transmission system including a client 1 and a server 5 which are connected via the internet 10 .
  • the link between the client 1 and server 5 can be over other types of network links such as via a body area network (BAN), personal area network (PAN), local area network (LAN), storage area network (SAN), campus area network (CAN), metropolitan area network (MAN), wide area network (WAN), wireless wide area network (WWAN), or public switched telephony network (PSTN).
  • BAN body area network
  • PAN personal area network
  • LAN local area network
  • SAN storage area network
  • CAN campus area network
  • MAN metropolitan area network
  • WAN wide area network
  • WWAN wireless wide area network
  • PSTN public switched telephony network
  • the link can be a WWAN link compatible with a communication standard selected from the set of: Global System for Mobile communications (GSM), High Speed Packet Access (HSPA), Wideband Code Division Multiple Access (W-CDMA), Universal Mobile Telecommunications System (UMTS), Long Term Evolution (LTE), Worldwide interoperability for Microwave access (WiMax), Evolution-Data Optimized (EV-DO).
  • GSM Global System for Mobile communications
  • HSPA High Speed Packet Access
  • W-CDMA Wideband Code Division Multiple Access
  • UMTS Universal Mobile Telecommunications System
  • LTE Long Term Evolution
  • WiMax Worldwide interoperability for Microwave access
  • EV-DO Evolution-Data Optimized
  • the client 1 includes a plurality of interfaces 2 A, 2 B, . . . 2 n .
  • the client 1 includes a generation unit 15 and a communication unit 16 .
  • the generation unit 15 (using a combination of hardware and software) generates a plurality of range requests 4 A, 4 B, . . . 4 n .
  • the range requests 4 A, 4 B, . . . 4 n each correspond to a data range of a data resource 6 .
  • the communication unit 16 then transmits the plurality of range requests 4 A, 4 B, . . . 4 n to the communication unit 23 of the server 5 via the plurality of network interfaces 2 A, 2 B, . . .
  • the server 5 segments or splits, based on the range requests 4 A, 4 B, . . . 4 n and using the segmentation unit 24 shown in FIG. 1D , the data resource 6 into a plurality of segments which are transmitted back using the communication unit 23 to the communication unit 16 of the client 1 via the network interface 2 A, 2 B, . . . 2 n which transmitted the corresponding range request 4 A, 4 B, . . . 4 n .
  • the data resource 6 may be stored in the storage unit 25 of the server 5 . Additionally, before segmentation, the data resource 6 may be computed on request by the server 5 or retrieved from some other source of data.
  • Duplicates of the data resource 6 may also be stored on or obtained by multiple servers. Such a configuration enables the client 1 to obtain portions of the data resource 6 from different servers. In an embodiment, this can be accomplished by the client using metalinks to obtain the data resource 6 . For example, if the client knows that URI-1 and URI-2 are copies on different servers, range requests could be made to both servers. Thus, for example, a first portion of the plurality of range requests 4 A, 4 B, . . . 4 n could be transmitted to a first server, a second portion of the plurality range requests 4 A, 4 B, . . . 4 n could be transmitted to a second server, and an nth portion of the plurality of range requests 4 A, 4 B, . . . 4 n could be transmitted to an nth server.
  • interfaces 2 A, 2 B, . . . 2 n connect the client 1 to the server 5 .
  • the interfaces 2 A, 2 B, . . . 2 n each connect to a respective distinct ISP 3 A, 3 B, . . . 3 n or network provider by which the range requests 4 A, 4 B, . . . 4 n and respective data provision responses are sent.
  • FIG. 1C illustrates steps performed by the client 1 when obtaining the data resource 6 from the server 5 .
  • the client 1 generates (S 1 ) a plurality of range requests 4 A, 4 B, . . . 4 n , the plurality of range requests 4 A, 4 B, . . . 4 n each corresponding to a data range of the data resource 6 .
  • the client 1 transmits (S 2 ) the plurality of range requests 4 A, 4 B, . . . 4 n to the server 5 via the plurality of network interfaces 2 A, 2 B, . . . 2 n .
  • the client then receives (S 3 ) the plurality of segments via the plurality of network interfaces 2 A, 2 B, . . . 2 n such that each segment is received via the network interface which transmitted the corresponding range request, each segment including stored therein a portion of the data resource 6 . Then, the client 1 reassembles (S 4 ) the data resource 6 which can be displayed, stored, etc. by the client 1 .
  • FIG. 1E illustrates steps performed by the server 5 when transferring the data resource 6 to the client 1 .
  • the server 5 obtains (S 11 ) the data resource 6 .
  • the data resource may be obtained from the storage unit 25 of the server 5 , may be generated by the server 5 or may be obtained from an external source.
  • the server 5 then segments (S 12 ) the obtained data resource 6 into a plurality of segments based on the range requests 4 A, 4 B, . . . 4 n and transmits (S 13 ) the plurality of segments to the client 1 .
  • FIG. 2A illustrates an alternate embodiment in which the interfaces 2 A and 2 B are connected to the same ISP 3 A.
  • the interfaces 2 A and 2 B send separate range requests 4 A, 4 B and receive separate data provision responses
  • the network path at least part of the way between the client 1 and the server 5 , is by way of the same ISP 3 A or same network provider.
  • One result of the use of multiple interfaces is the increase in bandwidth on a client device 1 which has access to, and is able to connect to, one or more networks over more than one interface 2 A, 2 B, . . . 2 n .
  • a device which can utilize this configuration is a mobile phone (client) which has access to the Internet directly or indirectly using a WiMax link, a 3G link, a WiFi link, a cable, and/or Bluetooth, for example.
  • a laptop computer having an Ethernet cable interface as well as WiFi interface.
  • the various interfaces may use different protocols on the physical and link layers, but will typically use the same protocols on the network layer and the transport layer, for example TCP/IP or UDP/IP.
  • Each network interface can establish an end-to-end communication link using for example: Transmission Control Protocol (TCP), Stream Control Transmission Protocol (SCTP), Datagram Congestion Control Protocol (DCCP), User Datagram Protocol (UDP, which is connectionless), etc.
  • TCP Transmission Control Protocol
  • SCTP Stream Control Transmission Protocol
  • DCCP Datagram Congestion Control Protocol
  • UDP User Datagram Protocol
  • the end-to-end communication link can be one of Hypertext Transfer Protocol (HTTP), File Transfer Protocol (FTP), Session Initiation Protocol (SIP), Internet Message Access Protocol (IMAP), Session Description Protocol (SDP), etc.
  • HTTP Hypertext Transfer Protocol
  • FTP File Transfer Protocol
  • SIP Session Initiation Protocol
  • IMAP Internet Message Access Protocol
  • SDP Session Description Protocol
  • the multiple range requests are created from one request referencing an entire amount of data.
  • the ranges represent subsections of the entire amount of data.
  • the first single request is created by a browser, and the corresponding multiple requests are created by a middleware on the client that distributes the request over the multiple interfaces.
  • the efficiency of concurrent utilization of multiple network interfaces is improved.
  • the efficiency of the system is further increased by using range requests to subdivide the data resource 6 such that individual data resources (e.g. a file) can be distributed over the multiple links.
  • the data resources can be a file, data produced from a scanner, a web camera, a microphone, video, audio, etc.
  • the server 5 includes a data file 6 which is accessed by the client 1 .
  • the client 1 sends a plurality of range requests 4 A, 4 B, . . . 4 n using the plurality of interfaces 2 A, 2 B, . . . 2 n.
  • FIG. 2B illustrates an embodiment in which, instead of simply sending a range request, a plurality of interfaces 22 A, 22 B, . . . 22 n of the client 1 each send a request 21 A, 21 B, . . . 21 n which includes a range of some data resource, to a device 20 .
  • the system can utilize a server 5 or some other providing resource storing or generating data to be retrieved, such as a file server, a scanner, a web camera, etc. and can also utilize a device 20 which itself is or which hosts the providing resource which is capable of receiving data such as, for example, a file (e.g. on a file server), a printer, a program or script capable of processing received data.
  • the data resource or the providing resource is identified by a URI (uniform resource identifier), the URI being a string of characters used to identify a name or a resource.
  • a URI uniform resource identifier
  • Both the data resource and the providing resource can be identified by a URI (i.e. a URL or a URN) as a URI can identify a device, data or something else, depending on circumstances.
  • FIG. 2C illustrates the structure of the client 1 according to the embodiment of the invention described in FIG. 2B .
  • the client 1 shown in FIG. 2C includes a plurality of interfaces 2 A, 2 B, . . . 2 n .
  • the client 1 shown in FIG. 2C uses the segmentation unit 18 to segment the data resource which is either stored in the storage 17 , generated at by the client 1 , or retrieved from an external source by the client 1 , into a plurality of segments based on a current status of the plurality of network interfaces 2 A, 2 B, . . . 2 n and/or the size of the data resource.
  • the client 1 also generates, using the generation unit 15 , a plurality of range requests 4 A, 4 B, . . . 4 n each including a segment of the plurality of segments which corresponds to a data range of the data resource.
  • the segmentation may be performed before any portion of the data resource is transferred or may be performed while transfer of the data resource is in progress based on measured results of previously transmitted segments (measured bandwidth, round trip delay, etc.).
  • the client 1 then uses the communication unit 16 to transmit the plurality of range requests 4 A, 4 B, . . . 4 n including data resource segments to the device 20 via the plurality of network interfaces 2 A, 2 B, . . . 2 n .
  • the client 1 may also transmit the plurality of range requests 4 A, 4 B, . . . 4 n including data resource segments to a server 5 .
  • FIG. 2D illustrates steps performed by the client 1 and the server 5 when transferring the data resource 6 from the client 1 to the device 20 .
  • the client 1 segments (S 20 ) a data resource into a plurality of segments.
  • the client then generates (S 21 ) a plurality of range requests 4 A, 4 B, . . . 4 n each including a segment of the plurality of segments which corresponds to a data range of the data resource 6 .
  • the plurality of range requests are then transmitted (S 22 ) from the client 1 to the device 20 which receives the transmitted range requests (S 23 ).
  • the device 20 transmits (S 24 ) an acknowledgment for each received range request to the client 1 .
  • FIG. 3 illustrates an example of a data resource (data file 6 , for example) which is accessed by a plurality of range requests.
  • Each range request corresponds to a portion of the data resource.
  • the range of data is a subset of data retrievable from the providing resource in response to requests, and an identification of a range includes at least a beginning (and optionally an end or an extent) of that range.
  • the data represented by the range is expected to be received by the client device in response to the request and over the same network interface which transmitted the request.
  • the range of data is a subset of data stored, retrieved from an external device or created in the providing resource
  • an identification of a range includes at least a beginning (and optionally an end or an extent) of that range, and the data represented by the range is included as payload in the request.
  • An expected response to the request is a received confirmation that the data was received and is handled according to the request.
  • the beginning of the range can be defined as a first byte position in a byte range, a time stamp, a page number, etc.
  • the extent of the range can be defined as a number of bytes in a byte range, a duration of time (video, audio), a number of pages (document), a number of blocks (image), a number of frames (video), etc.
  • the invention can be implemented using existing protocols, provided that the protocol implementation at both the client and the server support ranges.
  • HTTP supports range requests.
  • the invention is, however, not limited to existing protocols.
  • the beginning and the end of the range has to be specified.
  • the server will send all the specified bytes and cannot be interrupted (unless the underlying TCP connection is cancelled and restarted).
  • the present invention allows ranges to have undefined extents (“open ends”), which is different from connections using the current http protocol.
  • a request which includes the segmented data resource, may be uploaded using an HTTP PUT or POST-like request. (Note that the current version of HTTP does not support POST requests with ranges.)
  • the response from the server is the data represented by the requested range and received over the same interface.
  • the response from the device is a status report confirming (or not) that the transmitted data was received and processed according to the request.
  • the size of the range for each request from each interface is dynamically determined and is based on certain factors. For example, the size of the range requests are determined such that they subdivide data payloads so that individual data resources (e.g. a file) can be distributed over several links in either predefined fixed sizes or dynamically determined by link capacity and/or delay.
  • individual data resources e.g. a file
  • the size of the range requests should be as small as possible but not so small that the available bandwidth is not optimally utilized, e.g. the size of the range should be larger than a bandwidth-delay product for the selected interface.
  • Small ranges provide a fine granularity which allows a “smooth”, “progressive” processing of the transferred resource.
  • the startup latency (the waiting time that is required before the transferred resource can be processed at an aggregated rate corresponding to the rate obtained from combining multiple interfaces) directly relates to the size of the ranges: the smaller the range, the smaller the required startup waiting time.
  • the startup waiting time increases.
  • a large range increases the chance of “unsmooth” processing if there are variances in transfer rate. For example, if one interface unexpectedly drops to 50% of the speed, it suddenly takes twice as long to “fill up a gap of missing data in the transferred resource.” Such an issue is mitigated by using small ranges.
  • a too small range is that the bandwidth is not optimally utilized. If less than (bandwidth *delay) bytes are requested (per interface), then the server is able to transfer an entire range before knowing what the next range should be. Therefore, the server will be idle, resulting in the bandwidth not being optimally utilized.
  • the size of the range request for an interface is determined based, at least in part, by the bandwidth available at the interface.
  • the bandwidth and/or delay may be determined during initial transmission.
  • the bandwidth can be estimated by counting the number of transferred bytes per time unit.
  • capacity can change during data transfer, tracking and adapting to the changes in bandwidth/delay may be performed. For example, as a bandwidth capacity value may change over time, some averaging can be performed to optimize this value.
  • the delay can be estimated by measuring the time between when a request is sent and an answer is received.
  • Another way to determine the bandwidth, which can be used in certain circumstances, is to use feedback from the TCP stack, as TCP keeps estimates of the round-trip delay.
  • An alternative approach is to probe the link with differently sized range requests which gives feedback on what the optimal range size should be based on current network conditions.
  • the size or particular range of the range requests can be determined based on obtaining metadata of the data resource.
  • the metadata of the data resource is obtained by transmitting a probing request and receiving the metadata in response. For example, transmitting an HTTP HEAD request (e.g. over the fastest network interface) using the same URI, and receiving at least one entity header field returned in response to the HTTP HEAD request.
  • the metadata of the data resource is obtained by identifying the metadata from the plurality of segments received from the server. For example, identifying entity header fields received in the responses to the plurality of HTTP GET requests sent over the multiple network interfaces.
  • the metadata may be among other things, a content-length header field, a content-encoding header field, content-type header field, or a content-range entity header field.
  • the metadata in a returned content-range header field may be used for determining a particular range of a range request.
  • the client would not have to remember the value of previously requested ranges as these values would be returned in the content-range header field.
  • the client could then generate a range request requesting a range consecutive to the immediately received range based on the metadata.
  • the particular interface which sends a particular range request can be selected based on certain factors.
  • the network interface can be selected such that traffic is distributed efficiently.
  • the network interface is selected based on monitoring, at least a subset of, the status of the network interfaces, the response times for previous requests, the number of sent/received bytes, etc. The selection can be made prior to or after the creation of the request i.e.
  • first a request is created and then a selection of an interface is made taking into account the nature of the request; or first a selection of an interface is made—for example because the interface becomes idle—and then a request is made, taking into account the capabilities and/or current status of the interface when creating the request by, for example, determining range size based on bandwidth-delay product, as described above.
  • an interface may be in one of several states including, idle, in the process of sending a request, in the process of waiting for a response, or in the process of receiving a response.
  • the latter three states may occur simultaneously (e.g. the interface may be sending and receiving at the same time, or sending a new request while waiting for a response to a previous request).
  • the requesting client may count the incoming bytes and compare them with the size of the requested range. Each response starts with a header that indicates that a new range is arriving.
  • the client by way of the communications unit, for example, is able to detect outages by monitoring the interfaces and rebalancing the load to an interface that is still working such that the data transfer can continue from the point at which the failure occurred. Any request which is only partially responded to would have to be sent again over one of the operating interfaces.
  • FIG. 4 illustrates an example of the pipelining process between client interface 2 A and the server 5 .
  • range request 1 is sent from the interface 2 A to the server 5 .
  • the server 5 begins data provision A. While the data provision A is being provided from the server 5 to the interface 2 A (and is not yet complete), a second request—range request 2 —is transmitted to the server 5 .
  • the server can immediately begin data provision B, and then data provision n, in response to range request n, etc., this ensures that the server is not idle.
  • transmission of the data is optimized.
  • FIG. 4 illustrates an example in which only one request is pipelined at a time. However, it is also possible that more than one request can be pipelined. Alternatively, a request can be sent before any response to a previous request is received. In an embodiment, the number of pipelined requests may be set to 1 (one “normal” request plus one “pipelined” request). In this embodiment, the optimal number of bytes to request is computed so that the range requested “fills up the pipe” or maximizes the bandwidth potential. In this way, there is almost always data being transmitted to the interface from the server during the transmission of the data resource. Alternatively, depending on the size of the bandwidth and the time delay, more than one pipelined request in the pipeline may be useful.
  • a method of pipelining requests is to pipeline a single request of X bytes, where X is the “number of bytes on flight,” (i.e. the amount of data that has been sent that has not yet been received).
  • the value X differs for each interface, because the X value is calculated as the bandwidth (e.g., bits/second)*the end-to-end delay (seconds) (bandwidth-delay product).
  • FIG. 5 illustrates the process whereby a plurality of interfaces are used in the pipeline process.
  • the client interface 2 A begins the process with range request 1 which is received by the server 5 .
  • the server 5 then begins data provision 2 A.
  • client interface 2 B sends range request 2 which is then in the pipeline.
  • the server 5 can immediately begin sending data provision 2 B.
  • This process continues with range requests 3 , 4 , . . . n and data provisions 2 B . . . 2 n , and 1 B . . . 1 n .
  • the client interfaces 2 A, 2 B, . . . 2 n , etc. are able to connect to the server 5 simultaneously and each send a request and each pipeline zero or more range requests.
  • HTTP pipelining is a method that allows a client to make multiple requests without waiting for each response (pipelined requests 61 ), allowing a single TCP connection to be used much more efficiently, with much lower elapsed time 63 .
  • each range retrieval request 64 is sequentially handled by the server before the client can send the next request (sequential requests 60 ).
  • an average time overhead of one round-trip time is incurred 62 .
  • this overhead significantly impairs the throughput of high-latency links (such as HSDPA).
  • HTTP pipelining eliminates the time overhead incurred by a sequential processing of requests and replies.
  • the optimal point in time for sending the next request may be determined.
  • Sending a range request too late will degrade the performance of pipelining, while sending the range request too early over the same link without having sent an intermediary request over a different interface increases the probability of two consecutive segments being transferred or sent by the same network interface, effectively resulting in double-size segments and one superfluous request.
  • One way to address this issue is to use adaptive segment sizes which tailor the segment size to the available bandwidth of a network interface. Such a configuration will avoid the superfluous request as the segment size will be determined to effectively utilize the available bandwidth.
  • Another way to address the timing issue is to schedule a predefined pipelining pattern during the startup phase. FIG.
  • FIG. 8 illustrates an example of a pipelining pattern during the startup phase.
  • the immediate requests 82 between “byte 0 ” 80 and “byte (n-segment size)” 81 are in an interleaved pattern and the pipelined requests 83 between “byte (n-segment size)” 81 and the end of the file are also in an interleaved pattern.
  • Requesting segments over interfaces 2 A, 2 B . . . 2 n in an interleaved pattern helps to provide a smooth playback, consistent sequential data reception, and reduces the initial response time.
  • each interface may send the range requests independently of the other interfaces (e.g. based on the receipt of data at each interface).
  • the server 5 requires 33 ms (i.e., 10 KB/300 KB/s) until the last byte of the segment is sent away.
  • the time t req is ⁇ 77 ms.
  • the amount of pipelined data can be determined such that it always exceeds the path's bandwidth-delay product, which can be addressed, at least, by increasing the request range size or by pipelining multiple segments.
  • the performance gain of multi-segment pipelining can be significant as is shown in an experimental example in FIG. 9 (throughput vs. segment size). While single-segment pipelining suffers from a rapid decrease in throughput for segment sizes below the bandwidth-delay product, multi-segment pipelining achieves close to optimal throughput with smaller range request sizes.
  • distribution of load (based on determined capacity) can be performed on two different bases.
  • the number of pipelined requests over each link may be limited, i.e. no additional requests are sent until after a certain number of requests have been responded to.
  • the number of pipelined requests does not have to be changed during transfer.
  • this embodiment may be implemented by initially, in the startup phase, 1) sending two consecutive (interleaved) requests per interface (see FIG. 8 ). Then based on received data from the server, 2) estimating delay and bandwidth. Once a segment corresponding to the first requested range has fully arrived, 3) generating and sending a third range request based on the estimate. The process then returns to step 2) where the delay and bandwidth estimate are again performed based on the data received from the server. A new request is again generated in step 3) using the updated delay and bandwidth estimate. This process continues looping between steps 2) and 3) until the process is complete.
  • a minimum delay (lower time limit) between requests may be enforced to prevent too many requests from being pipelined.
  • capacity for a link falls below a certain threshold (due to the link failing, or other significant reduction in link capacity etc.) the lower time limit for that link can be increased.
  • link capacity is significantly increased (due to reduction in other traffic etc) the lower time limit for that link can be reduced.
  • a maximum delay (upper time limit) can also be enforced.
  • the upper time limit ensures that the server always has access to a range request for each link. However, if capacity is reduced, the time limit can be increased (or be overridden by the lower time limit). Conversely, if capacity or server response improves the upper time limit can be reduced.
  • the invention can be implemented such that the distribution of the load can be performed based on the number of requests sent over each link (embodiment A) and/or based on the size of range requested over each link (embodiment B).
  • embodiments which regulate when requests are sent to the server such as by limiting the number of requests pipelined, or implementing a minimum time limit, can be used with either or both embodiments A and B.
  • the maximum delay (upper time limit) can be implemented with either or both embodiments A and B as well as embodiments in which the number or requests in the pipeline is limited and in which a minimum delay is implemented.
  • FIG. 10 illustrates experimental results from a direct performance comparison of a single WLAN interface versus using an additional HSDPA interface.
  • a 50 MB large video file is downloaded over multiple interfaces and, after a short startup delay of 3 seconds, played out at constant bitrate.
  • the present example compares pure success vs. any type of failure.
  • Each occurrence of the constant-bitrate file reader seeking ahead of the received data is counted as a failure.
  • a success is only recorded when the entire file was received and played back without any need for data re-buffering. As is shown in FIG.
  • the range requests sent from the client interfaces 2 A, 2 B . . . 2 n to the server 5 each include a data range identified as part of or retrievable from the resource.
  • the resource data provided by the server 5 is time dependent like video data or VoIP data, delay may become an issue.
  • Progressive downloading describes the method of sequentially transferring a media file from a server, allowing data playback on the client while the download is still in progress.
  • a particular property of progressive downloading is the use of client-side buffering for achieving smooth playback during periods of congestion. If the buffer is emptied too fast, the playback pauses until sufficient data is available again.
  • a startup latency for filling the buffer prior to playback is therefore required to reduce the chance of an interrupted video experience.
  • the tolerable startup latency is subjective and dependent on the media content, but may be as small as a few seconds.
  • Progressive downloading can be utilized over multiple network interfaces such as those described in the present invention. Moreover, such a configuration enables higher quality multimedia streams and is more robust against link variances, congestion and failures.
  • data received at the client may not be linear with respect to the playback order.
  • data that is received over one interface may be blocked by gaps because the client is still waiting for data with an earlier playback and requested over a different interface.
  • data that is received and available, but not ready for playback can be buffered by the client.
  • buffered data is made available for playback after a startup latency. This enables a playback bitrate substantially equivalent to the total aggregated throughput.
  • FIG. 11 illustrates an example snapshot of the received byte ranges during a progressing download. The shaded areas represent currently received data.
  • FIG. 11 includes a finished segment 110 and several unfinished segments 111 . The unfinished segment which is immediately after the finished segment 110 forms the data ready for playback 112 . However, data that is received but not ready for playback 113 may also exist in such a system. Thus, data that follows a gap is unready for playback and is therefore buffered.
  • the amount of data which is buffered can be determined based on the desired playback bitrate.
  • the desired playback bitrate can be determined dynamically during the data transfer. During a live data transfer, network conditions may vary. Thus, optimal numbers may be calculated up to the current time.
  • the “desired” playback bitrate may be equal to the current average aggregated throughput. For example, when a 6 Mbit/s download throughput is currently being achieved, the corresponding playback “quality” may be 6 Mbit/s.
  • the playback bitrate can dynamically adapt to the current throughput conditions (for example, by switching between resources of different quality levels or by utilizing a scalable video codec).
  • elements such as start-up latency and buffer size can also be determined from historical data. For example, once a file download with a given segment size has completed, the average aggregated throughput can be determined using the client and/or the server by dividing the total file size by the transfer duration. The average aggregated throughput can be determined by either or both the client and the server as both have used approximately the same time to receive/send the same amount of data.
  • the optimal startup latency and the required buffer size can be determined based on the history of received bytes.
  • the number of bytes that are ready for playback can be determined by the client at any given point in time (bytes_ready). Knowing the exact received byte ranges, the amount of data that must be buffered can also be determined by the client (bytes_unready).
  • the required startup latency 120 is illustrated as a race between 122 ( a ) the number of bytes ready for playback and 122 ( b ) the desired constant-bitrate playback. The startup latency 120 thus compensates for the maximum difference between the two pointers 122 ( a ) and 122 ( b ).
  • the required buffer size can generally be understood as the maximum capacity needed to store data that was received during the startup waiting time.
  • the buffer size is additionally influenced by received data that is not in correct playback order.
  • the buffer occupancy at a given point in time is the number of received but unready bytes plus the number of playback-ready bytes needed to guarantee smooth playback.
  • the maximum buffer requirements buf_req can thus be calculated as: buf_req max t (bytes_unready(t)+bytes_needed(t)).
  • One advantage of downloading a file over multiple network interfaces in parallel is the expected throughput gain compared to downloading over a single network interface.
  • each link should transfer data at its full capacity at all times; idle periods should be kept as rare as possible.
  • the size of the range request is important.
  • the utilization of the links is often interrupted due to frequent requests to the server. This causes a significant loss in bandwidth aggregation efficiency, which indicates that there exists a tradeoff between the segmentation overhead and the maximum achievable throughput. If the segmentation unit chooses too small a segment size, this increases the overhead and therefore reduces the aggregation efficiency.
  • the use of too large a range size is problematic for at least two reasons. First, there is the last segment problem, which is most obvious when a file is divided in two segments. Since the networks are heterogeneous, one interface will finish first and remain idle until the entire download has finished. Second, the use of too large a segment size creates a significant increase in the required startup latency and buffer requirements.
  • an optimal segment size must exist that maximizes the aggregated throughput.
  • the optimal segment size depends primarily on the delay heterogeneity of the used links. Therefore, the client may also dynamically adjust the segment size. Such a dynamic adjustment can be made based on link monitoring feedback (throughput and/or delay).
  • the required waiting time before the playback of an on-demand video is an important factor for the user-perceived quality of service. For example, when starting a full-length movie of two hours, users may be willing to wait a minute before show time, but for shorter clips, the startup should be perceived as instantaneous.
  • the optimal startup latency may be unknown in advance. Additionally, the required startup latency depends on the desired playback bitrate, which is bounded by the aggregated throughput over the combined links. If the desired playback bitrate is well below the aggregated throughput, then the startup latency can be reduced. However, in the present invention, the multiple links enable a playback bitrate that approaches the average aggregated throughput.
  • FIG. 14 illustrates two examples showing the startup latency which results in a playback bitrate equal to the average aggregated throughput.
  • file segmentation plays an important role.
  • segment size is small, it takes less time for transferring a full segment, which reduces the startup latency. From an end user's perspective, the choice of segment size equates to a balance between a shorter startup latency and a higher playback quality.
  • the startup delay is also dependent on the throughput heterogeneity of the links.
  • the startup phase itself can be used for link monitoring purposes.
  • the data is downloaded and not yet played back.
  • the data is buffered either because the first received bytes cannot be played back, due to the bytes not making up a complete display entity, such as a video frame, or because the first bytes received, while forming a complete video frame, correspond to a frame which is too short to be shown due to quality concerns.
  • a short startup latency is introduced to gather enough data to achieve a smoother playback.
  • the startup latency at the client can be reduced by using a TCP-like slow growth in segment sizes.
  • the maximum expected buffer size increases with the range size determined by the client and is in close correlation with the required startup latency.
  • the only major difference between the two metrics is their dependence on the number of used interfaces. As the startup latency is dominated by the slowest link used, the aggregation of additional faster interfaces will not have any significant impact on the startup latency.
  • the buffer capacity increases with the number of used network interfaces, because each interface causes additional out-of-order data to be received and buffered.
  • FIG. 15 illustrates the hardware structure of, for example, the client 1 according to an embodiment of the invention.
  • the client 1 includes a bus B or other communication mechanism for communicating information, and a processor/CPU 1004 coupled with the bus B for processing the information.
  • the client 1 also includes a main memory/memory unit 1003 , such as a random access memory (RAM) or other dynamic storage device (e.g., dynamic RAM (DRAM), static RAM (SRAM), and synchronous DRAM (SDRAM)), coupled to the bus B for storing information and instructions to be executed by processor/CPU 1004 .
  • the memory unit 1003 may be used for storing temporary variables or other intermediate information during the execution of instructions by the CPU 1004 .
  • the client 1 may also further include a read only memory (ROM) or other static storage device (e.g., programmable ROM (PROM), erasable PROM (EPROM), and electrically erasable PROM (EEPROM)) coupled to the bus B for storing static information and instructions for the CPU 1004 .
  • ROM read only memory
  • PROM programmable ROM
  • EPROM erasable PROM
  • EEPROM electrically erasable PROM
  • the client 1 may also include a disk controller coupled to the bus B to control one or more storage devices for storing information and instructions, such as mass storage 1002 , and drive device 1006 (e.g., floppy disk drive, read-only compact disc drive, read/write compact disc drive, compact disc jukebox, tape drive, and removable magneto-optical drive).
  • the storage devices may be added to the client 1 using an appropriate device interface (e.g., small computer system interface (SCSI), integrated device electronics (IDE), enhanced-IDE (E-IDE) also known as Parallel ATA (PATA), Serial ATA (SATA), direct memory access (DMA), or ultra-DMA).
  • SCSI small computer system interface
  • IDE integrated device electronics
  • E-IDE enhanced-IDE
  • PATA Parallel ATA
  • SATA Serial ATA
  • DMA direct memory access
  • ultra-DMA ultra-DMA
  • the client 1 may also include special purpose logic devices (e.g., application specific integrated circuits (ASICs)) or configurable logic devices (e.g., simple programmable logic devices (SPLDs), complex programmable logic devices (CPLDs), and field programmable gate arrays (FPGAs)).
  • ASICs application specific integrated circuits
  • SPLDs simple programmable logic devices
  • CPLDs complex programmable logic devices
  • FPGAs field programmable gate arrays
  • the client 1 may also include a display controller coupled to the bus B to control a display, such as a cathode ray tube (CRT) or a liquid crystal display (LCD), for displaying information to a computer user.
  • a display such as a cathode ray tube (CRT) or a liquid crystal display (LCD)
  • the computer system includes input devices, such as a keyboard and a pointing device, for interacting with a computer user and providing information to the processor.
  • the pointing device for example, may be a mouse, a trackball, or a pointing stick for communicating direction information and command selections to the processor and for controlling cursor movement on the display.
  • a printer may provide printed listings of data stored and/or generated by the computer system.
  • the client 1 performs at least a portion of the processing steps of the invention in response to the CPU 1004 executing one or more sequences of one or more instructions contained in a memory, such as the memory unit 1003 .
  • a memory such as the memory unit 1003 .
  • Such instructions may be read into the memory unit from another computer readable medium, such as the mass storage 1002 or a removable media 1001 .
  • One or more processors in a multi-processing arrangement may also be employed to execute the sequences of instructions contained in memory unit 1003 .
  • hard-wired circuitry may be used in place of or in combination with software instructions. Thus, embodiments are not limited to any specific combination of hardware circuitry and software.
  • the client 1 includes at least one computer readable medium 1001 or memory for holding instructions programmed according to the teachings of the invention and for containing data structures, tables, records, or other data described herein.
  • Examples of non-transitory computer readable media are compact discs, hard disks, floppy disks, tape, magneto-optical disks, PROMs (EPROM, EEPROM, flash EPROM), DRAM, SRAM, SDRAM, or any other magnetic medium, compact discs (e.g., CD-ROM), or any other medium from which a computer can read.
  • the present invention includes software for controlling the client 1 , for driving a device or devices for implementing the invention, and for enabling the client 1 to interact with a human user.
  • software may include, but is not limited to, device drivers, operating systems, development tools, and applications software.
  • Such computer readable media further includes the computer program product of the present invention for performing all or a portion (if processing is distributed) of the processing performed in implementing the invention.
  • the computer code elements on the medium of the present invention may be any interpretable or executable code mechanism, including but not limited to scripts, interpretable programs, dynamic link libraries (DLLs), Java classes, and complete executable programs. Moreover, parts of the processing of the present invention may be distributed for better performance, reliability, and/or cost.
  • Non-volatile media includes, for example, optical, magnetic disks, and magneto-optical disks, such as the mass storage 1002 or the removable media 1001 .
  • Volatile media includes dynamic memory, such as the memory unit 1003 .
  • Various forms of computer readable media may be involved in carrying out one or more sequences of one or more instructions to the CPU 1004 for execution.
  • the instructions may initially be carried on a magnetic disk of a remote computer.
  • An input coupled to the bus B can receive the data and place the data on the bus B.
  • the bus B carries the data to the memory unit 1003 , from which the CPU 1004 retrieves and executes the instructions.
  • the instructions received by the memory unit 1003 may optionally be stored on mass storage 1002 either before or after execution by the CPU 1004 .
  • the client 1 also includes a communication interface 1005 coupled to the bus B.
  • the communication interface 1004 provides a two-way data communication coupling to a network that is connected to, for example, a local area network (LAN), or to another communications network such as the Internet.
  • the communication interface 1005 may be a network interface card to attach to any packet switched LAN.
  • the communication interface 1005 may be an asymmetrical digital subscriber line (ADSL) card, an integrated services digital network (ISDN) card or a modem to provide a data communication link to a corresponding type of communications line.
  • Wireless links may also be implemented.
  • the communication interface 1005 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.
  • the network typically provides data communication through one or more networks to other data devices.
  • the network may provide a link to another computer through a local network (e.g., a LAN) or through equipment operated by a service provider, which provides communication services through a communications network.
  • the local network and the communications network use, for example, electrical, electromagnetic, or optical signals that carry digital data streams, and the associated physical layer (e.g., CAT 5 cable, coaxial cable, optical fiber, etc).
  • the network may provide a link to a mobile device such as a personal digital assistant (PDA) laptop computer, or cellular telephone.
  • PDA personal digital assistant

Abstract

A process for requesting information, the process including generating a plurality of range requests using a client having a plurality of network interfaces, the range requests each corresponding to a data range of a data resource, transmitting, from the client having the plurality of network interfaces, a plurality of range requests to a server via the plurality of network interfaces, receiving, at the client and from the server, a plurality of segments via the plurality of network interfaces such that each segment is received via a network interface which transmitted the corresponding range request, each segment including a portion of the data resource, and reassembling the data resource using the plurality of segments.

Description

    BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to a system and method of using range requests and pipelining to segment a data resource and transfer it over multiple network interfaces.
  • 2. Discussion of the Background
  • Even though user devices are often equipped with multiple network interfaces and in reach of several access networks at the same time, data to or from the Internet is normally communicated over only one of the access networks at a time. Also, even when a user device establishes multiple links to multiple access networks using its network interfaces, data to or from a particular data resource on the Internet is normally communicated over only one network link at a time. The result is that only some of the bandwidth that is theoretically available is actually used when content is delivered to a device. A multilink solution, where content is delivered over several network interfaces simultaneously would help remedy this issue. However, a major hurdle in the deployment of a multilink solution is the lack of server-side support. Although there have been suggested modifications to TCP and SCTP, standard transport protocols are unable to provide host-based bandwidth aggregation. For example, TCP is purely end-to-end and allows only one IP address (one interface) on each side. SCTP supports multiple IP addresses per endpoint, but uses them only for failover. A common approach is therefore to provide specialized libraries that transparently partition application-layer data into multiple transport streams. However, the implementation of such middleware requires software modifications to clients and servers.
  • SUMMARY OF THE INVENTION
  • Accordingly, the present invention provides, inter alia, a method of requesting information. The method includes a step of generating a plurality of range requests using a client having a plurality of network interfaces, the range requests each corresponding to a data range of a data resource. Also included in the method are the steps of transmitting, from the client having the plurality of network interfaces, a plurality of range requests to a server via the plurality of network interfaces, and receiving, at the client and from the server, a plurality of segments via the plurality of network interfaces such that each segment is received via a network interface which transmitted the corresponding range request, each segment including a portion of the data resource. Also included in the method is the step of reassembling the data resource using the plurality of segments.
  • In an embodiment of the invention, the method further includes a step of generating, using the client having the plurality of network interfaces, at least one additional range request, each additional range request corresponding to a data range of the data resource, a step of determining an upper time limit for each at least one additional range request, a step of transmitting, from the client having a plurality of network interfaces, the at least one additional range request to the server via at least one of the plurality of network interfaces before the determined upper time limit, a step of receiving, at the client and from the server, a plurality of additional segments via the plurality of network interfaces such that each additional segment is received via a network interface which transmitted the corresponding additional range request, each additional segment including a portion of the data resource, and a step of reassembling the data resource using at least the additional plurality of segments.
  • The present invention also includes a method of transferring information. The method includes a step of segmenting, using a client having a plurality of network interfaces, a data resource into a plurality of segments based on metadata of the data resource. Also included in the method are the steps of generating a plurality of range requests, the range requests each including a segment of the plurality of segments which corresponds to a data range of the data resource and transmitting, from the client having the plurality of network interfaces, the plurality of range requests to a server via the plurality of network interfaces for reassembly of the data resource using the plurality of segments at the server.
  • Also included in the present invention is a device for requesting information. The device includes a plurality of network interfaces and a generation unit that generates a plurality of range requests, the range requests each corresponding to a data range of a data resource. Also included in the device is a communications unit that transmits the plurality of range requests to a server via the plurality of network interfaces, the server segmenting the data resource based on the range requests into a plurality of segments, where the communications unit further receives the plurality of segments, which are transmitted by the server, via the plurality of network interfaces such that each segment is received via a network interface which transmitted the corresponding range request. Moreover, the device includes a reassembly unit that reassembles the data resource using the plurality of segments received by the communications unit.
  • It is to be understood that both the foregoing general description of the invention and the following detailed description are exemplary, but are not restrictive, of the invention.
  • BRIEF DESCRIPTION OF SEVERAL VIEWS OF THE DRAWINGS
  • Other objects, features and advantages of the present invention will become more apparent from the following detailed description when read in conjunction with the accompanying drawings, in which:
  • FIG. 1A illustrates a system according to an embodiment of the present invention including a client having multiple interfaces which communicate multiple range requests to a server;
  • FIG. 1B shows a block diagram illustrating an exemplary structure of the client according to an embodiment of the present invention;
  • FIG. 1C shows a flow diagram illustrating a method according to an embodiment of the present invention;
  • FIG. 1D shows a block diagram illustrating an exemplary structure of the server according to an embodiment of the present invention;
  • FIG. 1E shows a flow diagram illustrating a method according to an embodiment of the present invention;
  • FIG. 2A illustrates a system according to another embodiment of the present invention including a client having multiple interfaces which communicate multiple range requests to a server;
  • FIG. 2B illustrates a system according to another embodiment of the present invention including a client having multiple interfaces which communicate multiple data bearing range requests to a device;
  • FIG. 2C shows a block diagram illustrating an exemplary structure of the client according to another embodiment of the present invention;
  • FIG. 2D shows a flow diagram illustrating a method according to an embodiment of the present invention;
  • FIG. 3 illustrates a data resource according to an embodiment of the present invention that has multiple sections each corresponding to a range of the data resource;
  • FIG. 4 illustrates a pipelining process according to an embodiment of the present invention, the pipelining distributing multiple requests from the client interface to the server;
  • FIG. 5 illustrates a pipelining process according to another embodiment of the present invention, the pipelining distributing multiple requests from a plurality of client interfaces to the server;
  • FIG. 6 shows a time diagram illustrating comparison between the time elapsed in a sequential process and a pipelined process;
  • FIG. 7 illustrates experimental results obtained by downloading a file using two different interfaces;
  • FIG. 8 illustrates a pipelining pattern when executing pipelining range requests using multiple interfaces during the start-up phase;
  • FIG. 9 shows the performance gain of multi-segment pipelining in an exemplary embodiment;
  • FIG. 10 illustrates an exemplary embodiment providing results from a performance comparison of using a single interface versus using multiple interfaces;
  • FIG. 11 illustrates the received byte ranges during a progressing download according to an exemplary embodiment of the present invention;
  • FIG. 12 illustrates an example of a startup latency during which data is downloaded but not yet played back at the desired playback bitrate according to an embodiment of the present invention;
  • FIG. 13 illustrates a comparison of the throughput gain when using multiple interfaces and the throughput when using the fastest interface alone;
  • FIG. 14 illustrates a comparison of the start-up latency at various segment sizes between two exemplary embodiments of the claimed invention; and
  • FIG. 15 shows a block diagram illustrating a hardware configuration of a device of the present invention.
  • DETAILED DESCRIPTION OF THE INVENTIONS
  • Referring now to the drawings wherein like reference numbers designate identical or corresponding parts throughout the several views and more particularly to FIG. 1A thereof, there is illustrated a data transmission system including a client 1 and a server 5 which are connected via the internet 10. It should be noted that the invention is not limited such that the link between the client 1 and server 5 must be over the internet. The link between the client 1 and the server 5 can be over other types of network links such as via a body area network (BAN), personal area network (PAN), local area network (LAN), storage area network (SAN), campus area network (CAN), metropolitan area network (MAN), wide area network (WAN), wireless wide area network (WWAN), or public switched telephony network (PSTN). For example, the link can be a WWAN link compatible with a communication standard selected from the set of: Global System for Mobile communications (GSM), High Speed Packet Access (HSPA), Wideband Code Division Multiple Access (W-CDMA), Universal Mobile Telecommunications System (UMTS), Long Term Evolution (LTE), Worldwide interoperability for Microwave access (WiMax), Evolution-Data Optimized (EV-DO).
  • As is shown in FIG. 1B, the client 1 includes a plurality of interfaces 2A, 2B, . . . 2 n. In addition, the client 1 includes a generation unit 15 and a communication unit 16. The generation unit 15 (using a combination of hardware and software) generates a plurality of range requests 4A, 4B, . . . 4 n. The range requests 4A, 4B, . . . 4 n each correspond to a data range of a data resource 6. The communication unit 16 then transmits the plurality of range requests 4A, 4B, . . . 4 n to the communication unit 23 of the server 5 via the plurality of network interfaces 2A, 2B, . . . 2 n. In response, the server 5 segments or splits, based on the range requests 4A, 4B, . . . 4 n and using the segmentation unit 24 shown in FIG. 1D, the data resource 6 into a plurality of segments which are transmitted back using the communication unit 23 to the communication unit 16 of the client 1 via the network interface 2A, 2B, . . . 2 n which transmitted the corresponding range request 4A, 4B, . . . 4 n. Before segmentation, the data resource 6 may be stored in the storage unit 25 of the server 5. Additionally, before segmentation, the data resource 6 may be computed on request by the server 5 or retrieved from some other source of data.
  • Duplicates of the data resource 6 may also be stored on or obtained by multiple servers. Such a configuration enables the client 1 to obtain portions of the data resource 6 from different servers. In an embodiment, this can be accomplished by the client using metalinks to obtain the data resource 6. For example, if the client knows that URI-1 and URI-2 are copies on different servers, range requests could be made to both servers. Thus, for example, a first portion of the plurality of range requests 4A, 4B, . . . 4 n could be transmitted to a first server, a second portion of the plurality range requests 4A, 4B, . . . 4 n could be transmitted to a second server, and an nth portion of the plurality of range requests 4A, 4B, . . . 4 n could be transmitted to an nth server.
  • As is illustrated in FIG. 1A, interfaces 2A, 2B, . . . 2 n connect the client 1 to the server 5. In addition, in an embodiment, the interfaces 2A, 2B, . . . 2 n each connect to a respective distinct ISP 3A, 3B, . . . 3 n or network provider by which the range requests 4A, 4B, . . . 4 n and respective data provision responses are sent.
  • FIG. 1C illustrates steps performed by the client 1 when obtaining the data resource 6 from the server 5. In an embodiment of the invention, the client 1 generates (S1) a plurality of range requests 4A, 4B, . . . 4 n, the plurality of range requests 4A, 4B, . . . 4 n each corresponding to a data range of the data resource 6. The client 1 then transmits (S2) the plurality of range requests 4A, 4B, . . . 4 n to the server 5 via the plurality of network interfaces 2A, 2B, . . . 2 n. The client then receives (S3) the plurality of segments via the plurality of network interfaces 2A, 2B, . . . 2 n such that each segment is received via the network interface which transmitted the corresponding range request, each segment including stored therein a portion of the data resource 6. Then, the client 1 reassembles (S4) the data resource 6 which can be displayed, stored, etc. by the client 1.
  • FIG. 1E illustrates steps performed by the server 5 when transferring the data resource 6 to the client 1. In response to receiving (S10) a plurality of range requests 4A, 4B, . . . 4 n, the server 5 obtains (S11) the data resource 6. The data resource may be obtained from the storage unit 25 of the server 5, may be generated by the server 5 or may be obtained from an external source. The server 5 then segments (S12) the obtained data resource 6 into a plurality of segments based on the range requests 4A, 4B, . . . 4 n and transmits (S13) the plurality of segments to the client 1.
  • FIG. 2A illustrates an alternate embodiment in which the interfaces 2A and 2B are connected to the same ISP 3A. Thus, in the embodiment shown in FIG. 2A, although the interfaces 2A and 2B send separate range requests 4A,4B and receive separate data provision responses, the network path, at least part of the way between the client 1 and the server 5, is by way of the same ISP 3A or same network provider.
  • One result of the use of multiple interfaces is the increase in bandwidth on a client device 1 which has access to, and is able to connect to, one or more networks over more than one interface 2A, 2B, . . . 2 n. One example of a device which can utilize this configuration is a mobile phone (client) which has access to the Internet directly or indirectly using a WiMax link, a 3G link, a WiFi link, a cable, and/or Bluetooth, for example. Another example is a laptop computer having an Ethernet cable interface as well as WiFi interface. The various interfaces may use different protocols on the physical and link layers, but will typically use the same protocols on the network layer and the transport layer, for example TCP/IP or UDP/IP. Each network interface can establish an end-to-end communication link using for example: Transmission Control Protocol (TCP), Stream Control Transmission Protocol (SCTP), Datagram Congestion Control Protocol (DCCP), User Datagram Protocol (UDP, which is connectionless), etc. In addition, the end-to-end communication link can be one of Hypertext Transfer Protocol (HTTP), File Transfer Protocol (FTP), Session Initiation Protocol (SIP), Internet Message Access Protocol (IMAP), Session Description Protocol (SDP), etc.
  • In one example, the multiple range requests are created from one request referencing an entire amount of data. The ranges represent subsections of the entire amount of data. For example, the first single request is created by a browser, and the corresponding multiple requests are created by a middleware on the client that distributes the request over the multiple interfaces.
  • Thus, by utilizing multiple local network interfaces (being of the same or different types) to increase available bandwidth, the efficiency of concurrent utilization of multiple network interfaces is improved. In addition the efficiency of the system is further increased by using range requests to subdivide the data resource 6 such that individual data resources (e.g. a file) can be distributed over the multiple links. The data resources can be a file, data produced from a scanner, a web camera, a microphone, video, audio, etc. As is shown in FIG. 1A, the server 5 includes a data file 6 which is accessed by the client 1. In addition, as is illustrated in FIG. 1A, the client 1 sends a plurality of range requests 4A, 4B, . . . 4 n using the plurality of interfaces 2A, 2B, . . . 2 n.
  • FIG. 2B illustrates an embodiment in which, instead of simply sending a range request, a plurality of interfaces 22A, 22B, . . . 22 n of the client 1 each send a request 21A, 21B, . . . 21 n which includes a range of some data resource, to a device 20. Thus, the system can utilize a server 5 or some other providing resource storing or generating data to be retrieved, such as a file server, a scanner, a web camera, etc. and can also utilize a device 20 which itself is or which hosts the providing resource which is capable of receiving data such as, for example, a file (e.g. on a file server), a printer, a program or script capable of processing received data.
  • Moreover, in an embodiment the data resource or the providing resource is identified by a URI (uniform resource identifier), the URI being a string of characters used to identify a name or a resource. Both the data resource and the providing resource can be identified by a URI (i.e. a URL or a URN) as a URI can identify a device, data or something else, depending on circumstances.
  • FIG. 2C illustrates the structure of the client 1 according to the embodiment of the invention described in FIG. 2B. The client 1 shown in FIG. 2C includes a plurality of interfaces 2A, 2B, . . . 2 n. When transferring a data resource to a device 20, the client 1 shown in FIG. 2C uses the segmentation unit 18 to segment the data resource which is either stored in the storage 17, generated at by the client 1, or retrieved from an external source by the client 1, into a plurality of segments based on a current status of the plurality of network interfaces 2A, 2B, . . . 2 n and/or the size of the data resource. The client 1 also generates, using the generation unit 15, a plurality of range requests 4A, 4B, . . . 4 n each including a segment of the plurality of segments which corresponds to a data range of the data resource. The segmentation may be performed before any portion of the data resource is transferred or may be performed while transfer of the data resource is in progress based on measured results of previously transmitted segments (measured bandwidth, round trip delay, etc.). The client 1 then uses the communication unit 16 to transmit the plurality of range requests 4A, 4B, . . . 4 n including data resource segments to the device 20 via the plurality of network interfaces 2A, 2B, . . . 2 n. The client 1 may also transmit the plurality of range requests 4A, 4B, . . . 4 n including data resource segments to a server 5.
  • FIG. 2D illustrates steps performed by the client 1 and the server 5 when transferring the data resource 6 from the client 1 to the device 20. In an embodiment of the invention, the client 1 segments (S20) a data resource into a plurality of segments. The client then generates (S21) a plurality of range requests 4A, 4B, . . . 4 n each including a segment of the plurality of segments which corresponds to a data range of the data resource 6. The plurality of range requests are then transmitted (S22) from the client 1 to the device 20 which receives the transmitted range requests (S23). Optionally, the device 20 then transmits (S24) an acknowledgment for each received range request to the client 1.
  • FIG. 3 illustrates an example of a data resource (data file 6, for example) which is accessed by a plurality of range requests. Each range request corresponds to a portion of the data resource. Thus, when retrieving data, from a providing resource (server 5, Peer-to-peer device, element on a device, etc.), the range of data is a subset of data retrievable from the providing resource in response to requests, and an identification of a range includes at least a beginning (and optionally an end or an extent) of that range. The data represented by the range is expected to be received by the client device in response to the request and over the same network interface which transmitted the request.
  • When transmitting data, such as to the device 20, for example, the range of data is a subset of data stored, retrieved from an external device or created in the providing resource, an identification of a range includes at least a beginning (and optionally an end or an extent) of that range, and the data represented by the range is included as payload in the request. An expected response to the request is a received confirmation that the data was received and is handled according to the request.
  • For example, the beginning of the range (and optionally the end of the range) can be defined as a first byte position in a byte range, a time stamp, a page number, etc. In addition, the extent of the range can be defined as a number of bytes in a byte range, a duration of time (video, audio), a number of pages (document), a number of blocks (image), a number of frames (video), etc.
  • The invention can be implemented using existing protocols, provided that the protocol implementation at both the client and the server support ranges. For example, HTTP supports range requests. The invention is, however, not limited to existing protocols. For example, when using HTTP, the beginning and the end of the range has to be specified. The server will send all the specified bytes and cannot be interrupted (unless the underlying TCP connection is cancelled and restarted). However, the present invention allows ranges to have undefined extents (“open ends”), which is different from connections using the current http protocol.
  • A request, which includes the segmented data resource, may be uploaded using an HTTP PUT or POST-like request. (Note that the current version of HTTP does not support POST requests with ranges.)
  • Thus, when requesting data ranges, the response from the server is the data represented by the requested range and received over the same interface. When transmitting data ranges, the response from the device is a status report confirming (or not) that the transmitted data was received and processed according to the request.
  • In an embodiment, the size of the range for each request from each interface is dynamically determined and is based on certain factors. For example, the size of the range requests are determined such that they subdivide data payloads so that individual data resources (e.g. a file) can be distributed over several links in either predefined fixed sizes or dynamically determined by link capacity and/or delay.
  • In addition, the size of the range requests should be as small as possible but not so small that the available bandwidth is not optimally utilized, e.g. the size of the range should be larger than a bandwidth-delay product for the selected interface.
  • Small ranges provide a fine granularity which allows a “smooth”, “progressive” processing of the transferred resource. The smaller the segments, the quicker the system “adapts” to variations in transfer rate. Also, the startup latency (the waiting time that is required before the transferred resource can be processed at an aggregated rate corresponding to the rate obtained from combining multiple interfaces) directly relates to the size of the ranges: the smaller the range, the smaller the required startup waiting time. Thus, one disadvantage of large ranges is that the startup waiting time increases. Also, a large range increases the chance of “unsmooth” processing if there are variances in transfer rate. For example, if one interface unexpectedly drops to 50% of the speed, it suddenly takes twice as long to “fill up a gap of missing data in the transferred resource.” Such an issue is mitigated by using small ranges.
  • However, there can be a disadvantage to using too small of a range. For example, one disadvantage of a too small range is that the bandwidth is not optimally utilized. If less than (bandwidth *delay) bytes are requested (per interface), then the server is able to transfer an entire range before knowing what the next range should be. Therefore, the server will be idle, resulting in the bandwidth not being optimally utilized.
  • In an embodiment, the size of the range request for an interface is determined based, at least in part, by the bandwidth available at the interface. However, there is typically a lack of prior knowledge at the client regarding the bandwidth and round trip delay of the various links. As a result, the bandwidth and/or delay may be determined during initial transmission. For example, the bandwidth can be estimated by counting the number of transferred bytes per time unit. As capacity can change during data transfer, tracking and adapting to the changes in bandwidth/delay may be performed. For example, as a bandwidth capacity value may change over time, some averaging can be performed to optimize this value. The delay can be estimated by measuring the time between when a request is sent and an answer is received. Another way to determine the bandwidth, which can be used in certain circumstances, is to use feedback from the TCP stack, as TCP keeps estimates of the round-trip delay. An alternative approach is to probe the link with differently sized range requests which gives feedback on what the optimal range size should be based on current network conditions.
  • Further, in an embodiment, the size or particular range of the range requests can be determined based on obtaining metadata of the data resource. In an embodiment, the metadata of the data resource is obtained by transmitting a probing request and receiving the metadata in response. For example, transmitting an HTTP HEAD request (e.g. over the fastest network interface) using the same URI, and receiving at least one entity header field returned in response to the HTTP HEAD request. In another embodiment the metadata of the data resource is obtained by identifying the metadata from the plurality of segments received from the server. For example, identifying entity header fields received in the responses to the plurality of HTTP GET requests sent over the multiple network interfaces.
  • The metadata may be among other things, a content-length header field, a content-encoding header field, content-type header field, or a content-range entity header field.
  • For example, the metadata in a returned content-range header field may be used for determining a particular range of a range request. As a result, the client would not have to remember the value of previously requested ranges as these values would be returned in the content-range header field. The client could then generate a range request requesting a range consecutive to the immediately received range based on the metadata.
  • In addition, in an embodiment, the particular interface which sends a particular range request can be selected based on certain factors. For example, the network interface can be selected such that traffic is distributed efficiently. In an embodiment, the network interface is selected based on monitoring, at least a subset of, the status of the network interfaces, the response times for previous requests, the number of sent/received bytes, etc. The selection can be made prior to or after the creation of the request i.e. first a request is created and then a selection of an interface is made taking into account the nature of the request; or first a selection of an interface is made—for example because the interface becomes idle—and then a request is made, taking into account the capabilities and/or current status of the interface when creating the request by, for example, determining range size based on bandwidth-delay product, as described above.
  • With regard to the current status of the interface, an interface may be in one of several states including, idle, in the process of sending a request, in the process of waiting for a response, or in the process of receiving a response. The latter three states may occur simultaneously (e.g. the interface may be sending and receiving at the same time, or sending a new request while waiting for a response to a previous request). In order to determine when a response has been completely received, the requesting client may count the incoming bytes and compare them with the size of the requested range. Each response starts with a header that indicates that a new range is arriving.
  • If one of the interfaces goes down, delay may occur. In order to address this problem, the client by way of the communications unit, for example, is able to detect outages by monitoring the interfaces and rebalancing the load to an interface that is still working such that the data transfer can continue from the point at which the failure occurred. Any request which is only partially responded to would have to be sent again over one of the operating interfaces.
  • In addition, further efficiency can be obtained by pipelining requests, which means that one or more additional requests are sent down one link without waiting until a response to the previous request(s) has been received.
  • FIG. 4 illustrates an example of the pipelining process between client interface 2A and the server 5. In this example, range request 1 is sent from the interface 2A to the server 5. In response, the server 5 begins data provision A. While the data provision A is being provided from the server 5 to the interface 2A (and is not yet complete), a second request—range request 2—is transmitted to the server 5. Once the server 5 has finished executing data provision A, the server can immediately begin data provision B, and then data provision n, in response to range request n, etc., this ensures that the server is not idle. Thus, by having a range request in the “pipeline”, transmission of the data is optimized.
  • FIG. 4 illustrates an example in which only one request is pipelined at a time. However, it is also possible that more than one request can be pipelined. Alternatively, a request can be sent before any response to a previous request is received. In an embodiment, the number of pipelined requests may be set to 1 (one “normal” request plus one “pipelined” request). In this embodiment, the optimal number of bytes to request is computed so that the range requested “fills up the pipe” or maximizes the bandwidth potential. In this way, there is almost always data being transmitted to the interface from the server during the transmission of the data resource. Alternatively, depending on the size of the bandwidth and the time delay, more than one pipelined request in the pipeline may be useful.
  • Thus, by having range requests in the pipeline it is ensured that the server always is busy; always has a request waiting in the pipeline. It is not the number of requests in the pipeline, but the number of requested bytes in the pipeline that is important. Thus, the number of requests in the pipeline is also dependent on the size of the range requests. Nevertheless, if the range request is sufficiently large it is possible to include only one pipeline request per link. However, it is also possible to pipeline no request or more than one request per link.
  • In an embodiment, a method of pipelining requests is to pipeline a single request of X bytes, where X is the “number of bytes on flight,” (i.e. the amount of data that has been sent that has not yet been received). The value X differs for each interface, because the X value is calculated as the bandwidth (e.g., bits/second)*the end-to-end delay (seconds) (bandwidth-delay product).
  • FIG. 5 illustrates the process whereby a plurality of interfaces are used in the pipeline process. For example, in FIG. 5, the client interface 2A begins the process with range request 1 which is received by the server 5. The server 5 then begins data provision 2A. After the client interface 2A sends range request 1, client interface 2B sends range request 2 which is then in the pipeline. Once the server 5 has completed sending data provision 2A, the server can immediately begin sending data provision 2B. This process continues with range requests 3, 4, . . . n and data provisions 2B . . . 2 n, and 1B . . . 1 n. Alternatively, the client interfaces 2A, 2B, . . . 2 n, etc. are able to connect to the server 5 simultaneously and each send a request and each pipeline zero or more range requests.
  • One type of pipelining which can be used in the present invention is HTTP pipelining. This type of pipelining is a method that allows a client to make multiple requests without waiting for each response (pipelined requests 61), allowing a single TCP connection to be used much more efficiently, with much lower elapsed time 63. As depicted in FIG. 6, in the absence of pipelining, each range retrieval request 64 is sequentially handled by the server before the client can send the next request (sequential requests 60). Thus, for each request, an average time overhead of one round-trip time is incurred 62. For a large number of small segments (small range requests), this overhead significantly impairs the throughput of high-latency links (such as HSDPA). Thus, from a client-side perspective, HTTP pipelining eliminates the time overhead incurred by a sequential processing of requests and replies.
  • When pipelining is not enabled, the size of each of the range requests can be increased. However, when at least one request is pipelined, a request for the next byte range is sent before the currently downloading range has completed and thus much of the time overhead can be eliminated. Moreover, as is shown in FIG. 7, when using small range sizes, using HTTP pipelining helps to increase the aggregated throughput.
  • Thus, using a pipelining mechanism eliminates the waiting time between successive range requests and effectively increases the aggregated throughput. The experimental results shown in FIG. 7 were obtained by simultaneously downloading a 50 MB large file over HSDPA and WLAN interfaces.
  • In addition, when pipelining is enabled, the optimal point in time for sending the next request (illustrated with small triangles 64 in FIG. 6) may be determined. Sending a range request too late will degrade the performance of pipelining, while sending the range request too early over the same link without having sent an intermediary request over a different interface increases the probability of two consecutive segments being transferred or sent by the same network interface, effectively resulting in double-size segments and one superfluous request. One way to address this issue is to use adaptive segment sizes which tailor the segment size to the available bandwidth of a network interface. Such a configuration will avoid the superfluous request as the segment size will be determined to effectively utilize the available bandwidth. Another way to address the timing issue is to schedule a predefined pipelining pattern during the startup phase. FIG. 8 illustrates an example of a pipelining pattern during the startup phase. The immediate requests 82 between “byte 080 and “byte (n-segment size)” 81 are in an interleaved pattern and the pipelined requests 83 between “byte (n-segment size)” 81 and the end of the file are also in an interleaved pattern. Requesting segments over interfaces 2A, 2B . . . 2 n in an interleaved pattern helps to provide a smooth playback, consistent sequential data reception, and reduces the initial response time.
  • During the interleaved startup phase it is ensured that consecutive byte ranges are not requested over the same interface. After the interleaved startup phase, each interface may send the range requests independently of the other interfaces (e.g. based on the receipt of data at each interface).
  • In addition, as is noted above, when using small range requests to improve playback smoothness, there exists a risk of choosing a range request which is too small. In this case, the server spends so little time processing a request that the client does not have enough time to pipeline the next request. This issue arises when the transmission delay dn over interface 2 n is larger than the time tn it takes for the server 5 to send out a requested data segment, which is given by the segment size S and the average throughput Φn, experienced over interface 2 n: tn=S/Φn.
  • For example, in an exemplary embodiment, for a 10 KB segment over HSDPA, the server 5 requires 33 ms (i.e., 10 KB/300 KB/s) until the last byte of the segment is sent away. The problem here is that the 33 ms required to send out the data is much less than the average transmission delay dn=110 ms experienced over the HSDPA link. Therefore, in this case, pipelining is not as efficient as expected, because the server 5 has sent a segment and is available to send another one, but no range request has yet been pipelined. For pipelining to work efficiently over a path with delay dn, a new request should be sent from the client to the server at time treq after the current time tnow: treq=tnow+S/Φi−d.
  • In the example of sending 10 KB segments over HSDPA, the time treq is −77 ms. Thus, while the client is downloading a segment, the next segment should be requested before the current segment's download has even started. In other words, the segment size of 10 KB is too small. Thus, calculation of the minimal segment size Smin required for fully efficient pipelining, can be calculated by setting treg=0 and solving the equation treq=tnow+S/Φi−d for the segment size S: Smin=d*Φi. Thus, for example, the minimum segment size in the present example is therefore 110 ms*300 KB/s=30 KB, which is equivalent to the bandwidth-delay product of the path.
  • Thus, in order to alleviate the minimum segment size problem, the amount of pipelined data can be determined such that it always exceeds the path's bandwidth-delay product, which can be addressed, at least, by increasing the request range size or by pipelining multiple segments.
  • When using smaller range request sizes, the performance gain of multi-segment pipelining can be significant as is shown in an experimental example in FIG. 9 (throughput vs. segment size). While single-segment pipelining suffers from a rapid decrease in throughput for segment sizes below the bandwidth-delay product, multi-segment pipelining achieves close to optimal throughput with smaller range request sizes.
  • Thus, by pipelining multiple segments/requests, issues related to the minimum segment size can be addressed without the need for increased segment size and, as a result, a smooth playback can be achieved.
  • As is described above, distribution of load (based on determined capacity) can be performed on two different bases. First, the distribution can be performed by modifying the number of requests sent over each link. Second, the distribution can be performed by modifying the size of the range requested over each link.
  • However, due to the fact that the available capacity may dynamically change during transmission, the optimum distribution over the various links may not be known when requests are first sent. Therefore, it may be useful to delay sending client generated requests. Specifically, if the client were to create requests representing the entire resource and send them immediately, it would be difficult to properly take into consideration the capacities of the various links. Therefore, in order to more efficiently take into account changes in capacity, it may be useful to delay transmission such that each request is transmitted only a little bit ahead of the server becoming available. Such a delay can be accomplished a number of ways.
  • For example, the number of pipelined requests over each link may be limited, i.e. no additional requests are sent until after a certain number of requests have been responded to. The number of pipelined requests does not have to be changed during transfer. In addition, this embodiment may be implemented by initially, in the startup phase, 1) sending two consecutive (interleaved) requests per interface (see FIG. 8). Then based on received data from the server, 2) estimating delay and bandwidth. Once a segment corresponding to the first requested range has fully arrived, 3) generating and sending a third range request based on the estimate. The process then returns to step 2) where the delay and bandwidth estimate are again performed based on the data received from the server. A new request is again generated in step 3) using the updated delay and bandwidth estimate. This process continues looping between steps 2) and 3) until the process is complete.
  • Alternatively, a minimum delay (lower time limit) between requests may be enforced to prevent too many requests from being pipelined. In addition, if capacity for a link falls below a certain threshold (due to the link failing, or other significant reduction in link capacity etc.) the lower time limit for that link can be increased. Conversely, if link capacity is significantly increased (due to reduction in other traffic etc) the lower time limit for that link can be reduced.
  • In addition, a maximum delay (upper time limit) can also be enforced. The upper time limit ensures that the server always has access to a range request for each link. However, if capacity is reduced, the time limit can be increased (or be overridden by the lower time limit). Conversely, if capacity or server response improves the upper time limit can be reduced.
  • Thus, the invention can be implemented such that the distribution of the load can be performed based on the number of requests sent over each link (embodiment A) and/or based on the size of range requested over each link (embodiment B). Moreover, embodiments which regulate when requests are sent to the server such as by limiting the number of requests pipelined, or implementing a minimum time limit, can be used with either or both embodiments A and B.
  • Further, the maximum delay (upper time limit) can be implemented with either or both embodiments A and B as well as embodiments in which the number or requests in the pipeline is limited and in which a minimum delay is implemented.
  • In order to illustrate the advantages of using multiple interfaces and respective range requests, FIG. 10 illustrates experimental results from a direct performance comparison of a single WLAN interface versus using an additional HSDPA interface. In this example, a 50 MB large video file is downloaded over multiple interfaces and, after a short startup delay of 3 seconds, played out at constant bitrate. Rather than analyzing complex video metrics that depend on the type of content and codes used, the present example compares pure success vs. any type of failure. Each occurrence of the constant-bitrate file reader seeking ahead of the received data is counted as a failure. A success is only recorded when the entire file was received and played back without any need for data re-buffering. As is shown in FIG. 10, the simultaneous use of multiple interfaces increases the probability of an interruption-free playback at a given bit-rate. Another way to interpret these results is that the unreliable WLAN link, which performs poorly at any bitrates above 600 KB can be made 100% reliable by adding an additional HSDPA link.
  • As is noted above, the range requests sent from the client interfaces 2A, 2B . . . 2 n to the server 5 each include a data range identified as part of or retrievable from the resource. However, when the resource data provided by the server 5 is time dependent like video data or VoIP data, delay may become an issue.
  • One way to transport time-dependent data is by using progressive downloading. Progressive downloading describes the method of sequentially transferring a media file from a server, allowing data playback on the client while the download is still in progress. A particular property of progressive downloading is the use of client-side buffering for achieving smooth playback during periods of congestion. If the buffer is emptied too fast, the playback pauses until sufficient data is available again. A startup latency for filling the buffer prior to playback is therefore required to reduce the chance of an interrupted video experience. The tolerable startup latency is subjective and dependent on the media content, but may be as small as a few seconds.
  • Progressive downloading can be utilized over multiple network interfaces such as those described in the present invention. Moreover, such a configuration enables higher quality multimedia streams and is more robust against link variances, congestion and failures.
  • However, when using a progressive downloading type system, even though segments are requested sequentially over the available interfaces, data received at the client may not be linear with respect to the playback order. In particular, data that is received over one interface may be blocked by gaps because the client is still waiting for data with an earlier playback and requested over a different interface. In order to avoid data loss, data that is received and available, but not ready for playback, can be buffered by the client.
  • In an embodiment, buffered data is made available for playback after a startup latency. This enables a playback bitrate substantially equivalent to the total aggregated throughput. FIG. 11 illustrates an example snapshot of the received byte ranges during a progressing download. The shaded areas represent currently received data. FIG. 11 includes a finished segment 110 and several unfinished segments 111. The unfinished segment which is immediately after the finished segment 110 forms the data ready for playback 112. However, data that is received but not ready for playback 113 may also exist in such a system. Thus, data that follows a gap is unready for playback and is therefore buffered.
  • In an embodiment, the amount of data which is buffered can be determined based on the desired playback bitrate. The desired playback bitrate can be determined dynamically during the data transfer. During a live data transfer, network conditions may vary. Thus, optimal numbers may be calculated up to the current time. The “desired” playback bitrate may be equal to the current average aggregated throughput. For example, when a 6 Mbit/s download throughput is currently being achieved, the corresponding playback “quality” may be 6 Mbit/s. Moreover, the playback bitrate can dynamically adapt to the current throughput conditions (for example, by switching between resources of different quality levels or by utilizing a scalable video codec).
  • In an embodiment, elements such as start-up latency and buffer size can also be determined from historical data. For example, once a file download with a given segment size has completed, the average aggregated throughput can be determined using the client and/or the server by dividing the total file size by the transfer duration. The average aggregated throughput can be determined by either or both the client and the server as both have used approximately the same time to receive/send the same amount of data.
  • Taking into account a desired playback bitrate, the optimal startup latency and the required buffer size can be determined based on the history of received bytes.
  • The number of bytes that are ready for playback can be determined by the client at any given point in time (bytes_ready). Knowing the exact received byte ranges, the amount of data that must be buffered can also be determined by the client (bytes_unready).
  • In order to determine the minimum required startup latency, how many bytes are needed (bytes_needed) for achieving an uninterrupted and sustainable playback at the desired bitrate is calculated for every time t:
    bytes_needed(t)=bitrate*t−bytes_ready(t).
  • As depicted in FIG. 12, the required startup latency 120 corresponds to the maximum number of bytes needed 121 to satisfy the desired playback bitrate. It can be calculated as follows: startup_latency = max t ( bytes_needed ( t ) ) bitrate .
    Thus, in FIG. 12, the required startup latency 120 is illustrated as a race between 122(a) the number of bytes ready for playback and 122(b) the desired constant-bitrate playback. The startup latency 120 thus compensates for the maximum difference between the two pointers 122(a) and 122(b).
  • The required buffer size can generally be understood as the maximum capacity needed to store data that was received during the startup waiting time. However, in a scenario of the client receiving file segments over multiple links, the buffer size is additionally influenced by received data that is not in correct playback order. In other words, the buffer occupancy at a given point in time is the number of received but unready bytes plus the number of playback-ready bytes needed to guarantee smooth playback. The maximum buffer requirements buf_req can thus be calculated as: buf_req maxt(bytes_unready(t)+bytes_needed(t)).
  • The performance of progressive download when using a multihomed host (i.e. a client with multiple network interfaces) that is simultaneously connected to a WLAN and up to two independent HSPDA networks is considered below. The measured link characteristics between the used client and server are summarized in Table 1.
    TABLE 1
    OBSERVED CHARACTERISTICS OF USED LINKS
    WLAN HSDPA
    Maximum achievable 640 320
    throughput KB/s KB/s
    Average experienced 600 250
    throughput (Φ) KB/s KB/s
    Average RTT for header-only 20 ms 100 ms
    IP packets
    30 ms 220 ms
    Average RTT for full-size IP
    packets

    All the presented results are based on downloading a 25.2 MB large file from a geographically local web server (the path lengths from the client to the server are in the range of 7 to 9 IP hops).
  • One advantage of downloading a file over multiple network interfaces in parallel is the expected throughput gain compared to downloading over a single network interface. When simultaneously transferring data over n different links li (iε{1, . . . , n}) with average throughput Φ(li), an objective is to maximize the bandwidth aggregation efficiency n, which we define as: n = Φ ( l 1 + + l n ) Φ ( l 1 ) + + Φ ( l n ) .
  • In other words, for achieving ideal bandwidth aggregation, each link should transfer data at its full capacity at all times; idle periods should be kept as rare as possible.
  • FIG. 13 illustrates a comparison of the throughput gain of combined heterogeneous networks to the traditional method of using the fastest interface alone. For optimal segment sizes (around 1000 KB in these experiments) the aggregation efficiency is n=90%.
  • As was noted above, the size of the range request is important. When choosing too small a range size, the utilization of the links is often interrupted due to frequent requests to the server. This causes a significant loss in bandwidth aggregation efficiency, which indicates that there exists a tradeoff between the segmentation overhead and the maximum achievable throughput. If the segmentation unit chooses too small a segment size, this increases the overhead and therefore reduces the aggregation efficiency. On the other hand, the use of too large a range size is problematic for at least two reasons. First, there is the last segment problem, which is most obvious when a file is divided in two segments. Since the networks are heterogeneous, one interface will finish first and remain idle until the entire download has finished. Second, the use of too large a segment size creates a significant increase in the required startup latency and buffer requirements.
  • From the results presented in FIG. 13, it is evident that an optimal segment size must exist that maximizes the aggregated throughput. The optimal segment size depends primarily on the delay heterogeneity of the used links. Therefore, the client may also dynamically adjust the segment size. Such a dynamic adjustment can be made based on link monitoring feedback (throughput and/or delay).
  • The required waiting time before the playback of an on-demand video is an important factor for the user-perceived quality of service. For example, when starting a full-length movie of two hours, users may be willing to wait a minute before show time, but for shorter clips, the startup should be perceived as instantaneous.
  • Due to link variations, the optimal startup latency may be unknown in advance. Additionally, the required startup latency depends on the desired playback bitrate, which is bounded by the aggregated throughput over the combined links. If the desired playback bitrate is well below the aggregated throughput, then the startup latency can be reduced. However, in the present invention, the multiple links enable a playback bitrate that approaches the average aggregated throughput.
  • FIG. 14 illustrates two examples showing the startup latency which results in a playback bitrate equal to the average aggregated throughput. Not only for bandwidth aggregation, but also for the required startup latency, file segmentation plays an important role. When the segment size is small, it takes less time for transferring a full segment, which reduces the startup latency. From an end user's perspective, the choice of segment size equates to a balance between a shorter startup latency and a higher playback quality.
  • In addition to the segment size, the startup delay is also dependent on the throughput heterogeneity of the links. The larger the throughput difference between the slowest link and the others, the larger the amount of bytes needed to satisfy the desired playback rate (see bytes_needed in FIG. 12).
  • In order to address such issues a precise estimation of the startup latency may be used. For example the startup phase itself can be used for link monitoring purposes. When data is first received, the data is downloaded and not yet played back. The data is buffered either because the first received bytes cannot be played back, due to the bytes not making up a complete display entity, such as a video frame, or because the first bytes received, while forming a complete video frame, correspond to a frame which is too short to be shown due to quality concerns. Thus, a short startup latency is introduced to gather enough data to achieve a smoother playback.
  • In an alternate embodiment the startup latency at the client can be reduced by using a TCP-like slow growth in segment sizes.
  • In order to store received data that is not ready for playback, buffering is needed. From an end user's perspective, the required buffer size is of little relevance. However, for the development of multilink on-demand streaming solutions on devices with scarce memory, the maximum required buffer capacity represents an important design factor.
  • The maximum expected buffer size increases with the range size determined by the client and is in close correlation with the required startup latency. The only major difference between the two metrics is their dependence on the number of used interfaces. As the startup latency is dominated by the slowest link used, the aggregation of additional faster interfaces will not have any significant impact on the startup latency. On the other hand, the buffer capacity increases with the number of used network interfaces, because each interface causes additional out-of-order data to be received and buffered.
  • When using a reception device with memory constraints, it is possible to determine an optimal balance of the playback bitrate and the startup latency based on the limited available buffer capacity.
  • FIG. 15 illustrates the hardware structure of, for example, the client 1 according to an embodiment of the invention. The client 1 includes a bus B or other communication mechanism for communicating information, and a processor/CPU 1004 coupled with the bus B for processing the information. The client 1 also includes a main memory/memory unit 1003, such as a random access memory (RAM) or other dynamic storage device (e.g., dynamic RAM (DRAM), static RAM (SRAM), and synchronous DRAM (SDRAM)), coupled to the bus B for storing information and instructions to be executed by processor/CPU 1004. In addition, the memory unit 1003 may be used for storing temporary variables or other intermediate information during the execution of instructions by the CPU 1004. The client 1 may also further include a read only memory (ROM) or other static storage device (e.g., programmable ROM (PROM), erasable PROM (EPROM), and electrically erasable PROM (EEPROM)) coupled to the bus B for storing static information and instructions for the CPU 1004.
  • The client 1 may also include a disk controller coupled to the bus B to control one or more storage devices for storing information and instructions, such as mass storage 1002, and drive device 1006 (e.g., floppy disk drive, read-only compact disc drive, read/write compact disc drive, compact disc jukebox, tape drive, and removable magneto-optical drive). The storage devices may be added to the client 1 using an appropriate device interface (e.g., small computer system interface (SCSI), integrated device electronics (IDE), enhanced-IDE (E-IDE) also known as Parallel ATA (PATA), Serial ATA (SATA), direct memory access (DMA), or ultra-DMA).
  • The client 1 may also include special purpose logic devices (e.g., application specific integrated circuits (ASICs)) or configurable logic devices (e.g., simple programmable logic devices (SPLDs), complex programmable logic devices (CPLDs), and field programmable gate arrays (FPGAs)).
  • The client 1 may also include a display controller coupled to the bus B to control a display, such as a cathode ray tube (CRT) or a liquid crystal display (LCD), for displaying information to a computer user. The computer system includes input devices, such as a keyboard and a pointing device, for interacting with a computer user and providing information to the processor. The pointing device, for example, may be a mouse, a trackball, or a pointing stick for communicating direction information and command selections to the processor and for controlling cursor movement on the display. In addition, a printer may provide printed listings of data stored and/or generated by the computer system.
  • The client 1 performs at least a portion of the processing steps of the invention in response to the CPU 1004 executing one or more sequences of one or more instructions contained in a memory, such as the memory unit 1003. Such instructions may be read into the memory unit from another computer readable medium, such as the mass storage 1002 or a removable media 1001. One or more processors in a multi-processing arrangement may also be employed to execute the sequences of instructions contained in memory unit 1003. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions. Thus, embodiments are not limited to any specific combination of hardware circuitry and software.
  • As stated above, the client 1 includes at least one computer readable medium 1001 or memory for holding instructions programmed according to the teachings of the invention and for containing data structures, tables, records, or other data described herein. Examples of non-transitory computer readable media are compact discs, hard disks, floppy disks, tape, magneto-optical disks, PROMs (EPROM, EEPROM, flash EPROM), DRAM, SRAM, SDRAM, or any other magnetic medium, compact discs (e.g., CD-ROM), or any other medium from which a computer can read.
  • Stored on any one or on a combination of computer readable media, the present invention includes software for controlling the client 1, for driving a device or devices for implementing the invention, and for enabling the client 1 to interact with a human user. Such software may include, but is not limited to, device drivers, operating systems, development tools, and applications software. Such computer readable media further includes the computer program product of the present invention for performing all or a portion (if processing is distributed) of the processing performed in implementing the invention.
  • The computer code elements on the medium of the present invention may be any interpretable or executable code mechanism, including but not limited to scripts, interpretable programs, dynamic link libraries (DLLs), Java classes, and complete executable programs. Moreover, parts of the processing of the present invention may be distributed for better performance, reliability, and/or cost.
  • The term “computer readable medium” as used herein refers to any medium that participates in providing instructions to the CPU 1004 for execution. A computer readable medium may take many forms, including but not limited to, non-volatile media, and volatile media. Non-volatile media includes, for example, optical, magnetic disks, and magneto-optical disks, such as the mass storage 1002 or the removable media 1001. Volatile media includes dynamic memory, such as the memory unit 1003.
  • Various forms of computer readable media may be involved in carrying out one or more sequences of one or more instructions to the CPU 1004 for execution. For example, the instructions may initially be carried on a magnetic disk of a remote computer. An input coupled to the bus B can receive the data and place the data on the bus B. The bus B carries the data to the memory unit 1003, from which the CPU 1004 retrieves and executes the instructions. The instructions received by the memory unit 1003 may optionally be stored on mass storage 1002 either before or after execution by the CPU 1004.
  • The client 1 also includes a communication interface 1005 coupled to the bus B. The communication interface 1004 provides a two-way data communication coupling to a network that is connected to, for example, a local area network (LAN), or to another communications network such as the Internet. For example, the communication interface 1005 may be a network interface card to attach to any packet switched LAN. As another example, the communication interface 1005 may be an asymmetrical digital subscriber line (ADSL) card, an integrated services digital network (ISDN) card or a modem to provide a data communication link to a corresponding type of communications line. Wireless links may also be implemented. In any such implementation, the communication interface 1005 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.
  • The network typically provides data communication through one or more networks to other data devices. For example, the network may provide a link to another computer through a local network (e.g., a LAN) or through equipment operated by a service provider, which provides communication services through a communications network. The local network and the communications network use, for example, electrical, electromagnetic, or optical signals that carry digital data streams, and the associated physical layer (e.g., CAT 5 cable, coaxial cable, optical fiber, etc). Moreover, the network may provide a link to a mobile device such as a personal digital assistant (PDA) laptop computer, or cellular telephone.
  • Modifications to embodiments of the invention described in the foregoing are susceptible to being implemented without departing from the scope of the invention as defined by the appended claims.
  • “Enhancing Video-on-Demand Playout over Multiple Heterogeneous Access Networks” by D. Kasper, K. R. Evensen, P. E. Engelstad, A. F. Hansen, P. Halvorsen, and C. Griwodz, Consumer Communications and Networking Conference (CCNC), ed. by IEEE, IEEE (ISBN:978-1-4244-5176-0) is incorporated herein by reference.

Claims (50)

1. A method of requesting information, comprising:
generating a plurality of range requests using a client having a plurality of network interfaces, the range requests each corresponding to a data range of a data resource;
transmitting, from the client having the plurality of network interfaces, a plurality of range requests to a server via the plurality of network interfaces;
receiving, at the client and from the server, a plurality of segments via the plurality of network interfaces such that each segment is received via a network interface which transmitted the corresponding range request, each segment including a portion of the data resource; and
reassembling the data resource using the plurality of segments.
2. The method of claim 1, wherein the plurality of range requests are generated based on a current status of the plurality of network interfaces.
3. The method of claim 2, wherein the current status of the plurality of network interfaces is determined based on available bandwidth and round-trip delay of each of the plurality of network interfaces.
4. The method of claim 1, further comprising:
obtaining, at the client, metadata of the data resource; and
generating at least one of the plurality of range requests based on the obtained metadata of the data resource.
5. The method of claim 1, wherein each of the plurality of network interfaces is connected to the server, at least partially, by way of a different network provider and the plurality of range requests are transmitted to the server by way of the different network provider.
6. The method of claim 1, wherein at least two of the plurality of network interfaces are connected to the server, at least partially, by way of a different network provider and the plurality of range requests are transmitted to the server by way of the different network provider.
7. The method of claim 3, wherein at least one of the plurality of range requests is generated based on metadata of the data resource obtained after the client transmits at least one of the plurality of range requests to the server.
8. The method according to claim 1, comprising:
generating, using the client having the plurality of network interfaces, at least one additional range request, each additional range request corresponding to a data range of the data resource;
determining an upper time limit for each at least one additional range request;
transmitting, from the client having a plurality of network interfaces, the at least one additional range request to the server via at least one of the plurality of network interfaces before the determined upper time limit;
receiving, at the client and from the server, a plurality of additional segments via the plurality of network interfaces such that each additional segment is received via a network interface which transmitted the corresponding additional range request, each additional segment including a portion of the data resource; and
reassembling the data resource using at least the additional plurality of segments.
9. The method of claim 8, wherein the at least one additional range request is included in the plurality of range requests.
10. The method of claim 9, wherein the transmitting the plurality of range requests further comprises determining a distribution of range requests over the plurality of network interfaces, and
wherein the generating the plurality of range requests further comprises determining a size of each of the plurality of range requests.
11. The method of claim 9, wherein the plurality of range requests includes two initial range requests which are transmitted before any additional range request is transmitted.
12. The method of claim 8, further comprising:
determining a limit on a number of additional range requests generated and/or transmitted before reception of a segment corresponding to a previous range request is complete.
13. The method of claim 8, wherein the upper time limit corresponds to a latest time by which the at least one additional range request is sent such that the at least one additional range request, transmitted via one of the plurality of network interfaces, is expected to be received by the server before the server finishes transmitting a segment corresponding a last received range request from the one of the plurality of network interfaces.
14. The method of claim 8, wherein the transmitting further comprises transmitting the at least one additional range request to the server no earlier than a determined lower time limit.
15. The method of claim 14, wherein the lower time limit and the upper time limit are updated in response to any significant variation in bandwidth or round trip delay.
16. The method of claim 8, wherein the at least one additional range request is generated based on a current status of the plurality of network interfaces.
17. The method of claim 8, further comprising:
obtaining, at the client, metadata of the data resource; and
generating the at least one additional range request based on the obtained metadata of the data resource.
18. The method of claim 8, wherein each of the plurality of network interfaces are connected to the server, at least partially, by way of a different network provider and the at least one additional range request is transmitted to the server by way of the different network provider.
19. The method of claim 8, wherein at least two of the plurality of network interfaces are connected to the server, at least partially, by way of a different network provider and the at least one additional range request is transmitted to the server by way of the different network provider.
20. A method of transferring information, comprising:
segmenting, using a client having a plurality of network interfaces, a data resource into a plurality of segments based on metadata of the data resource;
generating a plurality of range requests, the range requests each including a segment of the plurality of segments which corresponds to a data range of the data resource;
transmitting, from the client having the plurality of network interfaces, the plurality of range requests to a server via the plurality of network interfaces for reassembly of the data resource using the plurality of segments at the server.
21. The method according to claim 20, wherein the segmenting the data resource into the plurality of segments is further based on a current status of the plurality of network interfaces.
22. A device for requesting information, comprising:
a plurality of network interfaces;
a generation unit configured to generate a plurality of range requests, the range requests each corresponding to a data range of a data resource;
a communications unit configured to transmit the plurality of range requests to a server via the plurality of network interfaces, the server segmenting the data resource based on the range requests into a plurality of segments,
wherein the communications unit is further configured to receive the plurality of segments, which are transmitted by the server, via the plurality of network interfaces such that each segment is received via a network interface which transmitted the corresponding range request; and
a reassembly unit configured to reassemble the data resource using the plurality of segments received by the communications unit.
23. The device of claim 22, wherein the plurality of range requests are generated based on a current status of the plurality of network interfaces.
24. The device of claim 23, wherein the current status of the plurality of network interfaces is determined based on available bandwidth and round-trip delay of each of the plurality of network interfaces.
25. The device of claim 22, wherein the generation unit obtains metadata of the data resource and generates at least one of the plurality of range requests based on the obtained metadata of the data resource.
26. The device of claim 22, wherein each of the plurality of network interfaces is connected to the server, at least partially, by way of a different network provider and the plurality of range requests are transmitted to the server by way of the different network provider.
27. The device of claim 22, wherein at least two of the plurality of network interfaces are connected to the server, at least partially, by way of a different network provider and the plurality of range requests are transmitted to the server by way of the different network provider.
28. The device according to claim 22, wherein
the generation unit is further configured to generate at least one additional range request, each additional range request corresponding to a data range of the data resource,
a determining unit configured to determine an upper time limit for each at least one additional range request;
the communication unit is further configured to transmit the at least one additional range request to the server via at least one of the plurality of network interfaces before the upper time limit determined by the determining unit,
the communications unit is further configured to receive a plurality of additional segments, which are transmitted by the server, via the plurality of network interfaces such that each additional segment is received via a network interface which transmitted the corresponding additional range request, and
the reassembly unit is further configured to reassemble the data resource using at least the plurality of additional segments received by the communications unit.
29. The device of claim 28, wherein the at least one additional range request is included in the plurality of range requests.
30. The device of claim 28, wherein the communication unit is further configured to determine a distribution of range requests over the plurality of network interfaces, and
wherein the generation unit is further configured to determine a size of each of the plurality of range requests.
31. The device of claim 28, wherein the plurality of range requests includes two initial range requests which are transmitted before any additional range request is transmitted.
32. The device of claim 28, wherein the determination unit is further configured to determine a limit on a number of additional range requests generated and/or transmitted before reception of a segment corresponding to a previous range request is complete.
33. The device of claim 28, wherein the upper time limit corresponds to a latest time by which the at least one additional range request is sent such that the at least one additional range request, transmitted via one of the plurality of network interfaces, is expected to be received by the server before the server finishes transmitting a segment corresponding a last received range request from the one of the plurality of network interfaces.
34. The device of claim 28, wherein the communication unit is further configured to transmit the at least one additional range request to the server no earlier than a determined lower time limit.
35. The device of claim 28, wherein the at least one additional range request is generated based on a current status of the plurality of network interfaces.
36. The device of claim 35, wherein the lower time limit and the upper time limit are updated in response to any significant variation in bandwidth or round trip delay.
37. The device of claim 28, wherein the generation unit obtains metadata of the data resource and generates at least one of the additional range requests based on the obtained metadata of the data resource.
38. The device of claim 28, wherein each of the plurality of network interfaces is connected to the server, at least partially, by way of a different network provider and the at least one additional range request is transmitted to the server by way of the different network provider.
39. The device of claim 28, wherein at least two of the plurality of network interfaces are connected to the server, at least partially, by way of a different network provider and the at least one additional range request is transmitted to the server by way of the different network provider.
40. The method according to claim 1, comprising:
receiving the plurality of range requests at the server;
obtaining the data resource;
segmenting the data resource based on the range requests received at the server into the plurality of segments; and
transmitting the plurality of segments to the client via the network interface which transmitted the corresponding range request.
41. The method according to claim 40, wherein the obtaining step further comprises obtaining the data resource from an external source.
42. The method according to claim 40, wherein the obtaining step further comprises generating the data resource.
43. The method according to claim 40, wherein the obtaining step further comprises reading the data resource from a storage unit of the server.
44. The method according to claim 1, further comprising:
transmitting, from the client, the plurality of range requests to a plurality of servers via the plurality of network interfaces, each of the plurality of servers having at least a portion of the data resource;
receiving, at the client and from the plurality of servers, the plurality of segments via the plurality of network interfaces.
45. The method according to claim 1, wherein the generating the plurality of range requests further includes:
adjusting segment sizes of the plurality of segments by adjusting the range requests based on a current status of the plurality of network interfaces, the current status including throughput or delay of the network interfaces.
46. A method of playing back segmented information, comprising:
generating a plurality of range requests using a client having a plurality of network interfaces, the range requests each corresponding to a data range of a data resource;
transmitting, from the client having the plurality of network interfaces, a plurality of range requests to a server via the plurality of network interfaces;
receiving, at the client and from the server, a plurality of segments via the plurality of network interfaces such that each segment is received via a network interface which transmitted the corresponding range request, each segment including a portion of the data resource corresponding to a corresponding range request;
playing back the data resource using the plurality of segments.
47. The method according to claim 46, wherein the playing back further comprises:
determining current throughput conditions for each network interface;
dynamically adapting a bitrate of the playing back based on the current throughput conditions.
48. The method according to claim 46, wherein the playing back further comprises, for any point in time within a predetermined time period:
determining a number of bytes ready for playback;
determining a maximum possible difference between a number of bytes ready for playback and a desired constant-rate playback bitrate; and
determining a required startup latency for the playing back based on the determined maximum possible difference.
49. The method according to claim 46, wherein the playing back further comprises, for any point in time within a predetermined time period:
determining a buffer requirement for the playing back of the data resource based on determining the number of received unready bytes and the number of playback-ready bytes needed for a desired constant-rate playback bitrate.
50. The method according to claim 46, wherein the generating the plurality of range requests further includes:
adjusting segment sizes of the plurality of segments by adjusting the range requests based on a current status of the plurality of network interfaces, the current status including throughput or delay of the network interfaces.
US12/713,939 2010-02-26 2010-02-26 Data segmentation, request and transfer method Active 2031-11-10 US8516147B2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US12/713,939 US8516147B2 (en) 2010-02-26 2010-02-26 Data segmentation, request and transfer method
US13/946,610 US20130304875A1 (en) 2010-02-26 2013-07-19 Data segmentation, request and transfer method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US12/713,939 US8516147B2 (en) 2010-02-26 2010-02-26 Data segmentation, request and transfer method

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US13/946,610 Division US20130304875A1 (en) 2010-02-26 2013-07-19 Data segmentation, request and transfer method

Publications (3)

Publication Number Publication Date
US20110213827A1 US20110213827A1 (en) 2011-09-01
US20120011191A2 true US20120011191A2 (en) 2012-01-12
US8516147B2 US8516147B2 (en) 2013-08-20

Family

ID=44505877

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/713,939 Active 2031-11-10 US8516147B2 (en) 2010-02-26 2010-02-26 Data segmentation, request and transfer method
US13/946,610 Abandoned US20130304875A1 (en) 2010-02-26 2013-07-19 Data segmentation, request and transfer method

Family Applications After (1)

Application Number Title Priority Date Filing Date
US13/946,610 Abandoned US20130304875A1 (en) 2010-02-26 2013-07-19 Data segmentation, request and transfer method

Country Status (1)

Country Link
US (2) US8516147B2 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130103849A1 (en) * 2011-09-21 2013-04-25 Qualcomm Incorporated Signaling characteristics of segments for network streaming of media data
US20160092136A1 (en) * 2014-09-30 2016-03-31 Emc Corporation Method and apparatus for cost-based load balancing for port selection
US9747040B1 (en) * 2015-06-30 2017-08-29 EMC IP Holding Company LLC Method and system for machine learning for write command selection based on technology feedback
CN107637046A (en) * 2015-03-17 2018-01-26 三星电子株式会社 It is multi-link to improve the method and apparatus of data transmission rate for controlling
US20180375922A1 (en) * 2017-06-21 2018-12-27 Walmart Apollo, Llc Systems and methods for dynamic file creation
US10785144B2 (en) * 2016-12-30 2020-09-22 Equinix, Inc. Latency equalization

Families Citing this family (43)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7515710B2 (en) 2006-03-14 2009-04-07 Divx, Inc. Federated digital rights management scheme including trusted systems
CN105072454B (en) 2009-01-07 2019-04-19 索尼克Ip股份有限公司 For the specialization of the media guidance of online content, centralization, automation creation
EP2507995A4 (en) 2009-12-04 2014-07-09 Sonic Ip Inc Elementary bitstream cryptographic material transport systems and methods
US9420022B2 (en) * 2010-12-17 2016-08-16 Microsoft Technology Licensing, Llc Media requests to counter latency and minimize network bursts
US20130166769A1 (en) * 2010-12-20 2013-06-27 Awind, Inc. Receiving device, screen frame transmission system and method
US8914534B2 (en) 2011-01-05 2014-12-16 Sonic Ip, Inc. Systems and methods for adaptive bitrate streaming of media stored in matroska container files using hypertext transfer protocol
US9467708B2 (en) 2011-08-30 2016-10-11 Sonic Ip, Inc. Selection of resolutions for seamless resolution switching of multimedia content
US8964977B2 (en) 2011-09-01 2015-02-24 Sonic Ip, Inc. Systems and methods for saving encoded media streamed using adaptive bitrate streaming
US8909922B2 (en) 2011-09-01 2014-12-09 Sonic Ip, Inc. Systems and methods for playing back alternative streams of protected content protected using common cryptographic information
WO2013042920A2 (en) * 2011-09-19 2013-03-28 에스케이텔레콤 주식회사 Device and method for supporting a simultaneous transmission service of multiple network based data
JP5799706B2 (en) * 2011-09-26 2015-10-28 富士通株式会社 Search request processing device
CN103095517B (en) * 2011-11-04 2016-12-07 华为技术有限公司 Stream media transmission quality assessment and information getting method and relevant device and system
US8640174B2 (en) 2012-03-01 2014-01-28 Motorola Mobility Llc Method for retrieving content, wireless communication device and communication system
US9143812B2 (en) 2012-06-29 2015-09-22 Sonic Ip, Inc. Adaptive streaming of multimedia
KR20140028961A (en) * 2012-08-31 2014-03-10 삼성전자주식회사 Method for data transmission and an electronic device thereof
US9438385B2 (en) * 2012-09-13 2016-09-06 First Principles, Inc. Data stream division to increase data transmission rates
US9100904B2 (en) * 2012-09-13 2015-08-04 First Principles, Inc. Data stream division to increase data transmission rates
TWI491228B (en) * 2012-12-26 2015-07-01 Ind Tech Res Inst Apparatus and method of bandwidth aggregation for radio accessing on multi-networks
US9313510B2 (en) 2012-12-31 2016-04-12 Sonic Ip, Inc. Use of objective quality measures of streamed content to reduce streaming bandwidth
US9191457B2 (en) 2012-12-31 2015-11-17 Sonic Ip, Inc. Systems, methods, and media for controlling delivery of content
CN103916448B (en) * 2013-01-09 2018-05-18 深圳市腾讯计算机系统有限公司 Data transmission method, system and corresponding cloud transmission platform based on cloud transmission platform
US9906785B2 (en) 2013-03-15 2018-02-27 Sonic Ip, Inc. Systems, methods, and media for transcoding video data according to encoding parameters indicated by received metadata
US10397292B2 (en) 2013-03-15 2019-08-27 Divx, Llc Systems, methods, and media for delivery of content
KR102164457B1 (en) 2013-04-25 2020-10-14 삼성전자주식회사 Electronic device for multiple radio access and method thereof
US9094737B2 (en) 2013-05-30 2015-07-28 Sonic Ip, Inc. Network video streaming with trick play based on separate trick play files
US9967305B2 (en) 2013-06-28 2018-05-08 Divx, Llc Systems, methods, and media for streaming media content
KR102096425B1 (en) * 2013-08-20 2020-04-03 삼성전자주식회사 Method and apparatus for distributing traffic using a plurality of network interfaces in a wireless communication system
US9742861B2 (en) * 2013-11-20 2017-08-22 Opanga Networks, Inc. Fractional pre-delivery of content to user devices for uninterrupted playback
US20150156264A1 (en) * 2013-12-04 2015-06-04 International Business Machines Corporation File access optimization using strategically partitioned and positioned data in conjunction with a collaborative peer transfer system
US9661106B2 (en) * 2013-12-04 2017-05-23 Facebook, Inc. Uploading and transcoding media files
US9866878B2 (en) 2014-04-05 2018-01-09 Sonic Ip, Inc. Systems and methods for encoding and playing back video at different frame rates using enhancement layers
KR101664423B1 (en) * 2014-08-07 2016-10-10 엘지전자 주식회사 Mobile terminal and method for controlling the same
FR3030172B1 (en) * 2014-12-11 2018-02-09 Sagemcom Broadband Sas METHOD AND DEVICES FOR TRANSMITTING A DATA STREAM ACCORDING TO A MULTIPOINT TRANSMISSION MODE
KR102352679B1 (en) 2015-03-25 2022-01-18 삼성전자주식회사 Device supporting multipath tcp, and method of receiving video data of device by streaming
JP6581665B2 (en) * 2015-04-10 2019-09-25 華為技術有限公司Huawei Technologies Co.,Ltd. Data transmission method and terminal
KR102372488B1 (en) * 2015-05-29 2022-03-10 삼성전자 주식회사 Method and Apparatus of scheduling for wireless packet network
KR102425517B1 (en) 2015-09-04 2022-07-27 삼성전자주식회사 Apparatus and method for uploading data in mobile communication system supporting a plurality of radio access interfaces
CN106101184B (en) * 2016-05-27 2019-10-22 华为技术有限公司 A kind of document down loading method and playback equipment
FR3060925B1 (en) * 2016-12-21 2020-02-21 Orange SERVER AND CLIENT SUITABLE FOR IMPROVING THE ROUNDTURN TIME OF AN HTTP REQUEST
US10498795B2 (en) 2017-02-17 2019-12-03 Divx, Llc Systems and methods for adaptive switching between multiple content delivery networks during adaptive bitrate streaming
US10447765B2 (en) * 2017-07-13 2019-10-15 International Business Machines Corporation Shared memory device
US11509747B2 (en) * 2018-06-21 2022-11-22 British Telecommunications Public Limited Company Path selection for content delivery network
US20200106699A1 (en) * 2018-10-01 2020-04-02 Citrix Systems, Inc. Systems and methods for multilink wan connectivity for saas applications

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5953506A (en) * 1996-12-17 1999-09-14 Adaptive Media Technologies Method and apparatus that provides a scalable media delivery system
US6128483A (en) * 1996-11-19 2000-10-03 Ericsson, Inc. Simultaneous over the air data download to multiple radios
US20020091789A1 (en) * 1998-12-03 2002-07-11 Sanjeev Katariya Scalable computing system for presenting customized aggregation of information
US20070011271A1 (en) * 2005-05-20 2007-01-11 Baker David V Multi-source data retrieval system
US20070286198A1 (en) * 2001-12-21 2007-12-13 Muirhead Charles S System for supply chain management of virtual private network services

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU6882998A (en) 1997-03-31 1998-10-22 Broadband Associates Method and system for providing a presentation on a network
US20030223450A1 (en) 2002-05-29 2003-12-04 Bender Paul E. Aggregating multiple air interfaces with a multi-link protocol
TWI232051B (en) 2002-08-09 2005-05-01 Quanta Comp Inc System and method for supporting mobile internet protocol using multiple separate tunnels
US9357033B2 (en) 2003-06-17 2016-05-31 Citrix Systems, Inc. Method and system for dynamic interleaving
WO2005069552A1 (en) 2004-01-14 2005-07-28 Gridiron Software, Inc. Redundant pipelined file transfer
US20050210122A1 (en) 2004-03-22 2005-09-22 Qualcomm Incorporated HTTP acceleration over a network link
US7539175B2 (en) 2004-11-19 2009-05-26 The Trustees Of Stevens Institute Of Technology Multi-access terminal with capability for simultaneous connectivity to multiple communication channels
US7860962B2 (en) 2005-08-09 2010-12-28 At&T Intellectual Property I, L.P. Media download method and system based on connection speed
EP1764970A1 (en) 2005-09-19 2007-03-21 Matsushita Electric Industrial Co., Ltd. Multiple interface mobile node with simultaneous home- and foreign network connection
US7783773B2 (en) 2006-07-24 2010-08-24 Microsoft Corporation Glitch-free media streaming
US7883419B2 (en) 2007-01-31 2011-02-08 Hewlett-Packard Development Company, L.P. Proxy device to packetize data to send to a wireless network
US8964757B2 (en) 2009-12-18 2015-02-24 Qualcomm Incorporated HTTP optimization, multi-homing, mobility and priority

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6128483A (en) * 1996-11-19 2000-10-03 Ericsson, Inc. Simultaneous over the air data download to multiple radios
US5953506A (en) * 1996-12-17 1999-09-14 Adaptive Media Technologies Method and apparatus that provides a scalable media delivery system
US20020091789A1 (en) * 1998-12-03 2002-07-11 Sanjeev Katariya Scalable computing system for presenting customized aggregation of information
US6564251B2 (en) * 1998-12-03 2003-05-13 Microsoft Corporation Scalable computing system for presenting customized aggregation of information
US20070286198A1 (en) * 2001-12-21 2007-12-13 Muirhead Charles S System for supply chain management of virtual private network services
US7764700B2 (en) * 2001-12-21 2010-07-27 Hewlett-Packard Development Company, L.P. System for supply chain management of virtual private network services
US20070011271A1 (en) * 2005-05-20 2007-01-11 Baker David V Multi-source data retrieval system

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9445136B2 (en) * 2011-09-21 2016-09-13 Qualcomm Incorporated Signaling characteristics of segments for network streaming of media data
US20130103849A1 (en) * 2011-09-21 2013-04-25 Qualcomm Incorporated Signaling characteristics of segments for network streaming of media data
US20160092136A1 (en) * 2014-09-30 2016-03-31 Emc Corporation Method and apparatus for cost-based load balancing for port selection
CN105573839A (en) * 2014-09-30 2016-05-11 伊姆西公司 Method and apparatus cost-based load balancing for port selection
US10725684B2 (en) * 2014-09-30 2020-07-28 EMC IP Holding Company LLC Method and apparatus for cost-based load balancing for port selection
CN107637046A (en) * 2015-03-17 2018-01-26 三星电子株式会社 It is multi-link to improve the method and apparatus of data transmission rate for controlling
US10999348B2 (en) 2015-03-17 2021-05-04 Samsung Electronics Co., Ltd. Method and apparatus for controlling multi-connection for data transmission rate improvement
CN112118177A (en) * 2015-03-17 2020-12-22 三星电子株式会社 Method and apparatus for controlling multiple connections to increase data transfer rate
US9747040B1 (en) * 2015-06-30 2017-08-29 EMC IP Holding Company LLC Method and system for machine learning for write command selection based on technology feedback
AU2017387920B2 (en) * 2016-12-30 2021-02-04 Equinix, Inc. Latency equalization
US10785144B2 (en) * 2016-12-30 2020-09-22 Equinix, Inc. Latency equalization
US20180375922A1 (en) * 2017-06-21 2018-12-27 Walmart Apollo, Llc Systems and methods for dynamic file creation
US10855744B2 (en) * 2017-06-21 2020-12-01 Walmart Apollo, Llc Systems and methods for dynamic file creation

Also Published As

Publication number Publication date
US20110213827A1 (en) 2011-09-01
US8516147B2 (en) 2013-08-20
US20130304875A1 (en) 2013-11-14

Similar Documents

Publication Publication Date Title
US8516147B2 (en) Data segmentation, request and transfer method
JP6178523B2 (en) Transport accelerator implementing request manager and connection manager functionality
EP2962435B1 (en) Link-aware streaming adaptation
US9043467B2 (en) Adaptive chunked and content-aware pacing of multi-media delivery over HTTP transport and network controlled bit rate selection
US8812712B2 (en) Proxy-driven content rate selection for streaming media servers
US9930097B2 (en) Transport accelerator systems and methods
US10951530B2 (en) Systems and methods for adjusting a congestion window value of a content delivery network
US20150271231A1 (en) Transport accelerator implementing enhanced signaling
US9596323B2 (en) Transport accelerator implementing client side transmission functionality
EP3465483A1 (en) System and method for improving an aggregated throughput of simultaneous connections
KR20040068880A (en) Reactive bandwidth control for streaming data
EP2391953A1 (en) Application, usage&radio link aware transport network scheduler
WO2018121742A1 (en) Method and device for transmitting stream data
Kaspar et al. Enhancing video-on-demand playout over multiple heterogeneous access networks
WO2016105846A1 (en) Link-aware streaming adaptation
US9131251B2 (en) Use of a receive-window size advertised by a client to a content server to change a video stream bitrate streamed by the content server
Schier et al. Using DCCP: Issues and improvements
JP2007515872A (en) Provide feedback using general NACK report block and erasure RLE report block
WO2023060029A1 (en) Techniques for client-controlled pacing of media streaming

Legal Events

Date Code Title Description
AS Assignment

Owner name: SIMULA INNOVATION AS, NORWAY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KASPAR, DOMINIK;EVENSEN, KRISTIAN R.;ENGELSTAD, PAAL E.;AND OTHERS;SIGNING DATES FROM 20100303 TO 20100308;REEL/FRAME:024375/0749

Owner name: TELENOR ASA, NORWAY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KASPAR, DOMINIK;EVENSEN, KRISTIAN R.;ENGELSTAD, PAAL E.;AND OTHERS;SIGNING DATES FROM 20100303 TO 20100308;REEL/FRAME:024375/0749

AS Assignment

Owner name: SIMULA INNOVATION SA, NORWAY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TELENOR ASA;REEL/FRAME:025588/0453

Effective date: 20110103

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: SIMULA INNOVATION AS, NORWAY

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE'S NAME PREVIOUSLY RECORDED ON REEL 025588 FRAME 0453. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT OF ASSIGNOR'S INTEREST;ASSIGNOR:TELENOR ASA;REEL/FRAME:031207/0874

Effective date: 20110103

AS Assignment

Owner name: CELERWAY COMMUNICATION AS, NORWAY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIMULA INNOVATION AS;REEL/FRAME:041495/0827

Effective date: 20170227

FPAY Fee payment

Year of fee payment: 4

SULP Surcharge for late payment
MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8